blob: 3cbdae2fed17822ab9ab27fa13d1ea66c1207711 [file] [log] [blame]
gstepanov023c1e42015-04-08 15:50:19 +03001import os
2import pickle
koder aka kdanilov7acd6bd2015-02-12 14:28:30 -08003import sys
koder aka kdanilovda45e882015-04-06 02:24:42 +03004import json
koder aka kdanilov2c473092015-03-29 17:12:13 +03005import Queue
koder aka kdanilov7acd6bd2015-02-12 14:28:30 -08006import pprint
koder aka kdanilove21d7472015-02-14 19:02:04 -08007import logging
koder aka kdanilov7acd6bd2015-02-12 14:28:30 -08008import argparse
koder aka kdanilov2c473092015-03-29 17:12:13 +03009import threading
10import collections
koder aka kdanilov7acd6bd2015-02-12 14:28:30 -080011
koder aka kdanilov2c473092015-03-29 17:12:13 +030012from concurrent.futures import ThreadPoolExecutor
Yulia Portnova407ca952015-04-10 10:38:15 +030013import formatters
14import report
koder aka kdanilov2c473092015-03-29 17:12:13 +030015
16import utils
koder aka kdanilove06762a2015-03-22 23:32:09 +020017import ssh_utils
koder aka kdanilovda45e882015-04-06 02:24:42 +030018import start_vms
koder aka kdanilove06762a2015-03-22 23:32:09 +020019from nodes import discover
koder aka kdanilov2c473092015-03-29 17:12:13 +030020from nodes.node import Node
gstepanovcd256d62015-04-07 17:47:32 +030021from config import cfg_dict, parse_config
koder aka kdanilovda45e882015-04-06 02:24:42 +030022from tests.itest import IOPerfTest, PgBenchTest
koder aka kdanilov2c473092015-03-29 17:12:13 +030023from sensors.api import start_monitoring
24
25
koder aka kdanilove21d7472015-02-14 19:02:04 -080026logger = logging.getLogger("io-perf-tool")
koder aka kdanilove21d7472015-02-14 19:02:04 -080027
28
koder aka kdanilov2c473092015-03-29 17:12:13 +030029def setup_logger(logger, level=logging.DEBUG):
30 logger.setLevel(level)
31 ch = logging.StreamHandler()
32 ch.setLevel(level)
33 logger.addHandler(ch)
Yulia Portnova7ddfa732015-02-24 17:32:58 +020034
koder aka kdanilov2c473092015-03-29 17:12:13 +030035 log_format = '%(asctime)s - %(levelname)-6s - %(name)s - %(message)s'
36 formatter = logging.Formatter(log_format,
37 "%H:%M:%S")
38 ch.setFormatter(formatter)
koder aka kdanilov7acd6bd2015-02-12 14:28:30 -080039
40
Yulia Portnova7ddfa732015-02-24 17:32:58 +020041def format_result(res, formatter):
koder aka kdanilove21d7472015-02-14 19:02:04 -080042 data = "\n{0}\n".format("=" * 80)
43 data += pprint.pformat(res) + "\n"
44 data += "{0}\n".format("=" * 80)
koder aka kdanilovfe056622015-02-19 08:46:15 -080045 templ = "{0}\n\n====> {1}\n\n{2}\n\n"
Yulia Portnova7ddfa732015-02-24 17:32:58 +020046 return templ.format(data, formatter(res), "=" * 80)
koder aka kdanilove21d7472015-02-14 19:02:04 -080047
48
koder aka kdanilov5d589b42015-03-26 12:25:51 +020049def connect_one(node):
50 try:
koder aka kdanilov2c473092015-03-29 17:12:13 +030051 ssh_pref = "ssh://"
52 if node.conn_url.startswith(ssh_pref):
53 url = node.conn_url[len(ssh_pref):]
54 node.connection = ssh_utils.connect(url)
55 else:
56 raise ValueError("Unknown url type {0}".format(node.conn_url))
koder aka kdanilov3a6633e2015-03-26 18:20:00 +020057 except Exception:
koder aka kdanilov2c473092015-03-29 17:12:13 +030058 logger.exception("During connect to {0}".format(node))
koder aka kdanilov5d589b42015-03-26 12:25:51 +020059
60
61def connect_all(nodes):
koder aka kdanilov2c473092015-03-29 17:12:13 +030062 logger.info("Connecting to nodes")
63 with ThreadPoolExecutor(32) as pool:
64 list(pool.map(connect_one, nodes))
koder aka kdanilovda45e882015-04-06 02:24:42 +030065 logger.info("All nodes connected successfully")
koder aka kdanilov2c473092015-03-29 17:12:13 +030066
67
68def save_sensors_data(q):
69 logger.info("Start receiving sensors data")
koder aka kdanilovda45e882015-04-06 02:24:42 +030070 sensor_data = []
koder aka kdanilov2c473092015-03-29 17:12:13 +030071 while True:
72 val = q.get()
73 if val is None:
koder aka kdanilovda45e882015-04-06 02:24:42 +030074 print sensor_data
75 q.put(sensor_data)
koder aka kdanilov2c473092015-03-29 17:12:13 +030076 break
koder aka kdanilovda45e882015-04-06 02:24:42 +030077 sensor_data.append(val)
koder aka kdanilov2c473092015-03-29 17:12:13 +030078 logger.info("Sensors thread exits")
79
80
81def test_thread(test, node, barrier):
82 try:
83 logger.debug("Run preparation for {0}".format(node.conn_url))
84 test.pre_run(node.connection)
85 logger.debug("Run test for {0}".format(node.conn_url))
86 test.run(node.connection, barrier)
87 except:
88 logger.exception("In test {0} for node {1}".format(test, node))
89
90
91def run_tests(config, nodes):
92 tool_type_mapper = {
93 "io": IOPerfTest,
94 "pgbench": PgBenchTest,
95 }
96
97 test_nodes = [node for node in nodes
98 if 'testnode' in node.roles]
99
100 res_q = Queue.Queue()
101
koder aka kdanilovda45e882015-04-06 02:24:42 +0300102 for test in config['tests']:
gstepanov023c1e42015-04-08 15:50:19 +0300103 for test in config['tests'][test]['internal_tests']:
104 for name, params in test.items():
105 logger.info("Starting {0} tests".format(name))
koder aka kdanilov2c473092015-03-29 17:12:13 +0300106
gstepanov023c1e42015-04-08 15:50:19 +0300107 threads = []
108 barrier = utils.Barrier(len(test_nodes))
109 for node in test_nodes:
110 msg = "Starting {0} test on {1} node"
111 logger.debug(msg.format(name, node.conn_url))
112 test = tool_type_mapper[name](params, res_q.put)
113 th = threading.Thread(None, test_thread, None,
114 (test, node, barrier))
115 threads.append(th)
116 th.daemon = True
117 th.start()
koder aka kdanilov2c473092015-03-29 17:12:13 +0300118
gstepanov023c1e42015-04-08 15:50:19 +0300119 for th in threads:
120 th.join()
koder aka kdanilov2c473092015-03-29 17:12:13 +0300121
gstepanov023c1e42015-04-08 15:50:19 +0300122 results = []
123 while not res_q.empty():
124 results.append(res_q.get())
125 # logger.info("Get test result {0!r}".format(results[-1]))
126 yield name, results
koder aka kdanilov2c473092015-03-29 17:12:13 +0300127
128
129def parse_args(argv):
130 parser = argparse.ArgumentParser(
131 description="Run disk io performance test")
132
133 parser.add_argument("-l", dest='extra_logs',
134 action='store_true', default=False,
135 help="print some extra log info")
136
gstepanov4861d712015-04-09 13:28:02 +0300137 parser.add_argument("-b", '--build_description',
138 type=str, default="Build info")
gstepanovaffcdb12015-04-07 17:18:29 +0300139 parser.add_argument("-i", '--build_id', type=str, default="id")
140 parser.add_argument("-t", '--build_type', type=str, default="GA")
141 parser.add_argument("-u", '--username', type=str, default="admin")
koder aka kdanilovda45e882015-04-06 02:24:42 +0300142 parser.add_argument("-o", '--output-dest', nargs="*")
143 parser.add_argument("config_file", nargs="?", default="config.yaml")
koder aka kdanilov2c473092015-03-29 17:12:13 +0300144
145 return parser.parse_args(argv[1:])
146
147
koder aka kdanilovda45e882015-04-06 02:24:42 +0300148def log_nodes_statistic(_, ctx):
149 nodes = ctx.nodes
koder aka kdanilov2c473092015-03-29 17:12:13 +0300150 logger.info("Found {0} nodes total".format(len(nodes)))
151 per_role = collections.defaultdict(lambda: 0)
152 for node in nodes:
153 for role in node.roles:
154 per_role[role] += 1
155
156 for role, count in sorted(per_role.items()):
157 logger.debug("Found {0} nodes with role {1}".format(count, role))
158
159
160def log_sensors_config(cfg):
koder aka kdanilov5d589b42015-03-26 12:25:51 +0200161 pass
162
163
koder aka kdanilovda45e882015-04-06 02:24:42 +0300164def connect_stage(cfg, ctx):
165 ctx.clear_calls_stack.append(disconnect_stage)
166 connect_all(ctx.nodes)
167
168
169def discover_stage(cfg, ctx):
170 if 'discover' in cfg:
171 discover_objs = [i.strip() for i in cfg['discover'].strip().split(",")]
172 ctx.nodes.extend(discover.discover(discover_objs, cfg['clouds']))
173
174 for url, roles in cfg.get('explicit_nodes', {}).items():
175 ctx.nodes.append(Node(url, roles.split(",")))
176
177
178def deploy_sensors_stage(cfg_dict, ctx):
179 ctx.clear_calls_stack.append(remove_sensors_stage)
180 if 'sensors' not in cfg_dict:
181 return
182
183 cfg = cfg_dict.get('sensors')
184 sens_cfg = []
185
186 for role, sensors_str in cfg["roles_mapping"].items():
187 sensors = [sens.strip() for sens in sensors_str.split(",")]
188
189 collect_cfg = dict((sensor, {}) for sensor in sensors)
190
191 for node in ctx.nodes:
192 if role in node.roles:
193 sens_cfg.append((node.connection, collect_cfg))
194
195 log_sensors_config(sens_cfg)
196
197 ctx.sensor_cm = start_monitoring(cfg["receiver_uri"], None,
198 connected_config=sens_cfg)
199
200 ctx.sensors_control_queue = ctx.sensor_cm.__enter__()
201
202 th = threading.Thread(None, save_sensors_data, None,
203 (ctx.sensors_control_queue,))
204 th.daemon = True
205 th.start()
206 ctx.sensor_listen_thread = th
207
208
209def remove_sensors_stage(cfg, ctx):
210 ctx.sensors_control_queue.put(None)
211 ctx.sensor_listen_thread.join()
212 ctx.sensor_data = ctx.sensors_control_queue.get()
213
214
gstepanov023c1e42015-04-08 15:50:19 +0300215def run_all_test(cfg, ctx, store_nodes):
koder aka kdanilovda45e882015-04-06 02:24:42 +0300216 ctx.results = []
217
gstepanov023c1e42015-04-08 15:50:19 +0300218 if 'start_test_nodes' in cfg['tests']:
219 params = cfg['tests']['start_test_nodes']['openstack']
220 for new_node in start_vms.launch_vms(params):
221 new_node.roles.append('testnode')
222 ctx.nodes.append(new_node)
223
koder aka kdanilovda45e882015-04-06 02:24:42 +0300224 if 'tests' in cfg:
gstepanov023c1e42015-04-08 15:50:19 +0300225 store_nodes(ctx.nodes)
koder aka kdanilovda45e882015-04-06 02:24:42 +0300226 ctx.results.extend(run_tests(cfg_dict, ctx.nodes))
227
gstepanov023c1e42015-04-08 15:50:19 +0300228
229def shut_down_vms(cfg, ctx):
230 with open('vm_journal.log') as f:
231 data = str(f.read())
232 nodes = pickle.loads(data)
233
234 for node in nodes:
235 logger.info("Node " + str(node) + " has been loaded")
236
237 logger.info("Removing nodes")
238 start_vms.clear_nodes()
239 logger.info("Nodes has been removed")
240
241
242def store_nodes(nodes):
243 with open('vm_journal.log', 'w+') as f:
244 f.write(pickle.dumps([nodes]))
245 for node in nodes:
246 logger.info("Node " + str(node) + " has been stored")
247
248
249def clear_enviroment(cfg, ctx):
250 if os.path.exists('vm_journal.log'):
251 shut_down_vms(cfg, ctx)
252 os.remove('vm_journal.log')
253
254
255def run_tests_stage(cfg, ctx):
256 # clear nodes that possible were created on previous test running
257 clear_enviroment(cfg, ctx)
258 ctx.clear_calls_stack.append(shut_down_vms)
259 run_all_test(cfg, ctx, store_nodes)
koder aka kdanilovda45e882015-04-06 02:24:42 +0300260
261
262def disconnect_stage(cfg, ctx):
263 for node in ctx.nodes:
264 if node.connection is not None:
265 node.connection.close()
266
267
268def report_stage(cfg, ctx):
269 output_dest = cfg.get('output_dest')
270 if output_dest is not None:
Yulia Portnova407ca952015-04-10 10:38:15 +0300271 if output_dest.endswith(".html"):
272 report.render_html_results(ctx, output_dest)
273 logger.info("Results were stored in %s" % output_dest)
274 else:
275 with open(output_dest, "w") as fd:
276 data = {"sensor_data": ctx.sensor_data,
277 "results": ctx.results}
278 fd.write(json.dumps(data))
koder aka kdanilovda45e882015-04-06 02:24:42 +0300279 else:
280 print "=" * 20 + " RESULTS " + "=" * 20
281 pprint.pprint(ctx.results)
282 print "=" * 60
283
284
285def complete_log_nodes_statistic(cfg, ctx):
286 nodes = ctx.nodes
287 for node in nodes:
288 logger.debug(str(node))
289
290
291class Context(object):
292 def __init__(self):
gstepanovaffcdb12015-04-07 17:18:29 +0300293 self.build_meta = {}
koder aka kdanilovda45e882015-04-06 02:24:42 +0300294 self.nodes = []
295 self.clear_calls_stack = []
296
297
gstepanovcd256d62015-04-07 17:47:32 +0300298def load_config(path):
299 global cfg_dict
300 cfg_dict = parse_config(path)
301
302
koder aka kdanilov3f356262015-02-13 08:06:14 -0800303def main(argv):
koder aka kdanilove06762a2015-03-22 23:32:09 +0200304 opts = parse_args(argv)
koder aka kdanilov2c473092015-03-29 17:12:13 +0300305
306 level = logging.DEBUG if opts.extra_logs else logging.WARNING
307 setup_logger(logger, level)
308
koder aka kdanilovda45e882015-04-06 02:24:42 +0300309 stages = [
310 discover_stage,
311 connect_stage,
312 complete_log_nodes_statistic,
313 # deploy_sensors_stage,
314 run_tests_stage,
315 report_stage
316 ]
koder aka kdanilov2c473092015-03-29 17:12:13 +0300317
gstepanovcd256d62015-04-07 17:47:32 +0300318 load_config(opts.config_file)
319
koder aka kdanilovda45e882015-04-06 02:24:42 +0300320 ctx = Context()
gstepanovaffcdb12015-04-07 17:18:29 +0300321 ctx.build_meta['build_id'] = opts.build_id
322 ctx.build_meta['build_descrption'] = opts.build_description
323 ctx.build_meta['build_type'] = opts.build_type
324 ctx.build_meta['username'] = opts.username
gstepanov023c1e42015-04-08 15:50:19 +0300325 logger.setLevel(logging.INFO)
326 logger.addHandler(logging.FileHandler('log.txt'))
koder aka kdanilovda45e882015-04-06 02:24:42 +0300327 try:
328 for stage in stages:
329 logger.info("Start {0.__name__} stage".format(stage))
gstepanov023c1e42015-04-08 15:50:19 +0300330 print "Start {0.__name__} stage".format(stage)
koder aka kdanilovda45e882015-04-06 02:24:42 +0300331 stage(cfg_dict, ctx)
332 finally:
333 exc, cls, tb = sys.exc_info()
334 for stage in ctx.clear_calls_stack[::-1]:
335 try:
336 logger.info("Start {0.__name__} stage".format(stage))
337 stage(cfg_dict, ctx)
338 except:
339 pass
koder aka kdanilov2c473092015-03-29 17:12:13 +0300340
koder aka kdanilovda45e882015-04-06 02:24:42 +0300341 if exc is not None:
342 raise exc, cls, tb
koder aka kdanilov2c473092015-03-29 17:12:13 +0300343
koder aka kdanilove06762a2015-03-22 23:32:09 +0200344 return 0
koder aka kdanilov3f356262015-02-13 08:06:14 -0800345
koder aka kdanilov7acd6bd2015-02-12 14:28:30 -0800346
koder aka kdanilov7acd6bd2015-02-12 14:28:30 -0800347if __name__ == '__main__':
koder aka kdanilove06762a2015-03-22 23:32:09 +0200348 exit(main(sys.argv))