blob: 1841bc8b39403bef0d113e3341a9264ec1163625 [file] [log] [blame]
gstepanov023c1e42015-04-08 15:50:19 +03001import os
koder aka kdanilov7acd6bd2015-02-12 14:28:30 -08002import sys
koder aka kdanilovda45e882015-04-06 02:24:42 +03003import json
koder aka kdanilov2c473092015-03-29 17:12:13 +03004import Queue
koder aka kdanilov7acd6bd2015-02-12 14:28:30 -08005import pprint
koder aka kdanilove21d7472015-02-14 19:02:04 -08006import logging
koder aka kdanilov7acd6bd2015-02-12 14:28:30 -08007import argparse
koder aka kdanilov1c2b5112015-04-10 16:53:51 +03008import traceback
koder aka kdanilov2c473092015-03-29 17:12:13 +03009import threading
10import collections
koder aka kdanilov7acd6bd2015-02-12 14:28:30 -080011
koder aka kdanilov66839a92015-04-11 13:22:31 +030012import yaml
koder aka kdanilov2c473092015-03-29 17:12:13 +030013from concurrent.futures import ThreadPoolExecutor
koder aka kdanilov6c491062015-04-09 22:33:13 +030014
koder aka kdanilov2c473092015-03-29 17:12:13 +030015import utils
koder aka kdanilov66839a92015-04-11 13:22:31 +030016import report
koder aka kdanilove06762a2015-03-22 23:32:09 +020017import ssh_utils
koder aka kdanilovda45e882015-04-06 02:24:42 +030018import start_vms
koder aka kdanilov66839a92015-04-11 13:22:31 +030019import pretty_yaml
koder aka kdanilove06762a2015-03-22 23:32:09 +020020from nodes import discover
koder aka kdanilov2c473092015-03-29 17:12:13 +030021from nodes.node import Node
koder aka kdanilov66839a92015-04-11 13:22:31 +030022from config import cfg_dict, load_config
koder aka kdanilovda45e882015-04-06 02:24:42 +030023from tests.itest import IOPerfTest, PgBenchTest
koder aka kdanilov2c473092015-03-29 17:12:13 +030024from sensors.api import start_monitoring
koder aka kdanilov66839a92015-04-11 13:22:31 +030025from formatters import format_results_for_console
koder aka kdanilov2c473092015-03-29 17:12:13 +030026
27
koder aka kdanilove21d7472015-02-14 19:02:04 -080028logger = logging.getLogger("io-perf-tool")
koder aka kdanilove21d7472015-02-14 19:02:04 -080029
30
koder aka kdanilov2c473092015-03-29 17:12:13 +030031def setup_logger(logger, level=logging.DEBUG):
32 logger.setLevel(level)
33 ch = logging.StreamHandler()
34 ch.setLevel(level)
35 logger.addHandler(ch)
Yulia Portnova7ddfa732015-02-24 17:32:58 +020036
koder aka kdanilov2c473092015-03-29 17:12:13 +030037 log_format = '%(asctime)s - %(levelname)-6s - %(name)s - %(message)s'
38 formatter = logging.Formatter(log_format,
39 "%H:%M:%S")
40 ch.setFormatter(formatter)
koder aka kdanilov7acd6bd2015-02-12 14:28:30 -080041
koder aka kdanilov6c491062015-04-09 22:33:13 +030042 # logger.setLevel(logging.INFO)
43 # logger.addHandler(logging.FileHandler('log.txt'))
44
koder aka kdanilov7acd6bd2015-02-12 14:28:30 -080045
Yulia Portnova7ddfa732015-02-24 17:32:58 +020046def format_result(res, formatter):
koder aka kdanilove21d7472015-02-14 19:02:04 -080047 data = "\n{0}\n".format("=" * 80)
48 data += pprint.pformat(res) + "\n"
49 data += "{0}\n".format("=" * 80)
koder aka kdanilovfe056622015-02-19 08:46:15 -080050 templ = "{0}\n\n====> {1}\n\n{2}\n\n"
Yulia Portnova7ddfa732015-02-24 17:32:58 +020051 return templ.format(data, formatter(res), "=" * 80)
koder aka kdanilove21d7472015-02-14 19:02:04 -080052
53
koder aka kdanilov1c2b5112015-04-10 16:53:51 +030054class Context(object):
55 def __init__(self):
56 self.build_meta = {}
57 self.nodes = []
58 self.clear_calls_stack = []
59 self.openstack_nodes_ids = []
60
61
koder aka kdanilov5d589b42015-03-26 12:25:51 +020062def connect_one(node):
63 try:
koder aka kdanilov2c473092015-03-29 17:12:13 +030064 ssh_pref = "ssh://"
65 if node.conn_url.startswith(ssh_pref):
66 url = node.conn_url[len(ssh_pref):]
67 node.connection = ssh_utils.connect(url)
68 else:
69 raise ValueError("Unknown url type {0}".format(node.conn_url))
koder aka kdanilov3a6633e2015-03-26 18:20:00 +020070 except Exception:
koder aka kdanilov2c473092015-03-29 17:12:13 +030071 logger.exception("During connect to {0}".format(node))
koder aka kdanilov5d589b42015-03-26 12:25:51 +020072
73
74def connect_all(nodes):
koder aka kdanilov2c473092015-03-29 17:12:13 +030075 logger.info("Connecting to nodes")
76 with ThreadPoolExecutor(32) as pool:
77 list(pool.map(connect_one, nodes))
koder aka kdanilovda45e882015-04-06 02:24:42 +030078 logger.info("All nodes connected successfully")
koder aka kdanilov2c473092015-03-29 17:12:13 +030079
80
81def save_sensors_data(q):
82 logger.info("Start receiving sensors data")
koder aka kdanilovda45e882015-04-06 02:24:42 +030083 sensor_data = []
koder aka kdanilov2c473092015-03-29 17:12:13 +030084 while True:
85 val = q.get()
86 if val is None:
koder aka kdanilovda45e882015-04-06 02:24:42 +030087 print sensor_data
88 q.put(sensor_data)
koder aka kdanilov2c473092015-03-29 17:12:13 +030089 break
koder aka kdanilovda45e882015-04-06 02:24:42 +030090 sensor_data.append(val)
koder aka kdanilov2c473092015-03-29 17:12:13 +030091 logger.info("Sensors thread exits")
92
93
94def test_thread(test, node, barrier):
95 try:
96 logger.debug("Run preparation for {0}".format(node.conn_url))
97 test.pre_run(node.connection)
98 logger.debug("Run test for {0}".format(node.conn_url))
99 test.run(node.connection, barrier)
100 except:
101 logger.exception("In test {0} for node {1}".format(test, node))
102
103
104def run_tests(config, nodes):
105 tool_type_mapper = {
106 "io": IOPerfTest,
107 "pgbench": PgBenchTest,
108 }
109
110 test_nodes = [node for node in nodes
111 if 'testnode' in node.roles]
112
113 res_q = Queue.Queue()
114
koder aka kdanilovda45e882015-04-06 02:24:42 +0300115 for test in config['tests']:
gstepanov82489e72015-04-10 16:18:03 +0300116 for test in config['tests'][test]['tests']:
gstepanov023c1e42015-04-08 15:50:19 +0300117 for name, params in test.items():
118 logger.info("Starting {0} tests".format(name))
koder aka kdanilov2c473092015-03-29 17:12:13 +0300119
gstepanov023c1e42015-04-08 15:50:19 +0300120 threads = []
121 barrier = utils.Barrier(len(test_nodes))
122 for node in test_nodes:
123 msg = "Starting {0} test on {1} node"
124 logger.debug(msg.format(name, node.conn_url))
125 test = tool_type_mapper[name](params, res_q.put)
126 th = threading.Thread(None, test_thread, None,
127 (test, node, barrier))
128 threads.append(th)
129 th.daemon = True
130 th.start()
koder aka kdanilov2c473092015-03-29 17:12:13 +0300131
gstepanov023c1e42015-04-08 15:50:19 +0300132 for th in threads:
133 th.join()
koder aka kdanilov2c473092015-03-29 17:12:13 +0300134
gstepanov023c1e42015-04-08 15:50:19 +0300135 results = []
koder aka kdanilov66839a92015-04-11 13:22:31 +0300136
gstepanov023c1e42015-04-08 15:50:19 +0300137 while not res_q.empty():
138 results.append(res_q.get())
koder aka kdanilov66839a92015-04-11 13:22:31 +0300139
140 yield name, test.merge_results(results)
koder aka kdanilov2c473092015-03-29 17:12:13 +0300141
142
143def parse_args(argv):
144 parser = argparse.ArgumentParser(
145 description="Run disk io performance test")
146
147 parser.add_argument("-l", dest='extra_logs',
148 action='store_true', default=False,
149 help="print some extra log info")
150
gstepanov4861d712015-04-09 13:28:02 +0300151 parser.add_argument("-b", '--build_description',
152 type=str, default="Build info")
gstepanovaffcdb12015-04-07 17:18:29 +0300153 parser.add_argument("-i", '--build_id', type=str, default="id")
154 parser.add_argument("-t", '--build_type', type=str, default="GA")
155 parser.add_argument("-u", '--username', type=str, default="admin")
koder aka kdanilov66839a92015-04-11 13:22:31 +0300156 parser.add_argument("-p", '--post-process-only', default=None)
koder aka kdanilovda45e882015-04-06 02:24:42 +0300157 parser.add_argument("-o", '--output-dest', nargs="*")
158 parser.add_argument("config_file", nargs="?", default="config.yaml")
koder aka kdanilov2c473092015-03-29 17:12:13 +0300159
160 return parser.parse_args(argv[1:])
161
162
koder aka kdanilovda45e882015-04-06 02:24:42 +0300163def log_nodes_statistic(_, ctx):
164 nodes = ctx.nodes
koder aka kdanilov2c473092015-03-29 17:12:13 +0300165 logger.info("Found {0} nodes total".format(len(nodes)))
166 per_role = collections.defaultdict(lambda: 0)
167 for node in nodes:
168 for role in node.roles:
169 per_role[role] += 1
170
171 for role, count in sorted(per_role.items()):
172 logger.debug("Found {0} nodes with role {1}".format(count, role))
173
174
175def log_sensors_config(cfg):
koder aka kdanilov5d589b42015-03-26 12:25:51 +0200176 pass
177
178
koder aka kdanilovda45e882015-04-06 02:24:42 +0300179def connect_stage(cfg, ctx):
180 ctx.clear_calls_stack.append(disconnect_stage)
181 connect_all(ctx.nodes)
182
183
184def discover_stage(cfg, ctx):
185 if 'discover' in cfg:
186 discover_objs = [i.strip() for i in cfg['discover'].strip().split(",")]
187 ctx.nodes.extend(discover.discover(discover_objs, cfg['clouds']))
188
189 for url, roles in cfg.get('explicit_nodes', {}).items():
190 ctx.nodes.append(Node(url, roles.split(",")))
191
192
193def deploy_sensors_stage(cfg_dict, ctx):
194 ctx.clear_calls_stack.append(remove_sensors_stage)
195 if 'sensors' not in cfg_dict:
196 return
197
198 cfg = cfg_dict.get('sensors')
199 sens_cfg = []
200
201 for role, sensors_str in cfg["roles_mapping"].items():
202 sensors = [sens.strip() for sens in sensors_str.split(",")]
203
204 collect_cfg = dict((sensor, {}) for sensor in sensors)
205
206 for node in ctx.nodes:
207 if role in node.roles:
208 sens_cfg.append((node.connection, collect_cfg))
209
210 log_sensors_config(sens_cfg)
211
212 ctx.sensor_cm = start_monitoring(cfg["receiver_uri"], None,
213 connected_config=sens_cfg)
214
215 ctx.sensors_control_queue = ctx.sensor_cm.__enter__()
216
217 th = threading.Thread(None, save_sensors_data, None,
218 (ctx.sensors_control_queue,))
219 th.daemon = True
220 th.start()
221 ctx.sensor_listen_thread = th
222
223
224def remove_sensors_stage(cfg, ctx):
225 ctx.sensors_control_queue.put(None)
226 ctx.sensor_listen_thread.join()
227 ctx.sensor_data = ctx.sensors_control_queue.get()
228
229
koder aka kdanilov1c2b5112015-04-10 16:53:51 +0300230def run_all_test(cfg, ctx):
koder aka kdanilovda45e882015-04-06 02:24:42 +0300231 ctx.results = []
232
gstepanov023c1e42015-04-08 15:50:19 +0300233 if 'start_test_nodes' in cfg['tests']:
234 params = cfg['tests']['start_test_nodes']['openstack']
koder aka kdanilov1c2b5112015-04-10 16:53:51 +0300235 os_nodes_ids = []
koder aka kdanilov6c491062015-04-09 22:33:13 +0300236
koder aka kdanilov1c2b5112015-04-10 16:53:51 +0300237 os_creds = params['creds']
238
239 if os_creds == 'fuel':
240 raise NotImplementedError()
241
242 elif os_creds == 'clouds':
243 os_cfg = cfg['clouds']['openstack']
244 tenant = os_cfg['OS_TENANT_NAME'].strip()
245 user = os_cfg['OS_USERNAME'].strip()
246 passwd = os_cfg['OS_PASSWORD'].strip()
247 auth_url = os_cfg['OS_AUTH_URL'].strip()
248
249 elif os_creds == 'ENV':
250 tenant = None
251 user = None
252 passwd = None
253 auth_url = None
254
255 else:
256 raise ValueError("Only 'ENV' creds are supported")
257
258 start_vms.nova_connect(user, passwd, tenant, auth_url)
259
260 new_nodes = []
261 for new_node, node_id in start_vms.launch_vms(params):
262 new_node.roles.append('testnode')
263 ctx.nodes.append(new_node)
264 os_nodes_ids.append(node_id)
265 new_nodes.append(new_node)
266
koder aka kdanilov66839a92015-04-11 13:22:31 +0300267 store_nodes_in_log(cfg, os_nodes_ids)
koder aka kdanilov1c2b5112015-04-10 16:53:51 +0300268 ctx.openstack_nodes_ids = os_nodes_ids
269
270 connect_all(new_nodes)
gstepanov023c1e42015-04-08 15:50:19 +0300271
koder aka kdanilovda45e882015-04-06 02:24:42 +0300272 if 'tests' in cfg:
273 ctx.results.extend(run_tests(cfg_dict, ctx.nodes))
274
gstepanov023c1e42015-04-08 15:50:19 +0300275
koder aka kdanilov1c2b5112015-04-10 16:53:51 +0300276def shut_down_vms_stage(cfg, ctx):
koder aka kdanilov66839a92015-04-11 13:22:31 +0300277 vm_ids_fname = cfg_dict['vm_ids_fname']
koder aka kdanilov1c2b5112015-04-10 16:53:51 +0300278 if ctx.openstack_nodes_ids is None:
koder aka kdanilov66839a92015-04-11 13:22:31 +0300279 nodes_ids = open(vm_ids_fname).read().split()
koder aka kdanilov1c2b5112015-04-10 16:53:51 +0300280 else:
281 nodes_ids = ctx.openstack_nodes_ids
282
283 logger.info("Removing nodes")
284 start_vms.clear_nodes(nodes_ids)
285 logger.info("Nodes has been removed")
gstepanov023c1e42015-04-08 15:50:19 +0300286
koder aka kdanilov66839a92015-04-11 13:22:31 +0300287 if os.path.exists(vm_ids_fname):
288 os.remove(vm_ids_fname)
gstepanov023c1e42015-04-08 15:50:19 +0300289
koder aka kdanilov66839a92015-04-11 13:22:31 +0300290
291def store_nodes_in_log(cfg, nodes_ids):
292 with open(cfg['vm_ids_fname'], 'w') as fd:
293 fd.write("\n".join(nodes_ids))
gstepanov023c1e42015-04-08 15:50:19 +0300294
295
296def clear_enviroment(cfg, ctx):
koder aka kdanilov66839a92015-04-11 13:22:31 +0300297 if os.path.exists(cfg_dict['vm_ids_fname']):
koder aka kdanilov1c2b5112015-04-10 16:53:51 +0300298 shut_down_vms_stage(cfg, ctx)
gstepanov023c1e42015-04-08 15:50:19 +0300299
300
301def run_tests_stage(cfg, ctx):
302 # clear nodes that possible were created on previous test running
koder aka kdanilov1c2b5112015-04-10 16:53:51 +0300303 # clear_enviroment(cfg, ctx) << fix OS connection
304 ctx.clear_calls_stack.append(shut_down_vms_stage)
305 run_all_test(cfg, ctx)
koder aka kdanilovda45e882015-04-06 02:24:42 +0300306
307
308def disconnect_stage(cfg, ctx):
309 for node in ctx.nodes:
310 if node.connection is not None:
311 node.connection.close()
312
313
koder aka kdanilov66839a92015-04-11 13:22:31 +0300314def yamable(data):
315 if isinstance(data, (tuple, list)):
316 return map(yamable, data)
317
318 if isinstance(data, unicode):
319 return str(data)
320
321 if isinstance(data, dict):
322 res = {}
323 for k, v in data.items():
324 res[yamable(k)] = yamable(v)
325 return res
326
327 return data
328
329
330def store_raw_results_stage(cfg, ctx):
331
332 raw_results = os.path.join(cfg_dict['var_dir'], 'raw_results.yaml')
333
334 if os.path.exists(raw_results):
335 cont = yaml.load(open(raw_results).read())
336 else:
337 cont = []
338
339 cont.extend(yamable(ctx.results))
340 raw_data = pretty_yaml.dumps(cont)
341
342 with open(raw_results, "w") as fd:
343 fd.write(raw_data)
344
345
346def console_report_stage(cfg, ctx):
347 for tp, data in ctx.results:
348 if 'io' == tp:
349 print format_results_for_console(data)
350
351
koder aka kdanilovda45e882015-04-06 02:24:42 +0300352def report_stage(cfg, ctx):
353 output_dest = cfg.get('output_dest')
koder aka kdanilov1c2b5112015-04-10 16:53:51 +0300354
koder aka kdanilovda45e882015-04-06 02:24:42 +0300355 if output_dest is not None:
Yulia Portnova407ca952015-04-10 10:38:15 +0300356 if output_dest.endswith(".html"):
357 report.render_html_results(ctx, output_dest)
358 logger.info("Results were stored in %s" % output_dest)
359 else:
360 with open(output_dest, "w") as fd:
361 data = {"sensor_data": ctx.sensor_data,
362 "results": ctx.results}
363 fd.write(json.dumps(data))
koder aka kdanilovda45e882015-04-06 02:24:42 +0300364 else:
365 print "=" * 20 + " RESULTS " + "=" * 20
366 pprint.pprint(ctx.results)
367 print "=" * 60
368
369
370def complete_log_nodes_statistic(cfg, ctx):
371 nodes = ctx.nodes
372 for node in nodes:
373 logger.debug(str(node))
374
375
koder aka kdanilov66839a92015-04-11 13:22:31 +0300376def load_data_from(var_dir):
377 def closure(cfg, ctx):
378 raw_results = os.path.join(var_dir, 'raw_results.yaml')
379 ctx.results = yaml.load(open(raw_results).read())
380 return closure
gstepanovcd256d62015-04-07 17:47:32 +0300381
382
koder aka kdanilov3f356262015-02-13 08:06:14 -0800383def main(argv):
koder aka kdanilove06762a2015-03-22 23:32:09 +0200384 opts = parse_args(argv)
koder aka kdanilov2c473092015-03-29 17:12:13 +0300385
koder aka kdanilov66839a92015-04-11 13:22:31 +0300386 if opts.post_process_only is not None:
387 stages = [
388 load_data_from(opts.post_process_only),
389 console_report_stage,
390 # report_stage
391 ]
392 else:
393 stages = [
394 discover_stage,
395 log_nodes_statistic,
396 complete_log_nodes_statistic,
397 connect_stage,
398 complete_log_nodes_statistic,
399 deploy_sensors_stage,
400 run_tests_stage,
401 store_raw_results_stage,
402 console_report_stage,
403 report_stage
404 ]
405
koder aka kdanilov2c473092015-03-29 17:12:13 +0300406 level = logging.DEBUG if opts.extra_logs else logging.WARNING
407 setup_logger(logger, level)
408
gstepanovcd256d62015-04-07 17:47:32 +0300409 load_config(opts.config_file)
koder aka kdanilov66839a92015-04-11 13:22:31 +0300410 logger.info("Store all info into {0}".format(cfg_dict['var_dir']))
gstepanovcd256d62015-04-07 17:47:32 +0300411
koder aka kdanilovda45e882015-04-06 02:24:42 +0300412 ctx = Context()
gstepanovaffcdb12015-04-07 17:18:29 +0300413 ctx.build_meta['build_id'] = opts.build_id
414 ctx.build_meta['build_descrption'] = opts.build_description
415 ctx.build_meta['build_type'] = opts.build_type
416 ctx.build_meta['username'] = opts.username
koder aka kdanilov6c491062015-04-09 22:33:13 +0300417
koder aka kdanilovda45e882015-04-06 02:24:42 +0300418 try:
419 for stage in stages:
420 logger.info("Start {0.__name__} stage".format(stage))
421 stage(cfg_dict, ctx)
422 finally:
423 exc, cls, tb = sys.exc_info()
424 for stage in ctx.clear_calls_stack[::-1]:
425 try:
426 logger.info("Start {0.__name__} stage".format(stage))
427 stage(cfg_dict, ctx)
koder aka kdanilov1c2b5112015-04-10 16:53:51 +0300428 except Exception as exc:
429 logger.exception("During {0.__name__} stage".format(stage))
koder aka kdanilov2c473092015-03-29 17:12:13 +0300430
koder aka kdanilovda45e882015-04-06 02:24:42 +0300431 if exc is not None:
432 raise exc, cls, tb
koder aka kdanilov2c473092015-03-29 17:12:13 +0300433
koder aka kdanilove06762a2015-03-22 23:32:09 +0200434 return 0
koder aka kdanilov3f356262015-02-13 08:06:14 -0800435
koder aka kdanilov7acd6bd2015-02-12 14:28:30 -0800436
koder aka kdanilov7acd6bd2015-02-12 14:28:30 -0800437if __name__ == '__main__':
koder aka kdanilove06762a2015-03-22 23:32:09 +0200438 exit(main(sys.argv))