lot of fixes
diff --git a/results/perf-1-compute-hdd/charts/6fa7da0da5f789f40c0adc40fb125521.png b/results/perf-1-compute-hdd/charts/6fa7da0da5f789f40c0adc40fb125521.png
new file mode 100644
index 0000000..208f28c
--- /dev/null
+++ b/results/perf-1-compute-hdd/charts/6fa7da0da5f789f40c0adc40fb125521.png
Binary files differ
diff --git a/results/perf-1-compute-hdd/charts/9cfbfff2077f35018f164fa9cb4cc9f7.png b/results/perf-1-compute-hdd/charts/9cfbfff2077f35018f164fa9cb4cc9f7.png
new file mode 100644
index 0000000..3cf19a8
--- /dev/null
+++ b/results/perf-1-compute-hdd/charts/9cfbfff2077f35018f164fa9cb4cc9f7.png
Binary files differ
diff --git a/results/perf-1-compute-hdd/log.txt b/results/perf-1-compute-hdd/log.txt
new file mode 100644
index 0000000..4eb4324
--- /dev/null
+++ b/results/perf-1-compute-hdd/log.txt
@@ -0,0 +1,109 @@
+12:30:22 - INFO - io-perf-tool - All info would be stored into /tmp/perf_tests/rotund_elfriede
+12:30:22 - INFO - io-perf-tool - Start discover_stage stage
+12:30:22 - INFO - io-perf-tool - Start log_nodes_statistic stage
+12:30:22 - INFO - io-perf-tool - Found 1 nodes total
+12:30:22 - [1;34mDEBUG[0m - io-perf-tool - Found 1 nodes with role testnode
+12:30:22 - INFO - io-perf-tool - Start connect_stage stage
+12:30:22 - INFO - io-perf-tool - Connecting to nodes
+12:30:22 - INFO - io-perf-tool - All nodes connected successfully
+12:30:22 - INFO - io-perf-tool - Start deploy_sensors_stage stage
+12:30:22 - INFO - io-perf-tool - Start run_tests_stage stage
+12:30:22 - INFO - io-perf-tool - Starting io tests
+12:30:22 - [1;34mDEBUG[0m - io-perf-tool - Starting io test on ssh://root@172.16.52.112:3022:/home/koder/.ssh/id_rsa node
+12:30:22 - [1;34mDEBUG[0m - io-perf-tool - Run preparation for ssh://root@172.16.52.112:3022:/home/koder/.ssh/id_rsa
+12:30:22 - [1;34mDEBUG[0m - io-perf-tool - SSH: Exec 'which fio'
+12:30:23 - [1;34mDEBUG[0m - io-perf-tool - SSH: Exec 'dd if=/dev/zero of=/opt/xxx.bin bs=1048576 count=10240'
+12:31:26 - [1;34mDEBUG[0m - io-perf-tool - Run test for ssh://root@172.16.52.112:3022:/home/koder/.ssh/id_rsa
+12:31:26 - [1;34mDEBUG[0m - io-perf-tool - Waiting on barrier
+12:31:26 - INFO - io-perf-tool - Test will takes about 2:12:35
+12:31:27 - [1;34mDEBUG[0m - io-perf-tool - SSH: Exec 'env python2 /tmp/disk_test_agent.py --type fio --params NUM_ROUNDS=7 FILENAME=/opt/xxx.bin --json -'
+14:46:24 - INFO - io-perf-tool - Start store_raw_results_stage stage
+14:46:24 - INFO - io-perf-tool - Start console_report_stage stage
+14:46:24 - INFO - io-perf-tool - Start report_stage stage
+14:46:24 - INFO - io-perf-tool - Text report were stored in /tmp/perf_tests/rotund_elfriede/report.txt
+14:46:24 - INFO - io-perf-tool - Start shut_down_vms_stage stage
+14:46:24 - INFO - io-perf-tool - Start disconnect_stage stage
+14:46:24 - INFO - io-perf-tool - All info stotored into /tmp/perf_tests/rotund_elfriede
+17:43:04 - INFO - io-perf-tool - All info would be stored into /tmp/perf_tests/rotund_elfriede
+17:43:04 - INFO - io-perf-tool - Start load_data_from_file stage
+17:43:04 - INFO - io-perf-tool - Start console_report_stage stage
+17:43:04 - INFO - io-perf-tool - Start report_stage stage
+17:43:37 - INFO - io-perf-tool - All info would be stored into /tmp/perf_tests/rotund_elfriede
+17:43:37 - INFO - io-perf-tool - Start load_data_from_file stage
+17:43:37 - INFO - io-perf-tool - Start console_report_stage stage
+17:43:37 - INFO - io-perf-tool - Start report_stage stage
+17:43:37 - INFO - io-perf-tool - Html report were stored in /tmp/perf_tests/rotund_elfriede/report.html
+17:43:37 - INFO - io-perf-tool - Text report were stored in /tmp/perf_tests/rotund_elfriede/report.txt
+17:43:37 - INFO - io-perf-tool - All info stotored into /tmp/perf_tests/rotund_elfriede
+17:46:54 - INFO - io-perf-tool - All info would be stored into /tmp/perf_tests/rotund_elfriede
+17:46:54 - INFO - io-perf-tool - Start load_data_from_file stage
+17:46:54 - INFO - io-perf-tool - Start report_stage stage
+17:46:54 - INFO - io-perf-tool - Html report were stored in /tmp/perf_tests/rotund_elfriede/report.html
+17:46:54 - INFO - io-perf-tool - Text report were stored in /tmp/perf_tests/rotund_elfriede/report.txt
+17:46:54 - INFO - io-perf-tool - All info stotored into /tmp/perf_tests/rotund_elfriede
+17:47:54 - INFO - io-perf-tool - All info would be stored into /tmp/perf_tests/rotund_elfriede
+17:47:54 - INFO - io-perf-tool - Start load_data_from_file stage
+17:47:54 - INFO - io-perf-tool - Start report_stage stage
+17:47:54 - INFO - io-perf-tool - Html report were stored in /tmp/perf_tests/rotund_elfriede/report.html
+17:47:54 - INFO - io-perf-tool - Text report were stored in /tmp/perf_tests/rotund_elfriede/report.txt
+17:47:54 - INFO - io-perf-tool - All info stotored into /tmp/perf_tests/rotund_elfriede
+17:49:42 - INFO - io-perf-tool - All info would be stored into /tmp/perf_tests/rotund_elfriede
+17:49:42 - INFO - io-perf-tool - Start load_data_from_file stage
+17:49:42 - INFO - io-perf-tool - Start report_stage stage
+17:49:42 - INFO - io-perf-tool - Html report were stored in /tmp/perf_tests/rotund_elfriede/report.html
+17:49:42 - INFO - io-perf-tool - Text report were stored in /tmp/perf_tests/rotund_elfriede/report.txt
+17:49:42 - INFO - io-perf-tool - All info stotored into /tmp/perf_tests/rotund_elfriede
+17:50:59 - INFO - io-perf-tool - All info would be stored into /tmp/perf_tests/rotund_elfriede
+17:50:59 - INFO - io-perf-tool - Start load_data_from_file stage
+17:50:59 - INFO - io-perf-tool - Start report_stage stage
+17:50:59 - INFO - io-perf-tool - Html report were stored in /tmp/perf_tests/rotund_elfriede/report.html
+17:50:59 - INFO - io-perf-tool - Text report were stored in /tmp/perf_tests/rotund_elfriede/report.txt
+17:50:59 - INFO - io-perf-tool - All info stotored into /tmp/perf_tests/rotund_elfriede
+17:52:01 - INFO - io-perf-tool - All info would be stored into /tmp/perf_tests/rotund_elfriede
+17:52:01 - INFO - io-perf-tool - Start load_data_from_file stage
+17:52:01 - INFO - io-perf-tool - Start report_stage stage
+17:52:01 - INFO - io-perf-tool - Html report were stored in /tmp/perf_tests/rotund_elfriede/report.html
+17:52:01 - INFO - io-perf-tool - Text report were stored in /tmp/perf_tests/rotund_elfriede/report.txt
+17:52:01 - INFO - io-perf-tool - All info stotored into /tmp/perf_tests/rotund_elfriede
+17:52:37 - INFO - io-perf-tool - All info would be stored into /tmp/perf_tests/rotund_elfriede
+17:52:37 - INFO - io-perf-tool - Start load_data_from_file stage
+17:52:37 - INFO - io-perf-tool - Start report_stage stage
+17:52:37 - INFO - io-perf-tool - Html report were stored in /tmp/perf_tests/rotund_elfriede/report.html
+17:52:37 - INFO - io-perf-tool - Text report were stored in /tmp/perf_tests/rotund_elfriede/report.txt
+17:52:37 - INFO - io-perf-tool - All info stotored into /tmp/perf_tests/rotund_elfriede
+18:45:15 - INFO - io-perf-tool - All info would be stored into /tmp/perf_tests/rotund_elfriede
+18:45:15 - INFO - io-perf-tool - Start load_data_from_file stage
+18:45:15 - INFO - io-perf-tool - Start report_stage stage
+18:45:15 - INFO - io-perf-tool - Html report were stored in /tmp/perf_tests/rotund_elfriede/report.html
+18:45:15 - INFO - io-perf-tool - Text report were stored in /tmp/perf_tests/rotund_elfriede/report.txt
+18:45:15 - INFO - io-perf-tool - All info stotored into /tmp/perf_tests/rotund_elfriede
+18:45:19 - INFO - io-perf-tool - All info would be stored into /tmp/perf_tests/rotund_elfriede
+18:45:19 - INFO - io-perf-tool - Start load_data_from_file stage
+18:45:20 - INFO - io-perf-tool - Start report_stage stage
+18:45:20 - INFO - io-perf-tool - Html report were stored in /tmp/perf_tests/rotund_elfriede/report.html
+18:45:20 - INFO - io-perf-tool - Text report were stored in /tmp/perf_tests/rotund_elfriede/report.txt
+18:45:20 - INFO - io-perf-tool - All info stotored into /tmp/perf_tests/rotund_elfriede
+18:46:01 - INFO - io-perf-tool - All info would be stored into /tmp/perf_tests/rotund_elfriede
+18:46:01 - INFO - io-perf-tool - Start load_data_from_file stage
+18:46:01 - INFO - io-perf-tool - Start report_stage stage
+18:46:02 - INFO - io-perf-tool - Html report were stored in /tmp/perf_tests/rotund_elfriede/report.html
+18:46:02 - INFO - io-perf-tool - Text report were stored in /tmp/perf_tests/rotund_elfriede/report.txt
+18:46:02 - INFO - io-perf-tool - All info stotored into /tmp/perf_tests/rotund_elfriede
+18:48:51 - INFO - io-perf-tool - All info would be stored into /tmp/perf_tests/rotund_elfriede
+18:48:51 - INFO - io-perf-tool - Start load_data_from_file stage
+18:48:51 - INFO - io-perf-tool - Start report_stage stage
+18:48:52 - INFO - io-perf-tool - Html report were stored in /tmp/perf_tests/rotund_elfriede/report.html
+18:48:52 - INFO - io-perf-tool - Text report were stored in /tmp/perf_tests/rotund_elfriede/report.txt
+18:48:52 - INFO - io-perf-tool - All info stotored into /tmp/perf_tests/rotund_elfriede
+18:49:47 - INFO - io-perf-tool - All info would be stored into /tmp/perf_tests/rotund_elfriede
+18:49:47 - INFO - io-perf-tool - Start load_data_from_file stage
+18:49:47 - INFO - io-perf-tool - Start report_stage stage
+18:49:47 - INFO - io-perf-tool - Html report were stored in /tmp/perf_tests/rotund_elfriede/report.html
+18:49:47 - INFO - io-perf-tool - Text report were stored in /tmp/perf_tests/rotund_elfriede/report.txt
+18:49:47 - INFO - io-perf-tool - All info stotored into /tmp/perf_tests/rotund_elfriede
+18:50:09 - INFO - io-perf-tool - All info would be stored into /tmp/perf_tests/rotund_elfriede
+18:50:09 - INFO - io-perf-tool - Start load_data_from_file stage
+18:50:09 - INFO - io-perf-tool - Start report_stage stage
+18:50:10 - INFO - io-perf-tool - Html report were stored in /tmp/perf_tests/rotund_elfriede/report.html
+18:50:10 - INFO - io-perf-tool - Text report were stored in /tmp/perf_tests/rotund_elfriede/report.txt
+18:50:10 - INFO - io-perf-tool - All info stotored into /tmp/perf_tests/rotund_elfriede
diff --git a/results/perf-1-compute-hdd/raw_results.yaml b/results/perf-1-compute-hdd/raw_results.yaml
new file mode 100644
index 0000000..6a6b4f6
--- /dev/null
+++ b/results/perf-1-compute-hdd/raw_results.yaml
@@ -0,0 +1,452 @@
+- - io
+ - res:
+ hdd_test_swd1mth40:
+ rw: write
+ blocksize: 1m
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 40
+ timings: [30, 5]
+ jobname: hdd_test_swd1mth40
+ iops: [160, 174, 172, 171, 167, 172, 169]
+ bw: [163850, 178577, 176845, 175673, 171042, 176273, 173693]
+ lat: [249262.59, 226253.83, 228928.36, 232426.25, 235542.05, 228821.3, 229864.66]
+ clat: [249221.16, 226213.61, 228885.42, 232385.36, 235498.86, 228780.83, 229823.97]
+ hdd_test_srd1mth120:
+ rw: read
+ blocksize: 1m
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 120
+ timings: [30, 5]
+ jobname: hdd_test_srd1mth120
+ iops: [438, 473, 421, 469, 477, 451, 474]
+ bw: [448702, 485077, 431596, 481014, 488770, 462393, 485464]
+ lat: [272341.19, 252221.05, 283448.97, 254310.42, 250329.45, 264582.62, 252028.58]
+ clat: [272341.0, 252220.88, 283448.78, 254310.25, 250329.27, 264582.44, 252028.41]
+ hdd_test_rwd4kth1:
+ rw: randwrite
+ blocksize: 4k
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 1
+ timings: [30, 5]
+ jobname: hdd_test_rwd4kth1
+ iops: [404, 406, 405, 410, 404, 407, 408]
+ bw: [1616, 1624, 1623, 1640, 1617, 1630, 1632]
+ lat: [2470.42, 2457.72, 2459.19, 2433.45, 2469.15, 2449.21, 2446.34]
+ clat: [2470.1, 2457.38, 2458.88, 2433.11, 2468.82, 2448.89, 2446.02]
+ hdd_test_rrd4kth5:
+ rw: randread
+ blocksize: 4k
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 5
+ timings: [30, 5]
+ jobname: hdd_test_rrd4kth5
+ iops: [262, 287, 277, 292, 279, 290, 270]
+ bw: [1048, 1149, 1108, 1170, 1119, 1161, 1082]
+ lat: [19059.87, 17386.08, 18032.29, 17070.71, 17856.45, 17204.53, 18461.88]
+ clat: [19059.61, 17385.79, 18031.99, 17070.41, 17856.17, 17204.21, 18461.62]
+ hdd_test_rws4kth5:
+ rw: randwrite
+ blocksize: 4k
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: s
+ concurence: 5
+ timings: [30, 5]
+ jobname: hdd_test_rws4kth5
+ iops: [71, 74, 71, 73, 74, 73, 73]
+ bw: [285, 297, 287, 295, 296, 293, 295]
+ lat: [69964.5, 67239.55, 69604.95, 67606.77, 67311.6, 68136.38, 67596.02]
+ clat: [69964.09, 67239.18, 69604.56, 67606.38, 67311.23, 68136.01, 67595.64]
+ hdd_test_swd1mth1:
+ rw: write
+ blocksize: 1m
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 1
+ timings: [30, 5]
+ jobname: hdd_test_swd1mth1
+ iops: [177, 174, 173, 177, 175, 173, 173]
+ bw: [181320, 178232, 177891, 181685, 179892, 177816, 177999]
+ lat: [5644.35, 5742.4, 5753.36, 5633.08, 5689.31, 5755.67, 5749.82]
+ clat: [5629.54, 5727.65, 5738.58, 5618.43, 5674.7, 5740.86, 5735.11]
+ hdd_test_srd1mth80:
+ rw: read
+ blocksize: 1m
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 80
+ timings: [30, 5]
+ jobname: hdd_test_srd1mth80
+ iops: [488, 481, 474, 468, 489, 481, 489]
+ bw: [500256, 493161, 485776, 480072, 501654, 492882, 501021]
+ lat: [163227.78, 165575.86, 168155.67, 170139.34, 162830.11, 165750.03, 163235.84]
+ clat: [163227.61, 165575.69, 168155.52, 170139.17, 162829.92, 165749.88, 163235.67]
+ hdd_test_swd1mth120:
+ rw: write
+ blocksize: 1m
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 120
+ timings: [30, 5]
+ jobname: hdd_test_swd1mth120
+ iops: [201, 211, 205, 218, 200, 211, 216]
+ bw: [206004, 216334, 210607, 223384, 205479, 216865, 221524]
+ lat: [592248.31, 531388.81, 550845.62, 507271.84, 564501.19, 517807.72, 532421.38]
+ clat: [592203.56, 531346.5, 550801.75, 507227.88, 564458.19, 517765.69, 532375.44]
+ hdd_test_swd1mth5:
+ rw: write
+ blocksize: 1m
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 5
+ timings: [30, 5]
+ jobname: hdd_test_swd1mth5
+ iops: [151, 145, 152, 148, 151, 143, 152]
+ bw: [155361, 148761, 155975, 152010, 155283, 146807, 155969]
+ lat: [32873.79, 34407.42, 32816.2, 33670.63, 32944.83, 34865.57, 32814.71]
+ clat: [32857.0, 34389.8, 32799.38, 33653.61, 32928.36, 34848.3, 32798.34]
+ hdd_test_rws4kth15:
+ rw: randwrite
+ blocksize: 4k
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: s
+ concurence: 15
+ timings: [30, 5]
+ jobname: hdd_test_rws4kth15
+ iops: [80, 78, 83, 76, 82, 82, 78]
+ bw: [321, 312, 333, 307, 331, 329, 312]
+ lat: [186285.27, 190777.61, 179656.09, 194278.34, 180218.91, 181491.55, 191513.5]
+ clat: [186284.88, 190777.22, 179655.69, 194277.94, 180218.5, 181491.14, 191513.09]
+ hdd_test_rws4kth30:
+ rw: randwrite
+ blocksize: 4k
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: s
+ concurence: 30
+ timings: [30, 5]
+ jobname: hdd_test_rws4kth30
+ iops: [82, 77, 81, 82, 83, 76, 82]
+ bw: [328, 308, 326, 329, 333, 306, 329]
+ lat: [363845.31, 386380.12, 364771.66, 361253.5, 357521.56, 388269.97, 361152.19]
+ clat: [363844.91, 386379.72, 364771.25, 361253.06, 357521.12, 388269.56, 361151.78]
+ hdd_test_rws4kth10:
+ rw: randwrite
+ blocksize: 4k
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: s
+ concurence: 10
+ timings: [30, 5]
+ jobname: hdd_test_rws4kth10
+ iops: [82, 77, 83, 77, 82, 78, 83]
+ bw: [330, 311, 332, 309, 331, 313, 333]
+ lat: [120939.94, 128316.23, 120108.23, 128817.77, 120450.09, 127472.73, 119743.85]
+ clat: [120939.52, 128315.8, 120107.84, 128817.36, 120449.7, 127472.34, 119743.43]
+ hdd_test_srd1mth15:
+ rw: read
+ blocksize: 1m
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 15
+ timings: [30, 5]
+ jobname: hdd_test_srd1mth15
+ iops: [192, 176, 172, 181, 186, 186, 186]
+ bw: [197618, 180275, 177117, 185532, 190677, 190988, 190529]
+ lat: [77377.37, 85025.45, 86539.22, 82610.98, 80411.26, 80263.92, 80454.14]
+ clat: [77377.22, 85025.3, 86539.05, 82610.83, 80411.09, 80263.76, 80453.98]
+ hdd_test_srd1mth30:
+ rw: read
+ blocksize: 1m
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 30
+ timings: [30, 5]
+ jobname: hdd_test_srd1mth30
+ iops: [375, 421, 423, 325, 346, 405, 420]
+ bw: [384467, 431356, 433528, 333485, 355080, 414940, 431082]
+ lat: [79710.12, 71103.94, 70713.8, 92025.48, 86380.95, 73954.95, 71141.71]
+ clat: [79709.97, 71103.77, 70713.64, 92025.31, 86380.8, 73954.78, 71141.55]
+ hdd_test_srd1mth10:
+ rw: read
+ blocksize: 1m
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 10
+ timings: [30, 5]
+ jobname: hdd_test_srd1mth10
+ iops: [157, 174, 168, 166, 167, 175, 182]
+ bw: [161732, 178381, 172617, 170814, 171364, 179492, 186602]
+ lat: [63202.16, 57316.83, 59239.18, 59862.51, 59670.26, 56965.62, 54803.95]
+ clat: [63201.98, 57316.68, 59239.02, 59862.35, 59670.09, 56965.46, 54803.78]
+ hdd_test_rrd4kth30:
+ rw: randread
+ blocksize: 4k
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 30
+ timings: [30, 5]
+ jobname: hdd_test_rrd4kth30
+ iops: [469, 463, 462, 457, 460, 460, 457]
+ bw: [1879, 1855, 1850, 1828, 1843, 1842, 1828]
+ lat: [63604.93, 64526.77, 64710.67, 65465.62, 64910.79, 64978.81, 65426.57]
+ clat: [63604.78, 64526.62, 64710.52, 65465.47, 64910.63, 64978.67, 65426.43]
+ hdd_test_rrd4kth10:
+ rw: randread
+ blocksize: 4k
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 10
+ timings: [30, 5]
+ jobname: hdd_test_rrd4kth10
+ iops: [353, 353, 354, 358, 356, 349, 353]
+ bw: [1415, 1413, 1419, 1433, 1424, 1397, 1413]
+ lat: [28220.8, 28267.07, 28136.39, 27866.14, 28059.99, 28568.01, 28278.55]
+ clat: [28220.62, 28266.85, 28136.17, 27865.95, 28059.8, 28567.81, 28278.35]
+ hdd_test_rrd4kth15:
+ rw: randread
+ blocksize: 4k
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 15
+ timings: [30, 5]
+ jobname: hdd_test_rrd4kth15
+ iops: [395, 390, 394, 393, 399, 389, 392]
+ bw: [1582, 1560, 1576, 1574, 1599, 1559, 1570]
+ lat: [37844.47, 38370.14, 38009.38, 38071.2, 37473.45, 38400.98, 38149.96]
+ clat: [37844.32, 38369.99, 38009.23, 38071.05, 37473.29, 38400.84, 38149.82]
+ hdd_test_swd1mth20:
+ rw: write
+ blocksize: 1m
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 20
+ timings: [30, 5]
+ jobname: hdd_test_swd1mth20
+ iops: [151, 149, 154, 150, 155, 149, 150]
+ bw: [155326, 153496, 157881, 153975, 159391, 152655, 153698]
+ lat: [131614.89, 133148.73, 129457.04, 132566.56, 128049.26, 134008.36, 132869.22]
+ clat: [131585.36, 133118.5, 129427.01, 132534.31, 128020.7, 133976.36, 132838.5]
+ hdd_test_rws4kth120:
+ rw: randwrite
+ blocksize: 4k
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: s
+ concurence: 120
+ timings: [30, 5]
+ jobname: hdd_test_rws4kth120
+ iops: [79, 82, 81, 83, 77, 82, 77]
+ bw: [319, 328, 324, 332, 311, 329, 311]
+ lat: [1491354.75, 1415709.5, 1426159.62, 1390980.88, 1489516.88, 1403286.62, 1492233.62]
+ clat: [1491354.38, 1415709.12, 1426159.12, 1390980.5, 1489516.38, 1403286.12, 1492233.25]
+ hdd_test_rrd4kth120:
+ rw: randread
+ blocksize: 4k
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 120
+ timings: [30, 5]
+ jobname: hdd_test_rrd4kth120
+ iops: [478, 472, 484, 485, 484, 474, 488]
+ bw: [1913, 1889, 1938, 1941, 1936, 1898, 1952]
+ lat: [246951.8, 250621.23, 244948.47, 244863.33, 245347.81, 249996.78, 243654.67]
+ clat: [246951.64, 250621.08, 244948.31, 244863.19, 245347.66, 249996.62, 243654.53]
+ hdd_test_swd1mth15:
+ rw: write
+ blocksize: 1m
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 15
+ timings: [30, 5]
+ jobname: hdd_test_swd1mth15
+ iops: [149, 155, 149, 151, 153, 152, 148]
+ bw: [153004, 159443, 152738, 155260, 157541, 156277, 151712]
+ lat: [99993.16, 96187.08, 100372.34, 98703.3, 97320.62, 97898.12, 100343.73]
+ clat: [99967.08, 96163.2, 100345.82, 98680.35, 97293.82, 97873.88, 100317.48]
+ hdd_test_srd1mth1:
+ rw: read
+ blocksize: 1m
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 1
+ timings: [30, 5]
+ jobname: hdd_test_srd1mth1
+ iops: [180, 180, 181, 177, 180, 180, 180]
+ bw: [184751, 184931, 185614, 181816, 185235, 184429, 185272]
+ lat: [5539.57, 5534.02, 5513.71, 5628.99, 5524.93, 5549.08, 5523.87]
+ clat: [5539.29, 5533.76, 5513.46, 5628.72, 5524.68, 5548.82, 5523.62]
+ hdd_test_srd1mth5:
+ rw: read
+ blocksize: 1m
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 5
+ timings: [30, 5]
+ jobname: hdd_test_srd1mth5
+ iops: [137, 165, 154, 132, 141, 167, 159]
+ bw: [140534, 169331, 158458, 135958, 144873, 171172, 163514]
+ lat: [36426.63, 30212.38, 32287.75, 37629.91, 35317.55, 29885.17, 31288.82]
+ clat: [36426.43, 30212.16, 32287.56, 37629.69, 35317.33, 29884.99, 31288.62]
+ hdd_test_swd1mth80:
+ rw: write
+ blocksize: 1m
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 80
+ timings: [30, 5]
+ jobname: hdd_test_swd1mth80
+ iops: [203, 196, 202, 194, 188, 190, 194]
+ bw: [208326, 201104, 206881, 198828, 192591, 194757, 199225]
+ lat: [390070.41, 395894.09, 386283.0, 391240.69, 405315.09, 395800.53, 393988.16]
+ clat: [390023.34, 395848.88, 386237.84, 391197.75, 405272.56, 395758.59, 393945.19]
+ hdd_test_rrd4kth80:
+ rw: randread
+ blocksize: 4k
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 80
+ timings: [30, 5]
+ jobname: hdd_test_rrd4kth80
+ iops: [474, 478, 476, 469, 482, 477, 475]
+ bw: [1896, 1913, 1906, 1876, 1928, 1909, 1900]
+ lat: [166645.81, 165584.61, 166682.8, 169626.28, 164584.86, 166233.72, 167094.14]
+ clat: [166645.67, 165584.48, 166682.64, 169626.14, 164584.7, 166233.56, 167094.0]
+ hdd_test_rws4kth20:
+ rw: randwrite
+ blocksize: 4k
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: s
+ concurence: 20
+ timings: [30, 5]
+ jobname: hdd_test_rws4kth20
+ iops: [77, 82, 79, 82, 81, 82, 78]
+ bw: [311, 329, 319, 329, 326, 330, 312]
+ lat: [255892.94, 241595.78, 249261.97, 241525.91, 243757.88, 240959.8, 254951.75]
+ clat: [255892.52, 241595.38, 249261.56, 241525.5, 243757.47, 240959.38, 254951.34]
+ hdd_test_srd1mth20:
+ rw: read
+ blocksize: 1m
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 20
+ timings: [30, 5]
+ jobname: hdd_test_srd1mth20
+ iops: [286, 278, 252, 301, 266, 270, 253]
+ bw: [292956, 285286, 258565, 308458, 272386, 276708, 259966]
+ lat: [69842.21, 71653.4, 79006.03, 66292.67, 75043.28, 73879.11, 78661.34]
+ clat: [69842.04, 71653.23, 79005.87, 66292.51, 75043.12, 73878.95, 78661.18]
+ hdd_test_rrd4kth40:
+ rw: randread
+ blocksize: 4k
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 40
+ timings: [30, 5]
+ jobname: hdd_test_rrd4kth40
+ iops: [462, 463, 465, 457, 456, 460, 457]
+ bw: [1851, 1854, 1861, 1828, 1827, 1843, 1828]
+ lat: [85900.73, 86016.24, 85645.2, 87243.56, 87067.41, 86435.66, 87320.75]
+ clat: [85900.58, 86016.09, 85645.03, 87243.41, 87067.26, 86435.52, 87320.59]
+ hdd_test_rrd4kth20:
+ rw: randread
+ blocksize: 4k
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 20
+ timings: [30, 5]
+ jobname: hdd_test_rrd4kth20
+ iops: [419, 421, 429, 425, 423, 427, 424]
+ bw: [1679, 1685, 1716, 1702, 1692, 1709, 1697]
+ lat: [47528.56, 47403.04, 46551.27, 46879.47, 47227.83, 46707.35, 47008.94]
+ clat: [47528.41, 47402.89, 46551.11, 46879.32, 47227.68, 46707.2, 47008.79]
+ hdd_test_rws4kth80:
+ rw: randwrite
+ blocksize: 4k
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: s
+ concurence: 80
+ timings: [30, 5]
+ jobname: hdd_test_rws4kth80
+ iops: [75, 82, 78, 81, 82, 81, 76]
+ bw: [302, 329, 314, 327, 329, 325, 307]
+ lat: [1050249.5, 947348.38, 994411.31, 950421.94, 947885.5, 961421.5, 1013113.69]
+ clat: [1050249.12, 947347.94, 994410.88, 950421.5, 947885.0, 961421.06, 1013113.25]
+ hdd_test_rws4kth1:
+ rw: randwrite
+ blocksize: 4k
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: s
+ concurence: 1
+ timings: [30, 5]
+ jobname: hdd_test_rws4kth1
+ iops: [41, 41, 41, 41, 42, 41, 41]
+ bw: [166, 165, 167, 164, 168, 165, 167]
+ lat: [23983.87, 24110.1, 23813.04, 24365.01, 23743.05, 24174.92, 23816.97]
+ clat: [23983.49, 24109.79, 23812.69, 24364.66, 23742.73, 24174.54, 23816.61]
+ hdd_test_rws4kth40:
+ rw: randwrite
+ blocksize: 4k
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: s
+ concurence: 40
+ timings: [30, 5]
+ jobname: hdd_test_rws4kth40
+ iops: [78, 78, 83, 82, 77, 77, 83]
+ bw: [313, 312, 332, 331, 309, 311, 333]
+ lat: [506244.72, 507070.97, 476336.69, 476645.34, 512317.62, 507324.06, 473760.31]
+ clat: [506244.34, 507070.56, 476336.31, 476644.94, 512317.22, 507323.59, 473759.91]
+ hdd_test_rrd4kth1:
+ rw: randread
+ blocksize: 4k
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 1
+ timings: [30, 5]
+ jobname: hdd_test_rrd4kth1
+ iops: [143, 157, 149, 157, 152, 157, 155]
+ bw: [575, 629, 597, 631, 611, 628, 620]
+ lat: [6950.36, 6351.87, 6695.59, 6330.2, 6534.99, 6359.04, 6439.06]
+ clat: [6950.11, 6351.6, 6695.33, 6329.94, 6534.73, 6358.75, 6438.8]
+ hdd_test_swd1mth30:
+ rw: write
+ blocksize: 1m
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 30
+ timings: [30, 5]
+ jobname: hdd_test_swd1mth30
+ iops: [142, 155, 160, 167, 163, 157, 164]
+ bw: [145482, 158842, 164228, 171272, 167495, 161213, 168625]
+ lat: [210477.3, 191831.08, 185666.44, 178756.48, 181382.86, 187686.77, 181725.97]
+ clat: [210439.06, 191794.2, 185629.3, 178718.39, 181345.3, 187648.77, 181689.28]
+ hdd_test_swd1mth10:
+ rw: write
+ blocksize: 1m
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 10
+ timings: [30, 5]
+ jobname: hdd_test_swd1mth10
+ iops: [149, 154, 147, 150, 153, 148, 151]
+ bw: [152757, 158453, 150542, 153884, 156809, 152038, 155536]
+ lat: [66851.63, 64582.15, 67974.72, 66451.34, 65271.57, 67256.13, 65793.96]
+ clat: [66831.3, 64562.43, 67954.29, 66430.45, 65252.41, 67235.51, 65774.64]
+ hdd_test_srd1mth40:
+ rw: read
+ blocksize: 1m
+ slat: [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ sync_mode: d
+ concurence: 40
+ timings: [30, 5]
+ jobname: hdd_test_srd1mth40
+ iops: [458, 467, 424, 440, 457, 477, 447]
+ bw: [469523, 478406, 435021, 451389, 468090, 488883, 458133]
+ lat: [87134.27, 85494.91, 94015.34, 90609.92, 87388.05, 83693.32, 89268.15]
+ clat: [87134.11, 85494.76, 94015.19, 90609.77, 87387.9, 83693.16, 89267.99]
+ __meta__:
+ raw_cfg:
+ '[defaults]\nwait_for_previous\ngroup_reporting\ntime_based\nbuffered=0\niodepth=1\n\nfilename={FILENAME}\nNUM_ROUNDS=7\n\nramp_time=5\nsize=10Gb\nruntime=30\n\n# ---------------------------------------------------------------------\n# check different thread count, sync mode. (latency, iops) = func(th_count)\n# ---------------------------------------------------------------------\n[hdd_test_{TEST_SUMM} * {NUM_ROUNDS}]\nblocksize=4k\nrw=randwrite\nsync=1\nnumjobs={% 1, 5, 10, 15, 20, 30, 40, 80, 120 %}\n\n# ---------------------------------------------------------------------\n# check different thread count, direct read mode. (latency, iops) = func(th_count)\n# also check iops for randread\n# ---------------------------------------------------------------------\n[hdd_test_{TEST_SUMM} * {NUM_ROUNDS}]\nblocksize=4k\nrw=randread\ndirect=1\nnumjobs={% 1, 5, 10, 15, 20, 30, 40, 80, 120 %}\n\n# ---------------------------------------------------------------------\n# check different thread count, direct read/write mode. (bw, iops) = func(th_count)\n# also check BW for seq read/write.\n# ---------------------------------------------------------------------\n[hdd_test_{TEST_SUMM} * {NUM_ROUNDS}]\nblocksize=1m\nrw={% read, write %}\ndirect=1\nnumjobs={% 1, 5, 10, 15, 20, 30, 40, 80, 120 %}\n\n# ---------------------------------------------------------------------\n# check IOPS randwrite.\n# ---------------------------------------------------------------------\n[hdd_test_{TEST_SUMM} * {NUM_ROUNDS}]\nblocksize=4k\nrw=randwrite\ndirect=1\n'
+ params:
+ NUM_ROUNDS: 7
+ FILENAME: /opt/xxx.bin
\ No newline at end of file
diff --git a/results/perf-1-compute-hdd/report.html b/results/perf-1-compute-hdd/report.html
new file mode 100644
index 0000000..ef326eb
--- /dev/null
+++ b/results/perf-1-compute-hdd/report.html
@@ -0,0 +1,12 @@
+<!DOCTYPE html>
+<html>
+<head>
+ <title>Report</title>
+</head>
+
+<body>
+<div><ol><li><img src='charts/6fa7da0da5f789f40c0adc40fb125521.png'></li>
+<li><img src='charts/9cfbfff2077f35018f164fa9cb4cc9f7.png'></li></ol></div>
+</body>
+
+</html>
\ No newline at end of file
diff --git a/results/perf-1-compute-hdd/report.txt b/results/perf-1-compute-hdd/report.txt
new file mode 100644
index 0000000..9d4992f
--- /dev/null
+++ b/results/perf-1-compute-hdd/report.txt
@@ -0,0 +1,45 @@
++-------------+------+---------+-----------+--------+
+| Description | IOPS | BW KBps | Dev * 3 % | LAT ms |
++=============+======+=========+===========+========+
+| rrd4kth1 | 152 | 613 | 9 | 6 |
+| rrd4kth5 | 279 | 1119 | 11 | 17 |
+| rrd4kth10 | 353 | 1416 | 2 | 28 |
+| rrd4kth15 | 393 | 1574 | 2 | 38 |
+| rrd4kth20 | 424 | 1697 | 2 | 47 |
+| rrd4kth30 | 461 | 1846 | 2 | 64 |
+| rrd4kth40 | 460 | 1841 | 2 | 86 |
+| rrd4kth80 | 475 | 1904 | 2 | 166 |
+| rrd4kth120 | 480 | 1923 | 3 | 246 |
+| --- | --- | --- | --- | --- |
+| rwd4kth1 | 406 | 1626 | 1 | 2 |
+| --- | --- | --- | --- | --- |
+| rws4kth1 | 41 | 166 | 2 | 24 |
+| rws4kth5 | 72 | 292 | 4 | 68 |
+| rws4kth10 | 80 | 322 | 9 | 123 |
+| rws4kth15 | 79 | 320 | 9 | 186 |
+| rws4kth20 | 80 | 322 | 7 | 246 |
+| rws4kth30 | 80 | 322 | 9 | 369 |
+| rws4kth40 | 79 | 320 | 9 | 494 |
+| rws4kth80 | 79 | 319 | 9 | 980 |
+| rws4kth120 | 80 | 322 | 7 | 1444 |
+| --- | --- | --- | --- | --- |
+| srd1mth1 | 179 | 184578 | 1 | 5 |
+| srd1mth5 | 150 | 154834 | 25 | 33 |
+| srd1mth10 | 169 | 174428 | 12 | 58 |
+| srd1mth15 | 182 | 187533 | 10 | 81 |
+| srd1mth20 | 272 | 279189 | 17 | 73 |
+| srd1mth30 | 387 | 397705 | 28 | 77 |
+| srd1mth40 | 452 | 464206 | 10 | 88 |
+| srd1mth80 | 481 | 493546 | 4 | 165 |
+| srd1mth120 | 457 | 469002 | 13 | 261 |
+| --- | --- | --- | --- | --- |
+| swd1mth1 | 174 | 179262 | 2 | 5 |
+| swd1mth5 | 148 | 152880 | 6 | 33 |
+| swd1mth10 | 150 | 154288 | 5 | 66 |
+| swd1mth15 | 151 | 155139 | 5 | 98 |
+| swd1mth20 | 151 | 155203 | 4 | 131 |
+| swd1mth30 | 158 | 162451 | 14 | 188 |
+| swd1mth40 | 169 | 173707 | 7 | 233 |
+| swd1mth80 | 195 | 200244 | 8 | 394 |
+| swd1mth120 | 208 | 214313 | 9 | 542 |
++-------------+------+---------+-----------+--------+