VMware testing on Microsoft Windows 10 Home Build 18362 via the Phoronix Test Suite.
Compare your own system(s) to this result file with the
Phoronix Test Suite by running the command:
phoronix-test-suite benchmark 2101092-SP-2009246NE84
{
"title": "ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020",
"last_modified": "2021-01-09 15:56:59",
"description": "VMware testing on Microsoft Windows 10 Home Build 18362 via the Phoronix Test Suite.",
"reference_id": "2009246-NE-THREADR1W39",
"systems": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"identifier": "ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020",
"hardware": {
"Processor": "AMD Ryzen Threadripper 3960X 24-Core @ 3.80GHz (24 Cores \/ 48 Threads)",
"Motherboard": "MSI Creator TRX40 (MS-7C59) (1.50 BIOS)",
"Memory": "4 x 16384 MB 1800MHz CMT64GX4M4K3600C16",
"Disk": "932GB Samsung SSD 860 EVO 1TB + 2 x 466GB Force MP600 + 4 x 477GB Samsung SSD 970 PRO 512GB + 57GB SanDisk Cruzer Blade USB + 477GB Samsung SSD 860 PRO 512GB + 10GB USB DISK CD USB + 932GB Force MP600 + 932GB JMicron Generic Disk + 932GB Inateck ASM1153E Disk + 954GB Inateck Disk + 954GB Samsung SSD 860 PRO 1TB + 1863GB Inateck ASM1153E Disk",
"Graphics": "AMD Radeon VII 4GB",
"Network": "Intel I211 Gigabit Connection + Intel Wi-Fi 6 AX200 160MHz"
},
"software": {
"OS": "Microsoft Windows 10 Enterprise LTSC Build 17763",
"Kernel": "10.0 (x86_64)",
"Display Driver": "27.20.1027.2005",
"OpenCL": "OpenCL 2.1 AMD-APP (3075.13)",
"Compiler": "GCC 8.3.0",
"File-System": "NTFS",
"Screen Resolution": "1920x1080"
},
"user": "SpingBasic",
"timestamp": "2020-09-22 01:01:05",
"data": {
"java": "OpenJDK Runtime Environment 18.9 (build 11.0.6+10-LTS)",
"python": "Python 3.7.9",
"security": "__user pointer sanitization: Disabled + Retpoline: Full + IBPB: Always + STIBP: Enabled"
}
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"identifier": "gogz-win81-00vm-ccta-cpu-massive-8jan2021",
"hardware": {
"Processor": "AMD Ryzen Threadripper 3960X 24-Core @ 3.80GHz (8 Cores)",
"Motherboard": "Intel 440BX",
"Memory": "1 x 8192 MB 0MHz VMW-8192MB",
"Disk": "120GB VMware VMware Virtual S + 100GB VMware VMware Virtual S",
"Graphics": "VMware SVGA 3D",
"Network": "Intel 82574L Gigabit Connection #2 + Intel 82574L Gigabit Connection #3 + TAP-Windows V9"
},
"software": {
"OS": "Microsoft Windows 10 Home Build 18362",
"Kernel": "10.0 (x86_64)",
"Display Driver": "8.16.7.5",
"Compiler": "GCC 7.1.0",
"File-System": "NTFS",
"Screen Resolution": "1895x968",
"System Layer": "VMware"
},
"user": "Chi",
"timestamp": "2021-01-09 02:31:11"
}
},
"results": {
"7fda2c8031e473865874c39b8ba37efcccc384b0": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "MOSTLY_CACHE HEAVY_CONTENTION READ_WRITE",
"description": "Scaling: Mostly RAM - Test: Heavy Contention - Mode: Read Write",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 12277.0268439999999827705323696136474609375,
"raw_values": [
9483.0749250000008032657206058502197265625,
13520.292347000000518164597451686859130859375,
13014.86703699999998207204043865203857421875,
10312.590436000000408967025578022003173828125,
12732.59068900000056601129472255706787109375,
12866.281246000000464846380054950714111328125,
12623.7093349999995552934706211090087890625,
13017.466303999999581719748675823211669921875,
12922.369280000000799191184341907501220703125
],
"test_run_times": [
3355.670000000000072759576141834259033203125,
3361.69000000000005456968210637569427490234375,
3365.09999999999990905052982270717620849609375,
3447.53000000000020008883439004421234130859375,
3417.2899999999999636202119290828704833984375,
3402.25,
3334.9600000000000363797880709171295166015625,
3381.38999999999987267074175179004669189453125,
3400.239999999999781721271574497222900390625
]
}
}
},
"11649e745d454fd5d98aa6ad135c43296d8b2a0e": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "ON_DISK HEAVY_CONTENTION READ_ONLY",
"description": "Scaling: On-Disk - Test: Heavy Contention - Mode: Read Only",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 388299.911826999974437057971954345703125,
"raw_values": [
382171.41986000002361834049224853515625,
388620.913816999993287026882171630859375,
394107.4018049999722279608249664306640625
],
"test_run_times": [
4617.4600000000000363797880709171295166015625,
4607.9399999999995998223312199115753173828125,
4522.670000000000072759576141834259033203125
]
}
}
},
"d6ea74c179ada8147bd542d1e364b05c2763f25c": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "ON_DISK HEAVY_CONTENTION READ_WRITE",
"description": "Scaling: On-Disk - Test: Heavy Contention - Mode: Read Write",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 1273.20264300000008006463758647441864013671875,
"raw_values": [
1265.537861999999904583091847598552703857421875,
1281.586939000000029409420676529407501220703125,
1272.48312699999996766564436256885528564453125
],
"test_run_times": [
4475.899999999999636202119290828704833984375,
4486.02000000000043655745685100555419921875,
4498.399999999999636202119290828704833984375
]
}
}
},
"df2c810c84611ca3583a60283ccf36c34b2dbb5e": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "MOSTLY_CACHE HEAVY_CONTENTION READ_ONLY",
"description": "Scaling: Mostly RAM - Test: Heavy Contention - Mode: Read Only",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 126995.70524599999771453440189361572265625,
"raw_values": [
125126.335852000003797002136707305908203125,
129299.16046899999491870403289794921875,
126561.61941600000136531889438629150390625
],
"test_run_times": [
3381.829999999999927240423858165740966796875,
3418.55999999999994543031789362430572509765625,
3372.65999999999985448084771633148193359375
]
}
}
},
"4c35c7ef81c117cd77c00e1a2fca916073e5176b": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "MOSTLY_CACHE SINGLE_THREAD READ_WRITE",
"description": "Scaling: Mostly RAM - Test: Single Thread - Mode: Read Write",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 1339.339561999999887120793573558330535888671875,
"raw_values": [
1297.797956000000112908310256898403167724609375,
1352.552028000000063912011682987213134765625,
1367.66870100000005550100468099117279052734375
],
"test_run_times": [
3335.7899999999999636202119290828704833984375,
3411.38999999999987267074175179004669189453125,
3331.03000000000020008883439004421234130859375
]
}
}
},
"05cf268a1bd590ed44b0290394dc409a0705e797": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "MOSTLY_CACHE SINGLE_THREAD READ_ONLY",
"description": "Scaling: Mostly RAM - Test: Single Thread - Mode: Read Only",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 3775.96230299999979251879267394542694091796875,
"raw_values": [
3798.8768730000001596636138856410980224609375,
3716.22207699999989927164278924465179443359375,
3812.78796000000011190422810614109039306640625
],
"test_run_times": [
3371.01999999999998181010596454143524169921875,
3344.94000000000005456968210637569427490234375,
3347.73000000000001818989403545856475830078125
]
}
}
},
"2b19a6308da71d23f6fdf23cb294745a8899d8b1": {
"identifier": "pts\/nginx-1.2.2",
"title": "NGINX Benchmark",
"app_version": "1.9.9",
"description": "Static Web Page Serving",
"scale": "Requests Per Second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 731.0900000000000318323145620524883270263671875,
"raw_values": [
721.5399999999999636202119290828704833984375,
734.6200000000000045474735088646411895751953125,
737.1100000000000136424205265939235687255859375
],
"test_run_times": [
2787.55999999999994543031789362430572509765625,
2738.13999999999987267074175179004669189453125,
2728.94000000000005456968210637569427490234375
]
}
}
},
"fa028f5ee807c5bf67a72f64e2358dde1f0d0ea0": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "ON_DISK SINGLE_THREAD READ_ONLY",
"description": "Scaling: On-Disk - Test: Single Thread - Mode: Read Only",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 23456,
"raw_values": [
21497.85195200000089243985712528228759765625,
23847.321423000001232139766216278076171875,
25022.66853099999934784136712551116943359375
],
"test_run_times": [
4576.920000000000072759576141834259033203125,
4818.9499999999998181010596454143524169921875,
4839.670000000000072759576141834259033203125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 4399,
"raw_values": [
4415.3535060000003795721568167209625244140625,
4382.795822000000043772161006927490234375
],
"test_run_times": [
248.55000000000001136868377216160297393798828125,
201.44999999999998863131622783839702606201171875,
0.070000000000000006661338147750939242541790008544921875,
0.190000000000000002220446049250313080847263336181640625
]
}
}
},
"f5424b42195c71e9b11259bd62716f63ff3206b2": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "ON_DISK SINGLE_THREAD READ_WRITE",
"description": "Scaling: On-Disk - Test: Single Thread - Mode: Read Write",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 42.20312899999999700639818911440670490264892578125,
"raw_values": [
42.30132900000000262252797256223857402801513671875,
42.08528199999999941383066470734775066375732421875,
42.22277700000000066893335315398871898651123046875
],
"test_run_times": [
4506.72999999999956344254314899444580078125,
4647.6800000000002910383045673370361328125,
4623.0500000000001818989403545856475830078125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 180.80000000000001136868377216160297393798828125,
"raw_values": [
179.66464600000000473301042802631855010986328125,
181.927494999999993297024047933518886566162109375
],
"test_run_times": [
206.150000000000005684341886080801486968994140625,
210.6399999999999863575794734060764312744140625,
0.070000000000000006661338147750939242541790008544921875,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"4f92e19442e6e76294f4b303bf8059b6b49216b5": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "ON_DISK NORMAL_LOAD READ_WRITE",
"description": "Scaling: On-Disk - Test: Normal Load - Mode: Read Write",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 1308,
"raw_values": [
1315.034769999999980427674017846584320068359375,
1302.586645000000089567038230597972869873046875,
1304.958378999999922598362900316715240478515625
],
"test_run_times": [
3980.1300000000001091393642127513885498046875,
4625.739999999999781721271574497222900390625,
4661.1899999999995998223312199115753173828125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 1784,
"raw_values": [
1804.389535999999907289748080074787139892578125,
1763.532547999999906096491031348705291748046875
],
"test_run_times": [
212.729999999999989768184605054557323455810546875,
217.830000000000012505552149377763271331787109375,
0.11000000000000000055511151231257827021181583404541015625,
0.2200000000000000011102230246251565404236316680908203125
]
}
}
},
"ee40faab64e98100a54ff22c56593430d4219d8f": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "ON_DISK NORMAL_LOAD READ_ONLY",
"description": "Scaling: On-Disk - Test: Normal Load - Mode: Read Only",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 389451,
"raw_values": [
390212.47147400001995265483856201171875,
391526.1676309999893419444561004638671875,
386614.6295800000079907476902008056640625
],
"test_run_times": [
4079.829999999999927240423858165740966796875,
4000.920000000000072759576141834259033203125,
3976.88999999999987267074175179004669189453125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 78464,
"raw_values": [
78646.24599500000476837158203125,
78281.750723999997717328369617462158203125
],
"test_run_times": [
220.05000000000001136868377216160297393798828125,
219.1100000000000136424205265939235687255859375,
0.070000000000000006661338147750939242541790008544921875,
0.070000000000000006661338147750939242541790008544921875
]
}
}
},
"964cbe068efc3f8ad6e279d944e771f6d9d70e13": {
"identifier": "pts\/blender-1.7.1",
"title": "Blender",
"app_version": "2.82",
"arguments": "-b ..\/barbershop_interior_gpu.blend -o output.test -x 1 -F JPEG -f 1 OPENCL",
"description": "Blend File: Barbershop - Compute: OpenCL",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 614.029999999999972715158946812152862548828125,
"raw_values": [
666.200000000000045474735088646411895751953125,
613.6799999999999499777914024889469146728515625,
601.8200000000000500222085975110530853271484375,
603.6299999999999954525264911353588104248046875,
616.4099999999999681676854379475116729736328125,
619.4600000000000363797880709171295166015625,
602.5700000000000500222085975110530853271484375,
602.3300000000000409272615797817707061767578125,
600.1499999999999772626324556767940521240234375
],
"test_run_times": [
737.950000000000045474735088646411895751953125,
684.8200000000000500222085975110530853271484375,
672.8200000000000500222085975110530853271484375,
674.48000000000001818989403545856475830078125,
687.1499999999999772626324556767940521240234375,
690.44000000000005456968210637569427490234375,
673.200000000000045474735088646411895751953125,
673.4199999999999590727384202182292938232421875,
671.1000000000000227373675443232059478759765625
]
}
}
},
"38ffcce4263977c87ee9483e7777e7dc0813194a": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "MOSTLY_CACHE NORMAL_LOAD READ_WRITE",
"description": "Scaling: Mostly RAM - Test: Normal Load - Mode: Read Write",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 12878,
"raw_values": [
13104.69198300000061863102018833160400390625,
13068.005279999999402207322418689727783203125,
12462.700570000000880099833011627197265625
],
"test_run_times": [
3407.67999999999983629095368087291717529296875,
3424.51999999999998181010596454143524169921875,
3336.84999999999990905052982270717620849609375
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 3650,
"raw_values": [
3658.7931159999998271814547479152679443359375,
3640.34961599999996906262822449207305908203125
],
"test_run_times": [
123.8900000000000005684341886080801486968994140625,
137.8899999999999863575794734060764312744140625,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"19ea2443be9cb6407a151d0d773471a430064db7": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "MOSTLY_CACHE NORMAL_LOAD READ_ONLY",
"description": "Scaling: Mostly RAM - Test: Normal Load - Mode: Read Only",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 130419,
"raw_values": [
130186.705474000002141110599040985107421875,
130279.8973450000048615038394927978515625,
130791.614038999992772005498409271240234375
],
"test_run_times": [
3333.94000000000005456968210637569427490234375,
3357.5399999999999636202119290828704833984375,
3296.4499999999998181010596454143524169921875
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 76118,
"raw_values": [
80552.459516999995685182511806488037109375,
71683.37183899999945424497127532958984375
],
"test_run_times": [
135.94999999999998863131622783839702606201171875,
137.909999999999996589394868351519107818603515625,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"b399a6747a6ba64f46124b5671250713d00f6fcb": {
"identifier": "pts\/blender-1.7.1",
"title": "Blender",
"app_version": "2.82",
"arguments": "-b ..\/pavillon_barcelone_gpu.blend -o output.test -x 1 -F JPEG -f 1 OPENCL",
"description": "Blend File: Pabellon Barcelona - Compute: OpenCL",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 1211.240000000000009094947017729282379150390625,
"raw_values": [
1215.140000000000100044417195022106170654296875,
1203.799999999999954525264911353588104248046875,
1214.779999999999972715158946812152862548828125
],
"test_run_times": [
1237.640000000000100044417195022106170654296875,
1268.30999999999994543031789362430572509765625,
1279.359999999999899955582804977893829345703125
]
}
}
},
"7bfe52f038c5b130d71757b1f9876e09fb4490b1": {
"identifier": "pts\/blender-1.7.1",
"title": "Blender",
"app_version": "2.82",
"arguments": "-b ..\/fishy_cat_gpu.blend -o output.test -x 1 -F JPEG -f 1 OPENCL",
"description": "Blend File: Fishy Cat - Compute: OpenCL",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 1209.069999999999936335370875895023345947265625,
"raw_values": [
1224.549999999999954525264911353588104248046875,
1197.319999999999936335370875895023345947265625,
1205.339999999999918145476840436458587646484375
],
"test_run_times": [
1247.2100000000000363797880709171295166015625,
1262.009999999999990905052982270717620849609375,
1270
]
}
}
},
"5ea36db146ece8d78e82e6ec6526ac2bdae1f6db": {
"identifier": "pts\/t-test1-1.0.1",
"title": "t-test1",
"app_version": "2017-01-13",
"arguments": "2",
"description": "Threads: 2",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 130,
"raw_values": [
126.7339999999999946567186270840466022491455078125,
133.039999999999992041921359486877918243408203125,
129.9370000000000118234311230480670928955078125
],
"test_run_times": [
126.7300000000000039790393202565610408782958984375,
133.039999999999992041921359486877918243408203125,
129.93999999999999772626324556767940521240234375
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 3238,
"raw_values": [
3268.999800205199790070764720439910888671875,
3207.03075504300022657844237983226776123046875
],
"test_run_times": [
3269,
3207.03000000000020008883439004421234130859375,
0.070000000000000006661338147750939242541790008544921875,
0.070000000000000006661338147750939242541790008544921875
]
}
}
},
"94b48082d6e5461c9bc3ee3a0f96cb51327bc7ab": {
"identifier": "pts\/renaissance-1.1.1",
"title": "Renaissance",
"app_version": "0.10.0",
"arguments": "akka-uct",
"description": "Test: Akka Unbalanced Cobwebbed Tree",
"scale": "ms",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 10987,
"raw_values": [
10719.1630000000004656612873077392578125,
11533.85599999999976716935634613037109375,
10969.89700000000084401108324527740478515625,
11676.35299999999915598891675472259521484375,
11295.75300000000061118043959140777587890625,
10675.3199999999997089616954326629638671875,
10541.69399999999950523488223552703857421875,
11036.40999999999985448084771633148193359375,
11031.3389999999999417923390865325927734375,
11047.671000000000276486389338970184326171875,
10693.012000000000625732354819774627685546875,
11366.072000000000116415321826934814453125,
11061.96299999999973806552588939666748046875,
10598.618000000000392901711165904998779296875,
10553.69800000000032014213502407073974609375
],
"test_run_times": [
267,
269.8799999999999954525264911353588104248046875,
266.66000000000002501110429875552654266357421875,
267.720000000000027284841053187847137451171875,
277.70999999999997953636921010911464691162109375,
261.509999999999990905052982270717620849609375,
262.69999999999998863131622783839702606201171875,
269.18999999999999772626324556767940521240234375,
270.529999999999972715158946812152862548828125,
268.41000000000002501110429875552654266357421875,
264.33999999999997498889570124447345733642578125,
272.3799999999999954525264911353588104248046875,
265.1299999999999954525264911353588104248046875,
263.66000000000002501110429875552654266357421875,
262.68999999999999772626324556767940521240234375
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 21180,
"raw_values": [
21076.0970000000015716068446636199951171875,
21283.77200000000084401108324527740478515625
],
"test_run_times": [
640.6200000000000045474735088646411895751953125,
502.44999999999998863131622783839702606201171875,
0.070000000000000006661338147750939242541790008544921875,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"35040347f411be86c49ef7e27c1fae3e43e8daaf": {
"identifier": "pts\/lczero-1.4.0",
"title": "LeelaChessZero",
"app_version": "0.25",
"arguments": "-b blas",
"description": "Backend: BLAS",
"scale": "Nodes Per Second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 671,
"raw_values": [
696,
682,
638,
667,
655,
692,
676,
664
],
"test_run_times": [
357.259999999999990905052982270717620849609375,
355.51999999999998181010596454143524169921875,
354.970000000000027284841053187847137451171875,
355.43999999999999772626324556767940521240234375,
354.93999999999999772626324556767940521240234375,
355.67000000000001591615728102624416351318359375,
354.20999999999997953636921010911464691162109375,
354.8600000000000136424205265939235687255859375
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 638,
"raw_values": [
832,
443
],
"test_run_times": [
360.56999999999999317878973670303821563720703125,
356.470000000000027284841053187847137451171875,
0.0899999999999999966693309261245303787291049957275390625,
0.1000000000000000055511151231257827021181583404541015625,
0.11000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625,
0.13000000000000000444089209850062616169452667236328125
]
}
}
},
"5db7d6938ebc18fdd26330fa9b532afcca9b820b": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "BUFFER_TEST SINGLE_THREAD READ_ONLY",
"description": "Scaling: Buffer Test - Test: Single Thread - Mode: Read Only",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 14115.4693530000004102475941181182861328125,
"raw_values": [
14758.66267499999958090484142303466796875,
14762.03015299999970011413097381591796875,
13744.04439099999945028685033321380615234375,
13168.020868999999947845935821533203125,
14717.908695000000079744495451450347900390625,
14724.74614199999996344558894634246826171875,
13445.537173999999140505678951740264892578125,
13071.207488000000012107193470001220703125,
14791.44098999999914667569100856781005859375,
14818.76752800000031129457056522369384765625,
13658.59610400000019581057131290435791015625,
13090.9794089999995776452124118804931640625,
13619.0522899999996297992765903472900390625,
14559.79774600000018835999071598052978515625,
14801.248642000000472762621939182281494140625
],
"test_run_times": [
113.4200000000000017053025658242404460906982421875,
112.6200000000000045474735088646411895751953125,
111.340000000000003410605131648480892181396484375,
111.650000000000005684341886080801486968994140625,
111.8900000000000005684341886080801486968994140625,
111.9200000000000017053025658242404460906982421875,
112.1400000000000005684341886080801486968994140625,
112.31000000000000227373675443232059478759765625,
112.18999999999999772626324556767940521240234375,
112.2600000000000051159076974727213382720947265625,
111.5,
113.2699999999999960209606797434389591217041015625,
113.0400000000000062527760746888816356658935546875,
111.2600000000000051159076974727213382720947265625,
111.280000000000001136868377216160297393798828125
]
}
}
},
"a29989100c7301c7213b8f55d7200be7bebbc7fc": {
"identifier": "pts\/renaissance-1.1.1",
"title": "Renaissance",
"app_version": "0.10.0",
"arguments": "reactors",
"description": "Test: Savina Reactors.IO",
"scale": "ms",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 13141,
"raw_values": [
13284.6800000000002910383045673370361328125,
12295.9830000000001746229827404022216796875,
12437.686999999999898136593401432037353515625,
12797.264999999999417923390865325927734375,
16341.906000000000858562998473644256591796875,
12259.188000000000101863406598567962646484375,
15176.010000000000218278728425502777099609375,
13138.314000000000305590219795703887939453125,
13292.153000000000247382558882236480712890625,
13126.815000000000509317032992839813232421875,
11510.9680000000007566995918750762939453125,
12466.435999999999694409780204296112060546875,
12573.53900000000066938810050487518310546875,
10756.89700000000084401108324527740478515625,
13009.923000000000683940015733242034912109375,
14579.07300000000032014213502407073974609375,
15503.30700000000069849193096160888671875,
12281.290999999999257852323353290557861328125,
12169.81899999999950523488223552703857421875,
13818.576999999999316059984266757965087890625
],
"test_run_times": [
140,
123.5799999999999982946974341757595539093017578125,
120.7300000000000039790393202565610408782958984375,
139.740000000000009094947017729282379150390625,
149.5,
138.849999999999994315658113919198513031005859375,
148.259999999999990905052982270717620849609375,
139.75,
127.31000000000000227373675443232059478759765625,
135.280000000000001136868377216160297393798828125,
122.9500000000000028421709430404007434844970703125,
129.280000000000001136868377216160297393798828125,
135.81000000000000227373675443232059478759765625,
128.659999999999996589394868351519107818603515625,
137.740000000000009094947017729282379150390625,
139.19999999999998863131622783839702606201171875,
153.18999999999999772626324556767940521240234375,
126.2600000000000051159076974727213382720947265625,
137.830000000000012505552149377763271331787109375,
140.539999999999992041921359486877918243408203125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 15766,
"raw_values": [
14412.774999999999636202119290828704833984375,
17119.28299999999944702722132205963134765625
],
"test_run_times": [
153.729999999999989768184605054557323455810546875,
157.6100000000000136424205265939235687255859375,
0.08000000000000000166533453693773481063544750213623046875,
0.070000000000000006661338147750939242541790008544921875,
0.059999999999999997779553950749686919152736663818359375,
0.070000000000000006661338147750939242541790008544921875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"173fffdad56a6293614ce6fbd99dd45c3743dd33": {
"identifier": "pts\/renaissance-1.1.1",
"title": "Renaissance",
"app_version": "0.10.0",
"arguments": "page-rank",
"description": "Test: Apache Spark PageRank",
"scale": "ms",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 3373,
"raw_values": [
3433.44900000000006912159733474254608154296875,
3638.78400000000010550138540565967559814453125,
3785.20800000000008367351256310939788818359375,
3523.25,
3408.25300000000015643308870494365692138671875,
3388.45299999999997453414835035800933837890625,
3384.35899999999992360244505107402801513671875,
3422.009000000000014551915228366851806640625,
3022.12399999999979627318680286407470703125,
3351.8890000000001236912794411182403564453125,
3219.7730000000001382431946694850921630859375,
3328.78299999999990177457220852375030517578125,
3664.08800000000019281287677586078643798828125,
3225.58599999999978535925038158893585205078125,
3434.35399999999981446308083832263946533203125,
3308.72899999999981446308083832263946533203125,
3419.90900000000010550138540565967559814453125,
2997.56500000000005456968210637569427490234375,
3311.57299999999986539478413760662078857421875,
3678.13599999999996725819073617458343505859375,
3487.670000000000072759576141834259033203125,
3277.46700000000009822542779147624969482421875,
3118.052999999999883584678173065185546875,
3066.3429999999998472048901021480560302734375,
3422.22600000000011277734301984310150146484375
],
"test_run_times": [
88.93999999999999772626324556767940521240234375,
92.030000000000001136868377216160297393798828125,
93.1700000000000017053025658242404460906982421875,
90.349999999999994315658113919198513031005859375,
90.1099999999999994315658113919198513031005859375,
83.9800000000000039790393202565610408782958984375,
90.849999999999994315658113919198513031005859375,
89.280000000000001136868377216160297393798828125,
84.25,
88.5400000000000062527760746888816356658935546875,
87.969999999999998863131622783839702606201171875,
85.1099999999999994315658113919198513031005859375,
90.2600000000000051159076974727213382720947265625,
88.4599999999999937472239253111183643341064453125,
89.06999999999999317878973670303821563720703125,
88.1700000000000017053025658242404460906982421875,
89.06000000000000227373675443232059478759765625,
81.349999999999994315658113919198513031005859375,
88.93000000000000682121026329696178436279296875,
91.409999999999996589394868351519107818603515625,
89.31999999999999317878973670303821563720703125,
87.2000000000000028421709430404007434844970703125,
84.5400000000000062527760746888816356658935546875,
81.5499999999999971578290569595992565155029296875,
89.9500000000000028421709430404007434844970703125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 9426,
"raw_values": [
9766.5509999999994761310517787933349609375,
9085.158999999999650754034519195556640625
],
"test_run_times": [
250.830000000000012505552149377763271331787109375,
278.490000000000009094947017729282379150390625,
0.070000000000000006661338147750939242541790008544921875,
0.070000000000000006661338147750939242541790008544921875,
0.070000000000000006661338147750939242541790008544921875,
0.1000000000000000055511151231257827021181583404541015625,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"e1f890b694d827aa7548fefbbe292cf60939acc7": {
"identifier": "pts\/blender-1.7.1",
"title": "Blender",
"app_version": "2.82",
"arguments": "-b ..\/bmw27_gpu.blend -o output.test -x 1 -F JPEG -f 1 OPENCL",
"description": "Blend File: BMW27 - Compute: OpenCL",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 377.8700000000000045474735088646411895751953125,
"raw_values": [
389.29000000000002046363078989088535308837890625,
372.18999999999999772626324556767940521240234375,
372.1200000000000045474735088646411895751953125
],
"test_run_times": [
412.32999999999998408384271897375583648681640625,
394.240000000000009094947017729282379150390625,
394.259999999999990905052982270717620849609375
]
}
}
},
"62662fe4970af15356a2d2c6262ade3c5b300d37": {
"identifier": "pts\/mrbayes-1.4.0",
"title": "Timed MrBayes Analysis",
"app_version": "3.2.7",
"description": "Primate Phylogeny Analysis",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 401,
"raw_values": [
400.27899999999999636202119290828704833984375,
401.26699999999999590727384202182292938232421875,
400.09300000000001773514668457210063934326171875
],
"test_run_times": [
400.279999999999972715158946812152862548828125,
401.26999999999998181010596454143524169921875,
400.08999999999997498889570124447345733642578125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 419,
"raw_values": [
420.7404079437300197241711430251598358154296875,
416.664968013760017129243351519107818603515625
],
"test_run_times": [
420.740000000000009094947017729282379150390625,
416.66000000000002501110429875552654266357421875,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"c9883de3b5843918d50bb37d38bd72266ba8f906": {
"identifier": "pts\/blender-1.7.1",
"title": "Blender",
"app_version": "2.82",
"arguments": "-b ..\/barbershop_interior_gpu.blend -o output.test -x 1 -F JPEG -f 1 NONE",
"description": "Blend File: Barbershop - Compute: CPU-Only",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 302.8799999999999954525264911353588104248046875,
"raw_values": [
300.69999999999998863131622783839702606201171875,
303.529999999999972715158946812152862548828125,
304.41000000000002501110429875552654266357421875
],
"test_run_times": [
330.240000000000009094947017729282379150390625,
332.240000000000009094947017729282379150390625,
333.279999999999972715158946812152862548828125
]
}
}
},
"ac988c8aa133da34c672fdc3f1ac8d4017b1c759": {
"identifier": "pts\/blender-1.7.1",
"title": "Blender",
"app_version": "2.82",
"arguments": "-b ..\/classroom_gpu.blend -o output.test -x 1 -F JPEG -f 1 OPENCL",
"description": "Blend File: Classroom - Compute: OpenCL",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 303.3899999999999863575794734060764312744140625,
"raw_values": [
299.93000000000000682121026329696178436279296875,
308.029999999999972715158946812152862548828125,
302.220000000000027284841053187847137451171875
],
"test_run_times": [
322.08999999999997498889570124447345733642578125,
330.220000000000027284841053187847137451171875,
324.3799999999999954525264911353588104248046875
]
}
}
},
"97814e1f0e82fc4b949e5dda62f5c57ef7d9feba": {
"identifier": "pts\/lczero-1.4.0",
"title": "LeelaChessZero",
"app_version": "0.25",
"arguments": "-b eigen",
"description": "Backend: Eigen",
"scale": "Nodes Per Second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 545,
"raw_values": [
562,
541,
533
],
"test_run_times": [
358.06000000000000227373675443232059478759765625,
360.5,
356.6100000000000136424205265939235687255859375
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 332,
"raw_values": [
318,
345
],
"test_run_times": [
368.1000000000000227373675443232059478759765625,
361.8999999999999772626324556767940521240234375,
0.1000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"819d7d01ceccc0b0066bf5ba95f9d1258c7b2015": {
"identifier": "pts\/lczero-1.4.0",
"title": "LeelaChessZero",
"app_version": "0.25",
"arguments": "-b random",
"description": "Backend: Random",
"scale": "Nodes Per Second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 126764,
"raw_values": [
128571,
123181,
128541
],
"test_run_times": [
342.56000000000000227373675443232059478759765625,
342.56000000000000227373675443232059478759765625,
342.529999999999972715158946812152862548828125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 112996,
"raw_values": [
110416,
115576
],
"test_run_times": [
343.94999999999998863131622783839702606201171875,
343.279999999999972715158946812152862548828125,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"d26eb6b47a8b6e5be191497d72172b008d41f9c1": {
"identifier": "pts\/blogbench-1.1.0",
"title": "BlogBench",
"app_version": "1.1",
"arguments": "READ",
"description": "Test: Read",
"scale": "Final Score",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 21043,
"raw_values": [
21318,
21014,
20796
],
"test_run_times": [
368.68999999999999772626324556767940521240234375,
321.54000000000002046363078989088535308837890625,
320.6000000000000227373675443232059478759765625
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 6030,
"raw_values": [
6532,
5528
],
"test_run_times": [
315.30000000000001136868377216160297393798828125,
313.20999999999997953636921010911464691162109375,
0.08000000000000000166533453693773481063544750213623046875,
0.059999999999999997779553950749686919152736663818359375,
0.070000000000000006661338147750939242541790008544921875,
0.070000000000000006661338147750939242541790008544921875,
0.070000000000000006661338147750939242541790008544921875
]
}
}
},
"76cbd4b69aae2d0dcb99e18956775552f2395171": {
"identifier": "pts\/blender-1.7.1",
"title": "Blender",
"app_version": "2.82",
"arguments": "-b ..\/pavillon_barcelone_gpu.blend -o output.test -x 1 -F JPEG -f 1 NONE",
"description": "Blend File: Pabellon Barcelona - Compute: CPU-Only",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 225.080000000000012505552149377763271331787109375,
"raw_values": [
224.020000000000010231815394945442676544189453125,
224.400000000000005684341886080801486968994140625,
226.81000000000000227373675443232059478759765625
],
"test_run_times": [
288.48000000000001818989403545856475830078125,
246.8700000000000045474735088646411895751953125,
249.25
]
}
}
},
"fa60fd3ebf878d0448239bc9d7aaac894b28f0a8": {
"identifier": "pts\/renaissance-1.1.1",
"title": "Renaissance",
"app_version": "0.10.0",
"arguments": "db-shootout",
"description": "Test: In-Memory Database Shootout",
"scale": "ms",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 10996,
"raw_values": [
10945.0210000000006402842700481414794921875,
11118.51399999999921419657766819000244140625,
11054.468999999999141437001526355743408203125,
10806.78299999999944702722132205963134765625,
11056.533999999999650754034519195556640625
],
"test_run_times": [
179.94999999999998863131622783839702606201171875,
185.979999999999989768184605054557323455810546875,
186.1100000000000136424205265939235687255859375,
181.219999999999998863131622783839702606201171875,
187.1100000000000136424205265939235687255859375
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 15134,
"raw_values": [
19600.1310000000012223608791828155517578125,
10667.5360000000000582076609134674072265625
],
"test_run_times": [
212.520000000000010231815394945442676544189453125,
217.840000000000003410605131648480892181396484375,
0.0899999999999999966693309261245303787291049957275390625,
0.070000000000000006661338147750939242541790008544921875,
0.0899999999999999966693309261245303787291049957275390625,
0.1000000000000000055511151231257827021181583404541015625,
0.08000000000000000166533453693773481063544750213623046875,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"9a743b764f90241ba42dc43a928cd9361eaed2c9": {
"identifier": "pts\/blender-1.7.1",
"title": "Blender",
"app_version": "2.82",
"arguments": "-b ..\/classroom_gpu.blend -o output.test -x 1 -F JPEG -f 1 NONE",
"description": "Blend File: Classroom - Compute: CPU-Only",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 188.18999999999999772626324556767940521240234375,
"raw_values": [
188.969999999999998863131622783839702606201171875,
186.6200000000000045474735088646411895751953125,
188.979999999999989768184605054557323455810546875
],
"test_run_times": [
253.06999999999999317878973670303821563720703125,
208.669999999999987494447850622236728668212890625,
211.020000000000010231815394945442676544189453125
]
}
}
},
"48ca4e09e0763c37e3112e8028ef870a625c1da5": {
"identifier": "pts\/renaissance-1.1.1",
"title": "Renaissance",
"app_version": "0.10.0",
"arguments": "naive-bayes",
"description": "Test: Apache Spark Bayes",
"scale": "ms",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 681,
"raw_values": [
807.7569999999999481588019989430904388427734375,
798.7730000000000245563569478690624237060546875,
179,
797.597999999999956344254314899444580078125,
793.4370000000000118234311230480670928955078125,
807.5890000000000554791768081486225128173828125,
804.249000000000023646862246096134185791015625,
800.078999999999950887286104261875152587890625,
806.1670000000000300133251585066318511962890625,
815.0489999999999781721271574497222900390625,
812.0240000000000009094947017729282379150390625,
178.08699999999998908606357872486114501953125,
815.3229999999999790816218592226505279541015625,
822.115000000000009094947017729282379150390625,
182.955000000000012505552149377763271331787109375,
808.25199999999995270627550780773162841796875,
804.9740000000000463842297904193401336669921875,
806.9950000000000045474735088646411895751953125,
811.56899999999995998223312199115753173828125,
799.133000000000038198777474462985992431640625,
180.8990000000000009094947017729282379150390625,
798.7530000000000427462509833276271820068359375,
181.951999999999998181010596454143524169921875,
805.76800000000002910383045673370361328125,
796.008000000000038198777474462985992431640625
],
"test_run_times": [
36.1099999999999994315658113919198513031005859375,
35.75,
18.480000000000000426325641456060111522674560546875,
36.159999999999996589394868351519107818603515625,
35.63000000000000255795384873636066913604736328125,
35.8299999999999982946974341757595539093017578125,
36.219999999999998863131622783839702606201171875,
36.25999999999999801048033987171947956085205078125,
35.86999999999999744204615126363933086395263671875,
36.32000000000000028421709430404007434844970703125,
36.27000000000000312638803734444081783294677734375,
18.300000000000000710542735760100185871124267578125,
36.530000000000001136868377216160297393798828125,
36.18999999999999772626324556767940521240234375,
18.059999999999998721023075631819665431976318359375,
36.2000000000000028421709430404007434844970703125,
36.00999999999999801048033987171947956085205078125,
36.17999999999999971578290569595992565155029296875,
36.28999999999999914734871708787977695465087890625,
35.74000000000000198951966012828052043914794921875,
18.440000000000001278976924368180334568023681640625,
36.17999999999999971578290569595992565155029296875,
16.67999999999999971578290569595992565155029296875,
36.10000000000000142108547152020037174224853515625,
35.719999999999998863131622783839702606201171875
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 3918,
"raw_values": [
4027.58199999999987994669936597347259521484375,
3809.365999999999985448084771633148193359375
],
"test_run_times": [
149.05000000000001136868377216160297393798828125,
184.1399999999999863575794734060764312744140625,
0.1000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"44c9dca2ba5d3e14ac16e945e792a5f4d0dee1e6": {
"identifier": "pts\/apache-1.7.2",
"title": "Apache Benchmark",
"app_version": "2.4.29",
"description": "Static Web Page Serving",
"scale": "Requests Per Second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 6503.829999999999927240423858165740966796875,
"raw_values": [
6439.5399999999999636202119290828704833984375,
6537.239999999999781721271574497222900390625,
6534.6999999999998181010596454143524169921875
],
"test_run_times": [
155.479999999999989768184605054557323455810546875,
153.18000000000000682121026329696178436279296875,
153.25
]
}
}
},
"fc8b6de93b7723b344a80057c04367a883bac1ec": {
"identifier": "pts\/asmfish-1.1.1",
"title": "asmFish",
"app_version": "2018-07-23",
"description": "1024 Hash Memory, 26 Depth",
"scale": "Nodes\/second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 72340992,
"raw_values": [
71280014,
71889170,
73853791
],
"test_run_times": [
152.240000000000009094947017729282379150390625,
129.900000000000005684341886080801486968994140625,
136.530000000000001136868377216160297393798828125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 11498448,
"raw_values": [
11514707,
11482188
],
"test_run_times": [
259.06000000000000227373675443232059478759765625,
234.93000000000000682121026329696178436279296875,
0.11000000000000000055511151231257827021181583404541015625,
0.179999999999999993338661852249060757458209991455078125
]
}
}
},
"ffba90bc6e56b93d1e0102003a1dd9e41a8e227c": {
"identifier": "pts\/compress-7zip-1.7.1",
"title": "7-Zip Compression",
"app_version": "16.02",
"description": "Compress Speed Test",
"scale": "MIPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 106201,
"raw_values": [
100372,
114022,
99908,
100340,
99944,
115148,
112697,
115030,
99819,
99186,
99538,
116506,
101407,
102314,
116791
],
"test_run_times": [
48.02000000000000312638803734444081783294677734375,
44.07000000000000028421709430404007434844970703125,
48.43999999999999772626324556767940521240234375,
47.530000000000001136868377216160297393798828125,
48.11999999999999744204615126363933086395263671875,
44.92999999999999971578290569595992565155029296875,
45.469999999999998863131622783839702606201171875,
44.5,
48.21000000000000085265128291212022304534912109375,
47.6700000000000017053025658242404460906982421875,
48.13000000000000255795384873636066913604736328125,
42.969999999999998863131622783839702606201171875,
47.99000000000000198951966012828052043914794921875,
47.7999999999999971578290569595992565155029296875,
43.78999999999999914734871708787977695465087890625
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 19916,
"raw_values": [
19558,
20274
],
"test_run_times": [
36.17999999999999971578290569595992565155029296875,
36.7000000000000028421709430404007434844970703125,
0.08000000000000000166533453693773481063544750213623046875,
0.070000000000000006661338147750939242541790008544921875
]
}
}
},
"ea8b3eb302364079e21158085fac0b5f9d83b2f0": {
"identifier": "pts\/blender-1.7.1",
"title": "Blender",
"app_version": "2.82",
"arguments": "-b ..\/fishy_cat_gpu.blend -o output.test -x 1 -F JPEG -f 1 NONE",
"description": "Blend File: Fishy Cat - Compute: CPU-Only",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 104.4899999999999948840923025272786617279052734375,
"raw_values": [
104.5499999999999971578290569595992565155029296875,
102.8299999999999982946974341757595539093017578125,
106.099999999999994315658113919198513031005859375
],
"test_run_times": [
127.06999999999999317878973670303821563720703125,
125.31999999999999317878973670303821563720703125,
128.56000000000000227373675443232059478759765625
]
}
}
},
"f562bc9c757669fd20f1aa832e22b6648a239e2b": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "BUFFER_TEST HEAVY_CONTENTION READ_WRITE",
"description": "Scaling: Buffer Test - Test: Heavy Contention - Mode: Read Write",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 36148.2462949999971897341310977935791015625,
"raw_values": [
35851.4786890000032144598662853240966796875,
36723.38049199999659322202205657958984375,
35869.879702999998698942363262176513671875
],
"test_run_times": [
114.909999999999996589394868351519107818603515625,
115.7999999999999971578290569595992565155029296875,
115.0199999999999960209606797434389591217041015625
]
}
}
},
"07cf1bee84d2966a28d7285e8311bec9427a245e": {
"identifier": "pts\/compress-zstd-1.2.0",
"title": "Zstd Compression",
"app_version": "1.4.5",
"arguments": "-b19",
"description": "Compression Level: 19",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 40.7000000000000028421709430404007434844970703125,
"raw_values": [
40.89999999999999857891452847979962825775146484375,
40.39999999999999857891452847979962825775146484375,
40.7999999999999971578290569595992565155029296875
],
"test_run_times": [
69.599999999999994315658113919198513031005859375,
68.909999999999996589394868351519107818603515625,
69.599999999999994315658113919198513031005859375
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 15.199999999999999289457264239899814128875732421875,
"raw_values": [
14.5,
15.9000000000000003552713678800500929355621337890625
],
"test_run_times": [
317.1000000000000227373675443232059478759765625,
160.68000000000000682121026329696178436279296875,
0.2099999999999999922284388276239042170345783233642578125,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875,
0.070000000000000006661338147750939242541790008544921875
]
}
}
},
"48c6fecaafbf03ab4a1904bf55026f5006decc10": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "BUFFER_TEST SINGLE_THREAD READ_WRITE",
"description": "Scaling: Buffer Test - Test: Single Thread - Mode: Read Write",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 2015.742856000000074345734901726245880126953125,
"raw_values": [
2004.528317000000015468685887753963470458984375,
2012.307238000000097599695436656475067138671875,
2030.393013999999993757228367030620574951171875
],
"test_run_times": [
114.530000000000001136868377216160297393798828125,
113.9500000000000028421709430404007434844970703125,
112.18000000000000682121026329696178436279296875
]
}
}
},
"d088577125c72342c951cbeda831e10f4f40a72f": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "BUFFER_TEST HEAVY_CONTENTION READ_ONLY",
"description": "Scaling: Buffer Test - Test: Heavy Contention - Mode: Read Only",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 143212.4072370000067166984081268310546875,
"raw_values": [
142410.24336900000344030559062957763671875,
142658.97163700000965036451816558837890625,
144568.00670599998557008802890777587890625
],
"test_run_times": [
112.18999999999999772626324556767940521240234375,
111.780000000000001136868377216160297393798828125,
112.849999999999994315658113919198513031005859375
]
}
}
},
"0f845ac1a17cd256254d014ce39a409e29f136ab": {
"identifier": "pts\/renaissance-1.1.1",
"title": "Renaissance",
"app_version": "0.10.0",
"arguments": "dec-tree",
"description": "Test: Random Forest",
"scale": "ms",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 2091,
"raw_values": [
2052.64499999999998181010596454143524169921875,
2060.73500000000012732925824820995330810546875,
2068.07900000000017826096154749393463134765625,
2141.05499999999983629095368087291717529296875,
2130.510000000000218278728425502777099609375
],
"test_run_times": [
91.3900000000000005684341886080801486968994140625,
91.06999999999999317878973670303821563720703125,
90.9200000000000017053025658242404460906982421875,
93.3599999999999994315658113919198513031005859375,
93.7399999999999948840923025272786617279052734375
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 2067,
"raw_values": [
2092.03299999999990177457220852375030517578125,
2042.16300000000001091393642127513885498046875
],
"test_run_times": [
106.900000000000005684341886080801486968994140625,
97.8799999999999954525264911353588104248046875,
0.070000000000000006661338147750939242541790008544921875,
0.070000000000000006661338147750939242541790008544921875,
0.070000000000000006661338147750939242541790008544921875,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"f8b4f2829bbd82baf42a96ae5d0db82122541a34": {
"identifier": "pts\/cachebench-1.1.2",
"title": "CacheBench",
"arguments": "-b",
"description": "Test: Read \/ Modify \/ Write",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 31607,
"raw_values": [
31587.03443219999826396815478801727294921875,
31611.32414913300090120173990726470947265625,
31623.080722667000372894108295440673828125
],
"min_result": [
"31171.78"
],
"max_result": [
"31896.35"
],
"test_run_times": [
125.0799999999999982946974341757595539093017578125,
125.0799999999999982946974341757595539093017578125,
125.0799999999999982946974341757595539093017578125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 48931,
"raw_values": [
56135.6381274759987718425691127777099609375,
41725.6582458569973823614418506622314453125
],
"test_run_times": [
125.1700000000000017053025658242404460906982421875,
125.159999999999996589394868351519107818603515625,
0.11999999999999999555910790149937383830547332763671875,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"286caffe46f313fa6981b18b518b8f2ce68ade60": {
"identifier": "pts\/cachebench-1.1.2",
"title": "CacheBench",
"arguments": "-w",
"description": "Test: Write",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 31500,
"raw_values": [
31529.67021100000056321732699871063232421875,
31436.36740833300063968636095523834228515625,
31534.79451213300126255489885807037353515625
],
"min_result": [
"30621.71"
],
"max_result": [
"32079.18"
],
"test_run_times": [
125.0799999999999982946974341757595539093017578125,
125.0799999999999982946974341757595539093017578125,
125.0799999999999982946974341757595539093017578125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 13966,
"raw_values": [
13853.158130809999420307576656341552734375,
14079.101847666999674402177333831787109375
],
"test_run_times": [
125.1299999999999954525264911353588104248046875,
125.1700000000000017053025658242404460906982421875,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"52ee8b1c2f74fac119a6f4d3ec07a99d9a0a5cde": {
"identifier": "pts\/cachebench-1.1.2",
"title": "CacheBench",
"arguments": "-r",
"description": "Test: Read",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 10689,
"raw_values": [
10674.497442800000499119050800800323486328125,
10694.83450413299942738376557826995849609375,
10698.55232353300016256980597972869873046875
],
"min_result": [
"10642.59"
],
"max_result": [
"10718.22"
],
"test_run_times": [
125.090000000000003410605131648480892181396484375,
125.0799999999999982946974341757595539093017578125,
125.0799999999999982946974341757595539093017578125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 10287,
"raw_values": [
10277.876515714000561274588108062744140625,
10297.072729048000837792642414569854736328125
],
"test_run_times": [
125.150000000000005684341886080801486968994140625,
125.1200000000000045474735088646411895751953125,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"83c4f37052ad8eff43bd22c660b254c2d166f665": {
"identifier": "pts\/minion-1.5.2",
"title": "Minion",
"app_version": "1.8",
"arguments": "benchmarks\/Quasigroup\/qg-watchelement-7-10.minion",
"description": "Benchmark: Quasigroup",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 112,
"raw_values": [
112.3610000000000042064129956997931003570556640625,
112.1979999999999932924765744246542453765869140625,
112.1929999999999978399500832892954349517822265625
],
"test_run_times": [
112.43000000000000682121026329696178436279296875,
112.2600000000000051159076974727213382720947265625,
112.25
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 122,
"raw_values": [
122.0400000000000062527760746888816356658935546875,
122.0630000000000023874235921539366245269775390625
],
"test_run_times": [
122.1400000000000005684341886080801486968994140625,
122.150000000000005684341886080801486968994140625,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"f0635bf087c8c22b967aa795a85e5a48a3da1376": {
"identifier": "pts\/renaissance-1.1.1",
"title": "Renaissance",
"app_version": "0.10.0",
"arguments": "als",
"description": "Test: Apache Spark ALS",
"scale": "ms",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 2042,
"raw_values": [
2022.2899999999999636202119290828704833984375,
2043.138999999999896317603997886180877685546875,
2025.488000000000056388671509921550750732421875,
2061.90000000000009094947017729282379150390625,
2058.902000000000043655745685100555419921875
],
"test_run_times": [
68.400000000000005684341886080801486968994140625,
69.2900000000000062527760746888816356658935546875,
70.2399999999999948840923025272786617279052734375,
69.9200000000000017053025658242404460906982421875,
69.6200000000000045474735088646411895751953125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 3298,
"raw_values": [
2885.97100000000000363797880709171295166015625,
3710.3800000000001091393642127513885498046875
],
"test_run_times": [
92.7099999999999937472239253111183643341064453125,
111.0400000000000062527760746888816356658935546875,
0.070000000000000006661338147750939242541790008544921875,
0.070000000000000006661338147750939242541790008544921875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.070000000000000006661338147750939242541790008544921875,
0.070000000000000006661338147750939242541790008544921875
]
}
}
},
"da537b151d630d3406022e009f401b1b88c03aa6": {
"identifier": "pts\/blender-1.7.1",
"title": "Blender",
"app_version": "2.82",
"arguments": "-b ..\/bmw27_gpu.blend -o output.test -x 1 -F JPEG -f 1 NONE",
"description": "Blend File: BMW27 - Compute: CPU-Only",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 69.7699999999999960209606797434389591217041015625,
"raw_values": [
69.469999999999998863131622783839702606201171875,
69.3900000000000005684341886080801486968994140625,
70.4500000000000028421709430404007434844970703125
],
"test_run_times": [
91.4899999999999948840923025272786617279052734375,
91.340000000000003410605131648480892181396484375,
92.400000000000005684341886080801486968994140625
]
}
}
},
"a0a6ca760a440139eff96952ffeb95422093a58a": {
"identifier": "pts\/novabench-1.0.0",
"title": "Novabench",
"arguments": "--test=cpu",
"description": "Test: CPU",
"scale": "CPU Score",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 633,
"raw_values": [
523,
666,
597,
688,
647,
661,
670,
573,
623,
606,
642,
661,
645,
650,
650
],
"test_run_times": [
18.199999999999999289457264239899814128875732421875,
18.0799999999999982946974341757595539093017578125,
18.089999999999999857891452847979962825775146484375,
18.089999999999999857891452847979962825775146484375,
18.089999999999999857891452847979962825775146484375,
18.089999999999999857891452847979962825775146484375,
18.10000000000000142108547152020037174224853515625,
18.0799999999999982946974341757595539093017578125,
18.1099999999999994315658113919198513031005859375,
18.10000000000000142108547152020037174224853515625,
18.0799999999999982946974341757595539093017578125,
18.10000000000000142108547152020037174224853515625,
18.089999999999999857891452847979962825775146484375,
18.089999999999999857891452847979962825775146484375,
18.089999999999999857891452847979962825775146484375
]
}
}
},
"04b0d506572ee7ad24b2ddcb33b3b632ddf2f47d": {
"identifier": "pts\/dav1d-1.6.0",
"title": "dav1d",
"app_version": "0.7.0",
"arguments": "-i chimera_10b_1080p.ivf",
"description": "Video Input: Chimera 1080p 10-bit",
"scale": "FPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 102.1889579999999995152393239550292491912841796875,
"raw_values": [
100.4899999999999948840923025272786617279052734375,
103.4893749999999954525264911353588104248046875,
102.587500000000005684341886080801486968994140625
],
"min_result": [
"100.46"
],
"max_result": [
"103.51"
],
"test_run_times": [
89.2999999999999971578290569595992565155029296875,
86.4800000000000039790393202565610408782958984375,
87.2300000000000039790393202565610408782958984375
]
}
}
},
"9fb6b1f63bd87c00364226a39e5d2bb685cd0d83": {
"identifier": "pts\/cpuminer-opt-1.3.0",
"title": "Cpuminer-Opt",
"app_version": "3.8.8.1",
"arguments": "-a deep",
"description": "Algorithm: deep",
"scale": "kH\/s - Hash Speed",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 19221,
"raw_values": [
21610,
6291.4600000000000363797880709171295166015625,
22100,
21480,
22030,
21520,
41.93999999999999772626324556767940521240234375,
21390,
22070,
21650,
21960,
21740,
21470,
21170,
21790
],
"test_run_times": [
34.86999999999999744204615126363933086395263671875,
35.32000000000000028421709430404007434844970703125,
34.99000000000000198951966012828052043914794921875,
34.99000000000000198951966012828052043914794921875,
34.86999999999999744204615126363933086395263671875,
34.659999999999996589394868351519107818603515625,
35.42999999999999971578290569595992565155029296875,
34.77000000000000312638803734444081783294677734375,
35,
35.1400000000000005684341886080801486968994140625,
34.99000000000000198951966012828052043914794921875,
34.5499999999999971578290569595992565155029296875,
35.43999999999999772626324556767940521240234375
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 5235,
"raw_values": [
5266.0699999999997089616954326629638671875,
5203.1400000000003274180926382541656494140625
],
"test_run_times": [
32.75,
33.38000000000000255795384873636066913604736328125,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"d7314229cd48bfb4f238b2c179d7b33c6776ac0e": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "BUFFER_TEST NORMAL_LOAD READ_ONLY",
"description": "Scaling: Buffer Test - Test: Normal Load - Mode: Read Only",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 144452,
"raw_values": [
144164.27153200001339428126811981201171875,
144647.66216099998564459383487701416015625,
144544.79310199999599717557430267333984375
],
"test_run_times": [
112.2000000000000028421709430404007434844970703125,
112.8299999999999982946974341757595539093017578125,
112.9800000000000039790393202565610408782958984375
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 45727,
"raw_values": [
45031.1641820000004372559487819671630859375,
46421.8884109999999054707586765289306640625
],
"test_run_times": [
90.3299999999999982946974341757595539093017578125,
90.2699999999999960209606797434389591217041015625,
0.070000000000000006661338147750939242541790008544921875,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"58eea68af3edce96bd553fc212c717811fc393b7": {
"identifier": "pts\/cpuminer-opt-1.3.0",
"title": "Cpuminer-Opt",
"app_version": "3.8.8.1",
"arguments": "-a myr-gr",
"description": "Algorithm: myr-gr",
"scale": "kH\/s - Hash Speed",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 7687,
"raw_values": [
8768.379999999999199644662439823150634765625,
7645.7200000000002546585164964199066162109375,
7566.399999999999636202119290828704833984375,
7579.510000000000218278728425502777099609375,
7629.6800000000002910383045673370361328125,
7654.7200000000002546585164964199066162109375,
7593.850000000000363797880709171295166015625,
7573.65999999999985448084771633148193359375,
7622.3199999999997089616954326629638671875,
7609.2899999999999636202119290828704833984375,
7543.4499999999998181010596454143524169921875,
7535.7100000000000363797880709171295166015625,
7665.27000000000043655745685100555419921875,
7629.760000000000218278728425502777099609375
],
"test_run_times": [
32.0799999999999982946974341757595539093017578125,
31.489999999999998436805981327779591083526611328125,
32.47999999999999687361196265555918216705322265625,
31.620000000000000994759830064140260219573974609375,
32.3599999999999994315658113919198513031005859375,
31.60000000000000142108547152020037174224853515625,
32.3599999999999994315658113919198513031005859375,
31.730000000000000426325641456060111522674560546875,
32.219999999999998863131622783839702606201171875,
31.440000000000001278976924368180334568023681640625,
31.530000000000001136868377216160297393798828125,
31.35000000000000142108547152020037174224853515625,
31.620000000000000994759830064140260219573974609375,
31.550000000000000710542735760100185871124267578125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 7493,
"raw_values": [
6285.47999999999956344254314899444580078125,
8700.059999999999490682967007160186767578125
],
"test_run_times": [
31.239999999999998436805981327779591083526611328125,
30.8900000000000005684341886080801486968994140625,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.070000000000000006661338147750939242541790008544921875
]
}
}
},
"128f8170b40bf9f91c0f3e4e77094829ac809e06": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-ecrush,0",
"description": "Test: Crush 0 - Process: Decompression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 467,
"raw_values": [
468,
468,
467,
467,
468,
466,
466,
468,
468,
467,
468,
467,
467,
467,
468
],
"test_run_times": [
29.280000000000001136868377216160297393798828125,
30.78999999999999914734871708787977695465087890625,
28.3900000000000005684341886080801486968994140625,
28.769999999999999573674358543939888477325439453125,
28.739999999999998436805981327779591083526611328125,
22.309999999999998721023075631819665431976318359375,
30.21000000000000085265128291212022304534912109375,
28.260000000000001563194018672220408916473388671875,
28.730000000000000426325641456060111522674560546875,
22.809999999999998721023075631819665431976318359375,
30.6400000000000005684341886080801486968994140625,
30.980000000000000426325641456060111522674560546875,
28.53999999999999914734871708787977695465087890625,
31,
22.239999999999998436805981327779591083526611328125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 443,
"raw_values": [
445,
441
],
"test_run_times": [
27.089999999999999857891452847979962825775146484375,
27.6400000000000005684341886080801486968994140625,
0.0899999999999999966693309261245303787291049957275390625,
0.070000000000000006661338147750939242541790008544921875
]
}
}
},
"87df8d5b801ad3bd53e12a1906a9873b94369183": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-ecrush,0",
"description": "Test: Crush 0 - Process: Compression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 107,
"raw_values": [
115,
101,
116,
114,
119,
93,
106,
117,
116,
88,
102,
100,
119,
100,
93
],
"test_run_times": [
29.280000000000001136868377216160297393798828125,
30.78999999999999914734871708787977695465087890625,
28.3900000000000005684341886080801486968994140625,
28.769999999999999573674358543939888477325439453125,
28.739999999999998436805981327779591083526611328125,
22.309999999999998721023075631819665431976318359375,
30.21000000000000085265128291212022304534912109375,
28.260000000000001563194018672220408916473388671875,
28.730000000000000426325641456060111522674560546875,
22.809999999999998721023075631819665431976318359375,
30.6400000000000005684341886080801486968994140625,
30.980000000000000426325641456060111522674560546875,
28.53999999999999914734871708787977695465087890625,
31,
22.239999999999998436805981327779591083526611328125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 70,
"raw_values": [
71,
68
],
"test_run_times": [
27.089999999999999857891452847979962825775146484375,
27.6400000000000005684341886080801486968994140625,
0.0899999999999999966693309261245303787291049957275390625,
0.070000000000000006661338147750939242541790008544921875
]
}
}
},
"c5bfa87584fc49593a2851c8dbab5eac0b0bb10b": {
"identifier": "pts\/renaissance-1.1.1",
"title": "Renaissance",
"app_version": "0.10.0",
"arguments": "future-genetic",
"description": "Test: Genetic Algorithm Using Jenetics + Futures",
"scale": "ms",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 1094,
"raw_values": [
1101.548000000000001818989403545856475830078125,
1113.2629999999999199644662439823150634765625,
1098.642000000000052750692702829837799072265625,
1072.953999999999950887286104261875152587890625,
1084.83300000000008367351256310939788818359375
],
"test_run_times": [
56.68999999999999772626324556767940521240234375,
56.52000000000000312638803734444081783294677734375,
56.1099999999999994315658113919198513031005859375,
54.909999999999996589394868351519107818603515625,
55.72999999999999687361196265555918216705322265625
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 1757,
"raw_values": [
1729.1120000000000800355337560176849365234375,
1785.325000000000045474735088646411895751953125
],
"test_run_times": [
89.31000000000000227373675443232059478759765625,
92.9599999999999937472239253111183643341064453125,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.070000000000000006661338147750939242541790008544921875,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"7c62964246ec3d489ef68d9dd7de35b2a6057f19": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "BUFFER_TEST NORMAL_LOAD READ_WRITE",
"description": "Scaling: Buffer Test - Test: Normal Load - Mode: Read Write",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 37125,
"raw_values": [
36688.6589670000030309893190860748291015625,
37022.0126200000013341195881366729736328125,
37663.3456829999995534308254718780517578125
],
"test_run_times": [
115.840000000000003410605131648480892181396484375,
114.7399999999999948840923025272786617279052734375,
116.099999999999994315658113919198513031005859375
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 6822,
"raw_values": [
6822.1366310000003068125806748867034912109375
],
"test_run_times": [
66.75,
47.57000000000000028421709430404007434844970703125,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"7cfeee9e921038188a30cf66762a05c8877fa8fc": {
"identifier": "pts\/john-the-ripper-1.7.2",
"title": "John The Ripper",
"app_version": "1.9.0-jumbo-1",
"arguments": "--format=md5crypt",
"description": "Test: MD5",
"scale": "Real C\/S",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 1659667,
"raw_values": [
1662000,
1664000,
1653000
],
"test_run_times": [
92.9800000000000039790393202565610408782958984375,
92.68000000000000682121026329696178436279296875,
92.9800000000000039790393202565610408782958984375
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 380317,
"raw_values": [
359315,
401318
],
"test_run_times": [
69.7900000000000062527760746888816356658935546875,
98.68999999999999772626324556767940521240234375,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875,
0.070000000000000006661338147750939242541790008544921875,
0.070000000000000006661338147750939242541790008544921875,
0.070000000000000006661338147750939242541790008544921875
]
}
}
},
"bf4fff796527fd58f3aee1decf5986622a81a8db": {
"identifier": "pts\/go-benchmark-1.1.4",
"title": "Go Benchmarks",
"arguments": "build",
"description": "Test: build",
"scale": "Nanoseconds\/Operation",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 29396637323,
"raw_values": [
37940556500,
28315805500,
28212221400,
28397371000,
28562503900,
28770671600,
28629558400,
28445910400,
28611043100,
29335126500,
29129961200,
28911784800,
28893770900
],
"test_run_times": [
38.030000000000001136868377216160297393798828125,
28.39999999999999857891452847979962825775146484375,
28.280000000000001136868377216160297393798828125,
28.489999999999998436805981327779591083526611328125,
28.6400000000000005684341886080801486968994140625,
28.85000000000000142108547152020037174224853515625,
28.699999999999999289457264239899814128875732421875,
28.530000000000001136868377216160297393798828125,
28.690000000000001278976924368180334568023681640625,
29.4200000000000017053025658242404460906982421875,
29.21000000000000085265128291212022304534912109375,
28.989999999999998436805981327779591083526611328125,
28.96000000000000085265128291212022304534912109375
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 25334725850,
"raw_values": [
26258225500,
24411226200
],
"test_run_times": [
26.39999999999999857891452847979962825775146484375,
24.530000000000001136868377216160297393798828125,
0.059999999999999997779553950749686919152736663818359375,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"0a32a78f934d3c6f13e09c0f9aa99954b9a9ffc7": {
"identifier": "pts\/renaissance-1.1.1",
"title": "Renaissance",
"app_version": "0.10.0",
"arguments": "finagle-http",
"description": "Test: Twitter HTTP Requests",
"scale": "ms",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 4299,
"raw_values": [
4291.7200000000002546585164964199066162109375,
4243.5410000000001673470251262187957763671875,
4239.7349999999996725819073617458343505859375,
4328.1899999999995998223312199115753173828125,
4393.3249999999998181010596454143524169921875
],
"test_run_times": [
57.47999999999999687361196265555918216705322265625,
55.89999999999999857891452847979962825775146484375,
55.97999999999999687361196265555918216705322265625,
58.030000000000001136868377216160297393798828125,
57.35000000000000142108547152020037174224853515625
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 5734,
"raw_values": [
5120.3890000000001236912794411182403564453125,
6348.466000000000349245965480804443359375
],
"test_run_times": [
70.1700000000000017053025658242404460906982421875,
76.18999999999999772626324556767940521240234375,
0.08000000000000000166533453693773481063544750213623046875,
0.070000000000000006661338147750939242541790008544921875,
0.070000000000000006661338147750939242541790008544921875,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"2d198da79dae8a42c45d979302d1803b4c52bdd5": {
"identifier": "pts\/v-ray-1.2.1",
"title": "Chaos Group V-RAY",
"app_version": "4.10.07",
"arguments": "-m vray",
"description": "Mode: CPU",
"scale": "Ksamples",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 34538,
"raw_values": [
34100,
34940,
34573
],
"test_run_times": [
72.099999999999994315658113919198513031005859375,
71.7000000000000028421709430404007434844970703125,
71.780000000000001136868377216160297393798828125
]
}
}
},
"46a591eee486b9dbaf87a9ef7f3f9554435f353d": {
"identifier": "pts\/stockfish-1.1.1",
"title": "Stockfish",
"app_version": "9",
"description": "Total Time",
"scale": "Nodes Per Second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 69222386,
"raw_values": [
69317222,
70747396,
67602540
],
"test_run_times": [
60.8599999999999994315658113919198513031005859375,
56.25,
58.81000000000000227373675443232059478759765625
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 10704311,
"raw_values": [
11142985,
10265637
],
"test_run_times": [
117.1299999999999954525264911353588104248046875,
133.43000000000000682121026329696178436279296875,
0.070000000000000006661338147750939242541790008544921875,
0.070000000000000006661338147750939242541790008544921875,
0.070000000000000006661338147750939242541790008544921875,
0.070000000000000006661338147750939242541790008544921875,
0.070000000000000006661338147750939242541790008544921875
]
}
}
},
"b3275b0d8c80227f7222f198278a8d6e1564d846": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-ebrotli,2",
"description": "Test: Brotli 2 - Process: Decompression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 670,
"raw_values": [
672,
670,
666,
657,
670,
670,
670,
672,
669,
670,
670,
671,
671,
671,
674
],
"test_run_times": [
26.89999999999999857891452847979962825775146484375,
23.230000000000000426325641456060111522674560546875,
23.32000000000000028421709430404007434844970703125,
23.6099999999999994315658113919198513031005859375,
27.480000000000000426325641456060111522674560546875,
22.550000000000000710542735760100185871124267578125,
22.3299999999999982946974341757595539093017578125,
27.030000000000001136868377216160297393798828125,
27.410000000000000142108547152020037174224853515625,
27.519999999999999573674358543939888477325439453125,
22.780000000000001136868377216160297393798828125,
22.219999999999998863131622783839702606201171875,
22.190000000000001278976924368180334568023681640625,
27.050000000000000710542735760100185871124267578125,
27.32000000000000028421709430404007434844970703125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 638,
"raw_values": [
640,
635
],
"test_run_times": [
24.760000000000001563194018672220408916473388671875,
23.28999999999999914734871708787977695465087890625,
0.070000000000000006661338147750939242541790008544921875,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"f7bde4847668cfa74d587c36950b7ec182656eba": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-ebrotli,2",
"description": "Test: Brotli 2 - Process: Compression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 186,
"raw_values": [
192,
181,
173,
169,
191,
187,
187,
191,
190,
191,
183,
190,
188,
191,
191
],
"test_run_times": [
26.89999999999999857891452847979962825775146484375,
23.230000000000000426325641456060111522674560546875,
23.32000000000000028421709430404007434844970703125,
23.6099999999999994315658113919198513031005859375,
27.480000000000000426325641456060111522674560546875,
22.550000000000000710542735760100185871124267578125,
22.3299999999999982946974341757595539093017578125,
27.030000000000001136868377216160297393798828125,
27.410000000000000142108547152020037174224853515625,
27.519999999999999573674358543939888477325439453125,
22.780000000000001136868377216160297393798828125,
22.219999999999998863131622783839702606201171875,
22.190000000000001278976924368180334568023681640625,
27.050000000000000710542735760100185871124267578125,
27.32000000000000028421709430404007434844970703125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 173,
"raw_values": [
174,
171
],
"test_run_times": [
24.760000000000001563194018672220408916473388671875,
23.28999999999999914734871708787977695465087890625,
0.070000000000000006661338147750939242541790008544921875,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"c75ad335f359edae4899c29f49cde43aeb58e789": {
"identifier": "pts\/cpuminer-opt-1.3.0",
"title": "Cpuminer-Opt",
"app_version": "3.8.8.1",
"arguments": "-a skein",
"description": "Algorithm: skein",
"scale": "kH\/s - Hash Speed",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 72514,
"raw_values": [
77490,
2621.4499999999998181010596454143524169921875,
76860,
80840,
78870,
78940,
78610,
76870,
77420,
78400,
80920,
82330
],
"test_run_times": [
30.969999999999998863131622783839702606201171875,
31.120000000000000994759830064140260219573974609375,
31.199999999999999289457264239899814128875732421875,
30.989999999999998436805981327779591083526611328125,
30.980000000000000426325641456060111522674560546875,
30.989999999999998436805981327779591083526611328125,
30.75,
31.14999999999999857891452847979962825775146484375,
30.660000000000000142108547152020037174224853515625,
31.300000000000000710542735760100185871124267578125,
31.03999999999999914734871708787977695465087890625
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 12447,
"raw_values": [
23420,
524.2899999999999636202119290828704833984375,
24270,
1572.8699999999998908606357872486114501953125
],
"test_run_times": [
31.120000000000000994759830064140260219573974609375,
30.719999999999998863131622783839702606201171875,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.070000000000000006661338147750939242541790008544921875,
0.070000000000000006661338147750939242541790008544921875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"6f249b3604d44d4d3f31896bb0653b59dd533a28": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-ezstd,1",
"description": "Test: Zstd 1 - Process: Decompression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 829,
"raw_values": [
824,
833,
835,
821,
831,
832,
809,
842,
812,
832,
828,
839,
825,
840,
839
],
"test_run_times": [
23.35000000000000142108547152020037174224853515625,
22.1400000000000005684341886080801486968994140625,
21.969999999999998863131622783839702606201171875,
23.3299999999999982946974341757595539093017578125,
23.559999999999998721023075631819665431976318359375,
23.1700000000000017053025658242404460906982421875,
22.949999999999999289457264239899814128875732421875,
22.879999999999999005240169935859739780426025390625,
22.969999999999998863131622783839702606201171875,
22.78999999999999914734871708787977695465087890625,
23.050000000000000710542735760100185871124267578125,
24.17999999999999971578290569595992565155029296875,
22.739999999999998436805981327779591083526611328125,
23.8900000000000005684341886080801486968994140625,
24.019999999999999573674358543939888477325439453125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 780,
"raw_values": [
739,
820
],
"test_run_times": [
24.32000000000000028421709430404007434844970703125,
24.96000000000000085265128291212022304534912109375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"32c63cd3d6f9367ae720a9b9afc31d7ebf02dc32": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-ezstd,1",
"description": "Test: Zstd 1 - Process: Compression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 495,
"raw_values": [
505,
452,
456,
483,
475,
491,
509,
506,
510,
507,
508,
500,
509,
508,
507
],
"test_run_times": [
23.35000000000000142108547152020037174224853515625,
22.1400000000000005684341886080801486968994140625,
21.969999999999998863131622783839702606201171875,
23.3299999999999982946974341757595539093017578125,
23.559999999999998721023075631819665431976318359375,
23.1700000000000017053025658242404460906982421875,
22.949999999999999289457264239899814128875732421875,
22.879999999999999005240169935859739780426025390625,
22.969999999999998863131622783839702606201171875,
22.78999999999999914734871708787977695465087890625,
23.050000000000000710542735760100185871124267578125,
24.17999999999999971578290569595992565155029296875,
22.739999999999998436805981327779591083526611328125,
23.8900000000000005684341886080801486968994140625,
24.019999999999999573674358543939888477325439453125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 460,
"raw_values": [
438,
482
],
"test_run_times": [
24.32000000000000028421709430404007434844970703125,
24.96000000000000085265128291212022304534912109375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"d17be9fede4ecead5b0d4491becef37b02e8bfd8": {
"identifier": "pts\/john-the-ripper-1.7.2",
"title": "John The Ripper",
"app_version": "1.9.0-jumbo-1",
"arguments": "--format=bcrypt",
"description": "Test: Blowfish",
"scale": "Real C\/S",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 34003,
"raw_values": [
33694,
36054,
33571,
33526,
33596,
33575
],
"test_run_times": [
52.85000000000000142108547152020037174224853515625,
49.3299999999999982946974341757595539093017578125,
52.93999999999999772626324556767940521240234375,
53.02000000000000312638803734444081783294677734375,
52.909999999999996589394868351519107818603515625,
52.93999999999999772626324556767940521240234375
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 8144,
"raw_values": [
8445,
7843
],
"test_run_times": [
35.780000000000001136868377216160297393798828125,
37.97999999999999687361196265555918216705322265625,
0.1000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625,
0.070000000000000006661338147750939242541790008544921875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"3d6c0a94e5b200281e3eaf197b04f2628760514d": {
"identifier": "pts\/m-queens-1.1.0",
"title": "m-queens",
"app_version": "1.2",
"description": "Time To Solve",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 25.57600000000000051159076974727213382720947265625,
"raw_values": [
25.428000000000000824229573481716215610504150390625,
25.620999999999998664179656771011650562286376953125,
25.678999999999998493649400188587605953216552734375
],
"test_run_times": [
25.42999999999999971578290569595992565155029296875,
25.620000000000000994759830064140260219573974609375,
25.67999999999999971578290569595992565155029296875
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 144.159999999999996589394868351519107818603515625,
"raw_values": [
140.447530031199988798107369802892208099365234375,
147.864150047300000778704998083412647247314453125
],
"test_run_times": [
140.44999999999998863131622783839702606201171875,
147.8600000000000136424205265939235687255859375,
0.070000000000000006661338147750939242541790008544921875,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"3f13c4be7c50b339286084a06826f68c6931739b": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-ebrotli,0",
"description": "Test: Brotli 0 - Process: Decompression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 580,
"raw_values": [
579,
581,
581,
579,
579,
578,
581,
581,
580,
579,
579,
580
],
"test_run_times": [
22.519999999999999573674358543939888477325439453125,
23.82000000000000028421709430404007434844970703125,
23.550000000000000710542735760100185871124267578125,
23.82000000000000028421709430404007434844970703125,
23.440000000000001278976924368180334568023681640625,
23.8299999999999982946974341757595539093017578125,
23.85000000000000142108547152020037174224853515625,
23.35000000000000142108547152020037174224853515625,
23.53999999999999914734871708787977695465087890625,
23.5,
23.8599999999999994315658113919198513031005859375,
23.780000000000001136868377216160297393798828125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 556,
"raw_values": [
550,
561
],
"test_run_times": [
23.6099999999999994315658113919198513031005859375,
25.719999999999998863131622783839702606201171875,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"f38647cf6335ade425c0eac6f6df2d7613680603": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-ebrotli,0",
"description": "Test: Brotli 0 - Process: Compression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 399,
"raw_values": [
268,
410,
414,
409,
413,
408,
411,
414,
412,
411,
412,
408
],
"test_run_times": [
22.519999999999999573674358543939888477325439453125,
23.82000000000000028421709430404007434844970703125,
23.550000000000000710542735760100185871124267578125,
23.82000000000000028421709430404007434844970703125,
23.440000000000001278976924368180334568023681640625,
23.8299999999999982946974341757595539093017578125,
23.85000000000000142108547152020037174224853515625,
23.35000000000000142108547152020037174224853515625,
23.53999999999999914734871708787977695465087890625,
23.5,
23.8599999999999994315658113919198513031005859375,
23.780000000000001136868377216160297393798828125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 382,
"raw_values": [
379,
385
],
"test_run_times": [
23.6099999999999994315658113919198513031005859375,
25.719999999999998863131622783839702606201171875,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"317be2daeb2913752106421a5f105f2a4216240d": {
"identifier": "system\/tesseract-ocr-1.0.1",
"title": "Tesseract OCR",
"app_version": "v4.0.0.20181030",
"description": "Time To OCR 7 Images",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 53.88799999999999812416717759333550930023193359375,
"raw_values": [
53.46600000000000108002495835535228252410888671875,
54.11800000000000210320649784989655017852783203125,
54.07900000000000062527760746888816356658935546875
],
"test_run_times": [
53.469999999999998863131622783839702606201171875,
54.11999999999999744204615126363933086395263671875,
54.0799999999999982946974341757595539093017578125
]
}
}
},
"d5a9a5e26f759aa7fe88ee6cfe06ffafbc2f83e8": {
"identifier": "pts\/graphics-magick-2.0.1",
"title": "GraphicsMagick",
"app_version": "1.3.33",
"arguments": "-sharpen 0x2.0",
"description": "Operation: Sharpen",
"scale": "Iterations Per Minute",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 136,
"raw_values": [
135,
136,
136
],
"test_run_times": [
60.18999999999999772626324556767940521240234375,
60.25,
60.0799999999999982946974341757595539093017578125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 36,
"test_run_times": [
61.409999999999996589394868351519107818603515625,
61.63000000000000255795384873636066913604736328125,
0.11000000000000000055511151231257827021181583404541015625,
0.070000000000000006661338147750939242541790008544921875
]
}
}
},
"216ef4bdc8cdeeef23a07753dd79d116198ba190": {
"identifier": "pts\/graphics-magick-2.0.1",
"title": "GraphicsMagick",
"app_version": "1.3.33",
"arguments": "-resize 50%",
"description": "Operation: Resizing",
"scale": "Iterations Per Minute",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 561,
"raw_values": [
560,
562,
562
],
"test_run_times": [
60.090000000000003410605131648480892181396484375,
60.1700000000000017053025658242404460906982421875,
60.090000000000003410605131648480892181396484375
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 87,
"raw_values": [
83,
91
],
"test_run_times": [
60.340000000000003410605131648480892181396484375,
60.67999999999999971578290569595992565155029296875,
0.14000000000000001332267629550187848508358001708984375,
0.11999999999999999555910790149937383830547332763671875,
0.11999999999999999555910790149937383830547332763671875,
0.11999999999999999555910790149937383830547332763671875,
0.14000000000000001332267629550187848508358001708984375
]
}
}
},
"b29f4a46364bc8c91b0300c9af2da1a2d5a9993c": {
"identifier": "pts\/graphics-magick-2.0.1",
"title": "GraphicsMagick",
"app_version": "1.3.33",
"arguments": "-operator all Noise-Gaussian 30%",
"description": "Operation: Noise-Gaussian",
"scale": "Iterations Per Minute",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 262,
"raw_values": [
264,
260,
261
],
"test_run_times": [
60.17999999999999971578290569595992565155029296875,
60.18999999999999772626324556767940521240234375,
60.06000000000000227373675443232059478759765625
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 63,
"raw_values": [
49,
76
],
"test_run_times": [
60.72999999999999687361196265555918216705322265625,
60.25,
0.08000000000000000166533453693773481063544750213623046875,
0.1000000000000000055511151231257827021181583404541015625,
0.1000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"2932bde0f723a15670f91da170ab533913725f85": {
"identifier": "pts\/graphics-magick-2.0.1",
"title": "GraphicsMagick",
"app_version": "1.3.33",
"arguments": "-enhance",
"description": "Operation: Enhanced",
"scale": "Iterations Per Minute",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 381,
"raw_values": [
384,
377,
381
],
"test_run_times": [
60.21000000000000085265128291212022304534912109375,
60.21000000000000085265128291212022304534912109375,
60.090000000000003410605131648480892181396484375
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 94,
"raw_values": [
98,
90
],
"test_run_times": [
60.469999999999998863131622783839702606201171875,
60.159999999999996589394868351519107818603515625,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.070000000000000006661338147750939242541790008544921875,
0.070000000000000006661338147750939242541790008544921875,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"48c19a54c4daa4cf389fbfef01f5a1ec99df189d": {
"identifier": "pts\/graphics-magick-2.0.1",
"title": "GraphicsMagick",
"app_version": "1.3.33",
"arguments": "-rotate 90",
"description": "Operation: Rotate",
"scale": "Iterations Per Minute",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 353,
"raw_values": [
356,
350,
352
],
"test_run_times": [
60.159999999999996589394868351519107818603515625,
60.2000000000000028421709430404007434844970703125,
60.090000000000003410605131648480892181396484375
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 249,
"raw_values": [
268,
230
],
"test_run_times": [
60.25999999999999801048033987171947956085205078125,
60.31000000000000227373675443232059478759765625,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.070000000000000006661338147750939242541790008544921875,
0.070000000000000006661338147750939242541790008544921875,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"e9218196b66940de67c64189b01914d2cc2fb0e5": {
"identifier": "pts\/graphics-magick-2.0.1",
"title": "GraphicsMagick",
"app_version": "1.3.33",
"arguments": "-swirl 90",
"description": "Operation: Swirl",
"scale": "Iterations Per Minute",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 545,
"raw_values": [
541,
547,
547
],
"test_run_times": [
60.13000000000000255795384873636066913604736328125,
60.0799999999999982946974341757595539093017578125,
60.07000000000000028421709430404007434844970703125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 203,
"raw_values": [
193,
212
],
"test_run_times": [
60.38000000000000255795384873636066913604736328125,
60.1700000000000017053025658242404460906982421875,
0.1000000000000000055511151231257827021181583404541015625,
0.08000000000000000166533453693773481063544750213623046875,
0.070000000000000006661338147750939242541790008544921875,
0.08000000000000000166533453693773481063544750213623046875,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"accedc6e67d43152e34f7e88c587d92267c7b6a9": {
"identifier": "pts\/graphics-magick-2.0.1",
"title": "GraphicsMagick",
"app_version": "1.3.33",
"arguments": "-colorspace HWB",
"description": "Operation: HWB Color Space",
"scale": "Iterations Per Minute",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 452,
"raw_values": [
448,
454,
454
],
"test_run_times": [
60.1400000000000005684341886080801486968994140625,
60.0799999999999982946974341757595539093017578125,
60.11999999999999744204615126363933086395263671875
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 226,
"raw_values": [
229,
223
],
"test_run_times": [
60.21000000000000085265128291212022304534912109375,
60.340000000000003410605131648480892181396484375,
0.0899999999999999966693309261245303787291049957275390625,
0.1000000000000000055511151231257827021181583404541015625
]
}
}
},
"1ab5b4ed40c087b99cdeeacac2864e3fc9e296c3": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-exz,0",
"description": "Test: XZ 0 - Process: Decompression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 121,
"test_run_times": [
41.85000000000000142108547152020037174224853515625,
41.57000000000000028421709430404007434844970703125,
41.50999999999999801048033987171947956085205078125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 100,
"raw_values": [
94,
106
],
"test_run_times": [
88.840000000000003410605131648480892181396484375,
75.5799999999999982946974341757595539093017578125,
0.08000000000000000166533453693773481063544750213623046875,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"a97fff0392e0dc8b06e7492f95d75e65766f23e6": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-exz,0",
"description": "Test: XZ 0 - Process: Compression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 38,
"test_run_times": [
41.85000000000000142108547152020037174224853515625,
41.57000000000000028421709430404007434844970703125,
41.50999999999999801048033987171947956085205078125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 31,
"raw_values": [
30,
31
],
"test_run_times": [
88.840000000000003410605131648480892181396484375,
75.5799999999999982946974341757595539093017578125,
0.08000000000000000166533453693773481063544750213623046875,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"32f6380f8c703d3d84bd215727a68137f54f31cb": {
"identifier": "pts\/c-ray-1.2.0",
"title": "C-Ray",
"app_version": "1.1",
"description": "Total Time - 4K, 16 Rays Per Pixel",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 22.385000000000001563194018672220408916473388671875,
"raw_values": [
22.5150000000000005684341886080801486968994140625,
22.34100000000000108002495835535228252410888671875,
22.29899999999999948840923025272786617279052734375
],
"test_run_times": [
24.910000000000000142108547152020037174224853515625,
24.690000000000001278976924368180334568023681640625,
24.629999999999999005240169935859739780426025390625
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 93.2999999999999971578290569595992565155029296875,
"raw_values": [
93.7770000000000010231815394945442676544189453125,
92.820999999999997953636921010911464691162109375
],
"test_run_times": [
96.5,
95.349999999999994315658113919198513031005859375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"2f9e749dc7b6b26e2e18e63972e940a1aa5ad80c": {
"identifier": "pts\/minion-1.5.2",
"title": "Minion",
"app_version": "1.8",
"arguments": "benchmarks\/graceful\/k6p2_table.minion",
"description": "Benchmark: Graceful",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 49.10900000000000176214598468504846096038818359375,
"raw_values": [
49.0870000000000032969182939268648624420166015625,
48.98599999999999710098563809879124164581298828125,
49.2530000000000001136868377216160297393798828125
],
"test_run_times": [
49.17999999999999971578290569595992565155029296875,
49.06000000000000227373675443232059478759765625,
49.31000000000000227373675443232059478759765625
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 58.60000000000000142108547152020037174224853515625,
"raw_values": [
58.6940000000000026147972675971686840057373046875,
58.51200000000000045474735088646411895751953125
],
"test_run_times": [
59.28999999999999914734871708787977695465087890625,
58.6099999999999994315658113919198513031005859375,
0.08000000000000000166533453693773481063544750213623046875,
0.070000000000000006661338147750939242541790008544921875
]
}
}
},
"2b540121ed8dc4275bde294a726e7059480f5985": {
"identifier": "pts\/minion-1.5.2",
"title": "Minion",
"app_version": "1.8",
"arguments": "benchmarks\/solitaire\/solitaire_benchmark_8.minion",
"description": "Benchmark: Solitaire",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 48.143000000000000682121026329696178436279296875,
"raw_values": [
48.3250000000000028421709430404007434844970703125,
48.06000000000000227373675443232059478759765625,
48.0439999999999969304553815163671970367431640625
],
"test_run_times": [
48.409999999999996589394868351519107818603515625,
48.14999999999999857891452847979962825775146484375,
48.11999999999999744204615126363933086395263671875
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 59.159999999999996589394868351519107818603515625,
"raw_values": [
59.71000000000000085265128291212022304534912109375,
58.60900000000000176214598468504846096038818359375
],
"test_run_times": [
59.86999999999999744204615126363933086395263671875,
58.72999999999999687361196265555918216705322265625,
0.070000000000000006661338147750939242541790008544921875,
0.070000000000000006661338147750939242541790008544921875
]
}
}
},
"87114e45f9a85f8303291ae0736e9ffa7f5f9dcf": {
"identifier": "pts\/compress-xz-1.1.0",
"title": "XZ Compression",
"app_version": "5.2.4",
"description": "Compressing ubuntu-16.04.3-server-i386.img, Compression Level 9",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 23.245000000000000994759830064140260219573974609375,
"raw_values": [
23.434999999999998721023075631819665431976318359375,
23.106999999999999317878973670303821563720703125,
23.193000000000001392663762089796364307403564453125
],
"test_run_times": [
23.42999999999999971578290569595992565155029296875,
23.1099999999999994315658113919198513031005859375,
23.190000000000001278976924368180334568023681640625
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 67.4599999999999937472239253111183643341064453125,
"raw_values": [
67.7929770946499985484479111619293689727783203125,
67.136078834534004045053734444081783294677734375
],
"test_run_times": [
67.780000000000001136868377216160297393798828125,
67.1400000000000005684341886080801486968994140625,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"a405928c66376da5d339299827bb08a752c1990b": {
"identifier": "pts\/java-scimark2-1.1.2",
"title": "Java SciMark",
"app_version": "2.0",
"arguments": "TEST_COMPOSITE",
"description": "Computational Test: Composite",
"scale": "Mflops",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 2702,
"raw_values": [
2713.7839628837000418570823967456817626953125,
2720.23489172180006789858452975749969482421875,
2654.360415599499901873059570789337158203125,
2720.84462485599988212925381958484649658203125
],
"test_run_times": [
28.629999999999999005240169935859739780426025390625,
28.5799999999999982946974341757595539093017578125,
29.92999999999999971578290569595992565155029296875,
28.559999999999998721023075631819665431976318359375
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 2936,
"raw_values": [
2935.184241772500172373838722705841064453125,
2937.0878288704998340108431875705718994140625
],
"test_run_times": [
31.160000000000000142108547152020037174224853515625,
31.3299999999999982946974341757595539093017578125,
0.070000000000000006661338147750939242541790008544921875,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"07d46f075f5bd9e04f1eb7212d0c6c9fc50099af": {
"identifier": "pts\/cpuminer-opt-1.3.0",
"title": "Cpuminer-Opt",
"app_version": "3.8.8.1",
"arguments": "-a m7m",
"description": "Algorithm: m7m",
"scale": "kH\/s - Hash Speed",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 1101,
"raw_values": [
1122.25,
1087.40000000000009094947017729282379150390625,
1094.279999999999972715158946812152862548828125
],
"test_run_times": [
35.8599999999999994315658113919198513031005859375,
35.75,
36.14999999999999857891452847979962825775146484375
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 266,
"raw_values": [
265.3600000000000136424205265939235687255859375,
267.56999999999999317878973670303821563720703125
],
"test_run_times": [
34.00999999999999801048033987171947956085205078125,
34.10000000000000142108547152020037174224853515625,
0.070000000000000006661338147750939242541790008544921875,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"946f5e0360cf178c425ab9080de3597a148b632b": {
"identifier": "pts\/phpbench-1.1.5",
"title": "PHPBench",
"app_version": "0.8.1",
"description": "PHP Benchmark Suite",
"scale": "Score",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 382488,
"raw_values": [
383614,
381915,
381934
],
"test_run_times": [
26.199999999999999289457264239899814128875732421875,
26.260000000000001563194018672220408916473388671875,
26.280000000000001136868377216160297393798828125
]
}
}
},
"055f9243aefb2de6118ed0c85a05c6ca0ede0071": {
"identifier": "pts\/cpuminer-opt-1.3.0",
"title": "Cpuminer-Opt",
"app_version": "3.8.8.1",
"arguments": "-a sha256t",
"description": "Algorithm: sha256t",
"scale": "kH\/s - Hash Speed",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 168613,
"raw_values": [
169900,
168850,
167090
],
"test_run_times": [
30.4200000000000017053025658242404460906982421875,
30.870000000000000994759830064140260219573974609375,
30.989999999999998436805981327779591083526611328125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 49180,
"raw_values": [
49480,
48880
],
"test_run_times": [
30.589999999999999857891452847979962825775146484375,
31.300000000000000710542735760100185871124267578125,
0.059999999999999997779553950749686919152736663818359375,
0.070000000000000006661338147750939242541790008544921875
]
}
}
},
"fd8a26d206b1ff7e664749268590d8ee61acc1ad": {
"identifier": "pts\/cpuminer-opt-1.3.0",
"title": "Cpuminer-Opt",
"app_version": "3.8.8.1",
"arguments": "-a lbry",
"description": "Algorithm: lbry",
"scale": "kH\/s - Hash Speed",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 66793,
"raw_values": [
68200,
67250,
64930
],
"test_run_times": [
30.809999999999998721023075631819665431976318359375,
30.8299999999999982946974341757595539093017578125,
31.120000000000000994759830064140260219573974609375
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 11998,
"raw_values": [
18390,
17080,
524.2899999999999636202119290828704833984375
],
"test_run_times": [
30.129999999999999005240169935859739780426025390625,
31.019999999999999573674358543939888477325439453125,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"6b1815d46eb146239522c2b1521de3ae4512fb73": {
"identifier": "pts\/crafty-1.4.4",
"title": "Crafty",
"app_version": "25.2",
"description": "Elapsed Time",
"scale": "Nodes Per Second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 8204260,
"raw_values": [
8190463,
8219677,
8202639
],
"test_run_times": [
23.92999999999999971578290569595992565155029296875,
23.949999999999999289457264239899814128875732421875,
24.010000000000001563194018672220408916473388671875
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 4958276,
"raw_values": [
4730987,
5185564
],
"test_run_times": [
41.47999999999999687361196265555918216705322265625,
38.28999999999999914734871708787977695465087890625,
0.08000000000000000166533453693773481063544750213623046875,
0.070000000000000006661338147750939242541790008544921875,
0.070000000000000006661338147750939242541790008544921875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"edb61ba9199c6f5934fef86679c889a5f7bfdd5f": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-ezstd,8",
"description": "Test: Zstd 8 - Process: Decompression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 846,
"raw_values": [
840,
843,
855
],
"test_run_times": [
30.879999999999999005240169935859739780426025390625,
31.8599999999999994315658113919198513031005859375,
31.469999999999998863131622783839702606201171875
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 824,
"raw_values": [
828,
820
],
"test_run_times": [
24.32000000000000028421709430404007434844970703125,
25.35000000000000142108547152020037174224853515625,
0.070000000000000006661338147750939242541790008544921875,
0.070000000000000006661338147750939242541790008544921875
]
}
}
},
"25c9b4e1e91e600c606af13ee9cfca164a8525fc": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-ezstd,8",
"description": "Test: Zstd 8 - Process: Compression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 99,
"test_run_times": [
30.879999999999999005240169935859739780426025390625,
31.8599999999999994315658113919198513031005859375,
31.469999999999998863131622783839702606201171875
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 80,
"raw_values": [
81,
78
],
"test_run_times": [
24.32000000000000028421709430404007434844970703125,
25.35000000000000142108547152020037174224853515625,
0.070000000000000006661338147750939242541790008544921875,
0.070000000000000006661338147750939242541790008544921875
]
}
}
},
"61f7806218ca246f1cf151f32a69e40ea45376b6": {
"identifier": "pts\/t-test1-1.0.1",
"title": "t-test1",
"app_version": "2017-01-13",
"arguments": "1",
"description": "Threads: 1",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 23.33500000000000085265128291212022304534912109375,
"raw_values": [
23.870999999999998664179656771011650562286376953125,
23.251000000000001222133505507372319698333740234375,
22.8840000000000003410605131648480892181396484375
],
"test_run_times": [
23.870000000000000994759830064140260219573974609375,
23.25,
22.879999999999999005240169935859739780426025390625
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 32.030000000000001136868377216160297393798828125,
"raw_values": [
32.3015789985660006777834496460855007171630859375,
31.753429889679001263402824406512081623077392578125
],
"test_run_times": [
32.2999999999999971578290569595992565155029296875,
31.75,
0.070000000000000006661338147750939242541790008544921875,
0.070000000000000006661338147750939242541790008544921875
]
}
}
},
"cc5f7480baf5f1901713e506466f0fbe4342712d": {
"identifier": "pts\/ebizzy-1.0.4",
"title": "ebizzy",
"app_version": "0.3",
"scale": "Records\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 5682,
"raw_values": [
5695,
5674,
5676
],
"test_run_times": [
25.440000000000001278976924368180334568023681640625,
24.17999999999999971578290569595992565155029296875,
25.219999999999998863131622783839702606201171875
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 691,
"raw_values": [
373,
1008
],
"test_run_times": [
20.9200000000000017053025658242404460906982421875,
21.89999999999999857891452847979962825775146484375,
0.190000000000000002220446049250313080847263336181640625,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875,
0.070000000000000006661338147750939242541790008544921875,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"32c31601b81a0078b59cf31c4b9a5edfdb2d6ac7": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-elibdeflate,1",
"description": "Test: Libdeflate 1 - Process: Decompression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 1227,
"raw_values": [
1226,
1231,
1225
],
"test_run_times": [
23.370000000000000994759830064140260219573974609375,
22.199999999999999289457264239899814128875732421875,
23.300000000000000710542735760100185871124267578125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 1185,
"raw_values": [
1194,
1176
],
"test_run_times": [
24.8599999999999994315658113919198513031005859375,
24.010000000000001563194018672220408916473388671875,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"a481174911c170e0ad2a765ad48b6e015cc7b36e": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-elibdeflate,1",
"description": "Test: Libdeflate 1 - Process: Compression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 253,
"test_run_times": [
23.370000000000000994759830064140260219573974609375,
22.199999999999999289457264239899814128875732421875,
23.300000000000000710542735760100185871124267578125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 241,
"raw_values": [
240,
241
],
"test_run_times": [
24.8599999999999994315658113919198513031005859375,
24.010000000000001563194018672220408916473388671875,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"f7d9afeee0ad6d3ac57fa2ea86ed8a7dc9b45df4": {
"identifier": "pts\/dacapobench-1.0.1",
"title": "DaCapo Benchmark",
"app_version": "9.12-MR1",
"arguments": "tradesoap",
"description": "Java Test: Tradesoap",
"scale": "msec",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 6411,
"raw_values": [
6571,
6363,
6239,
6469
],
"test_run_times": [
16.120000000000000994759830064140260219573974609375,
15.8499999999999996447286321199499070644378662109375,
15.6899999999999995026200849679298698902130126953125,
15.9399999999999995026200849679298698902130126953125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 11110,
"raw_values": [
11143,
11077
],
"test_run_times": [
26.410000000000000142108547152020037174224853515625,
25.870000000000000994759830064140260219573974609375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.070000000000000006661338147750939242541790008544921875,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"440cd1be88623872ff0e6ee9547b1e6a6becc0bb": {
"identifier": "pts\/dav1d-1.6.0",
"title": "dav1d",
"app_version": "0.7.0",
"arguments": "-i summer_nature_4k.ivf",
"description": "Video Input: Summer Nature 4K",
"scale": "FPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 190.912920000000013942553778178989887237548828125,
"raw_values": [
190.923750000000012505552149377763271331787109375,
190.789999999999992041921359486877918243408203125,
191.025000000000005684341886080801486968994140625
],
"min_result": [
"190.51"
],
"max_result": [
"191.36"
],
"test_run_times": [
19.269999999999999573674358543939888477325439453125,
19.089999999999999857891452847979962825775146484375,
19.07000000000000028421709430404007434844970703125
]
}
}
},
"542e395bccf53724e8470b4fcc194dfecbb3512f": {
"identifier": "pts\/dacapobench-1.0.1",
"title": "DaCapo Benchmark",
"app_version": "9.12-MR1",
"arguments": "tradebeans",
"description": "Java Test: Tradebeans",
"scale": "msec",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 3198,
"raw_values": [
3178,
3174,
3230,
3209
],
"test_run_times": [
12.6300000000000007815970093361102044582366943359375,
12.6400000000000005684341886080801486968994140625,
12.71000000000000085265128291212022304534912109375,
12.78999999999999914734871708787977695465087890625
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 4963,
"raw_values": [
4325,
5600
],
"test_run_times": [
19.260000000000001563194018672220408916473388671875,
21.309999999999998721023075631819665431976318359375,
0.070000000000000006661338147750939242541790008544921875,
0.059999999999999997779553950749686919152736663818359375,
0.070000000000000006661338147750939242541790008544921875,
0.08000000000000000166533453693773481063544750213623046875,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"631d0ddd2f71400980ee09a97a6c7dfb5442e5a3": {
"identifier": "pts\/go-benchmark-1.1.4",
"title": "Go Benchmarks",
"arguments": "garbage",
"description": "Test: garbage",
"scale": "Nanoseconds\/Operation",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 702173,
"raw_values": [
705117,
700464,
700939
],
"test_run_times": [
20.300000000000000710542735760100185871124267578125,
20.0799999999999982946974341757595539093017578125,
20.1099999999999994315658113919198513031005859375
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 2275007,
"raw_values": [
2037583,
2512430
],
"test_run_times": [
12.2400000000000002131628207280300557613372802734375,
14.5099999999999997868371792719699442386627197265625,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"8a6ed830709f850982b3fef035ed22d9d2ff9c8f": {
"identifier": "pts\/dav1d-1.6.0",
"title": "dav1d",
"app_version": "0.7.0",
"arguments": "-i chimera_8b_1080p.ivf",
"description": "Video Input: Chimera 1080p",
"scale": "FPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 622.3899999999999863575794734060764312744140625,
"raw_values": [
622.1475000000000363797880709171295166015625,
622.904999999999972715158946812152862548828125,
622.1174999999999499777914024889469146728515625
],
"min_result": [
"621.97"
],
"max_result": [
"623.06"
],
"test_run_times": [
14.75,
14.5099999999999997868371792719699442386627197265625,
14.519999999999999573674358543939888477325439453125
]
}
}
},
"6102c5c5a72a9c836f5047b0e1f6704e0865349f": {
"identifier": "pts\/svt-av1-2.2.1",
"title": "SVT-AV1",
"app_version": "0.8",
"arguments": "-enc-mode 8 -n 320 -i Bosphorus_1920x1080_120fps_420_8bit_YUV.yuv -w 1920 -h 1080",
"description": "Encoder Mode: Enc Mode 8 - Input: 1080p",
"scale": "Frames Per Second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 71.102000000000003865352482534945011138916015625,
"raw_values": [
69.33899999999999863575794734060764312744140625,
71.748999999999995225152815692126750946044921875,
72.2180000000000035242919693700969219207763671875
],
"test_run_times": [
8.53999999999999914734871708787977695465087890625,
7.87999999999999989341858963598497211933135986328125,
8.42999999999999971578290569595992565155029296875
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 11.3499999999999996447286321199499070644378662109375,
"raw_values": [
11.0370000000000008100187187665142118930816650390625,
11.669000000000000483169060316868126392364501953125
],
"test_run_times": [
30.879999999999999005240169935859739780426025390625,
29.39999999999999857891452847979962825775146484375,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"9dba57b02628dcad5b62fa47d8ab13ed936d7880": {
"identifier": "pts\/svt-hevc-1.1.0",
"title": "SVT-HEVC",
"app_version": "1.4.1",
"description": "1080p 8-bit YUV To HEVC Video Encode",
"scale": "Frames Per Second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 131.18999999999999772626324556767940521240234375,
"raw_values": [
130.659999999999996589394868351519107818603515625,
133.18999999999999772626324556767940521240234375,
129.729999999999989768184605054557323455810546875
],
"test_run_times": [
5.20999999999999996447286321199499070644378662109375,
5.12000000000000010658141036401502788066864013671875,
5.21999999999999975131004248396493494510650634765625
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 21.17999999999999971578290569595992565155029296875,
"raw_values": [
20.32000000000000028421709430404007434844970703125,
22.03999999999999914734871708787977695465087890625
],
"test_run_times": [
30.809999999999998721023075631819665431976318359375,
28.57000000000000028421709430404007434844970703125,
0.1000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875,
0.11000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"fbcbf269b8c6e8f3ff08d0d5bc80c365ae4ad893": {
"identifier": "pts\/encode-flac-1.6.0",
"title": "FLAC Audio Encoding",
"app_version": "1.3.2",
"description": "WAV To FLAC",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 8.0570000000000003836930773104541003704071044921875,
"raw_values": [
8.1440000000000001278976924368180334568023681640625,
8.0359999999999995878852132591418921947479248046875,
8.0299999999999993605115378159098327159881591796875,
8.044000000000000483169060316868126392364501953125,
8.0310000000000005826450433232821524143218994140625
],
"test_run_times": [
8.1400000000000005684341886080801486968994140625,
8.03999999999999914734871708787977695465087890625,
8.0299999999999993605115378159098327159881591796875,
8.03999999999999914734871708787977695465087890625,
8.0299999999999993605115378159098327159881591796875
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 11.17999999999999971578290569595992565155029296875,
"raw_values": [
11.6986927986149993330400320701301097869873046875,
10.6575868129729993682985877967439591884613037109375
],
"test_run_times": [
11.699999999999999289457264239899814128875732421875,
10.660000000000000142108547152020037174224853515625,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.070000000000000006661338147750939242541790008544921875
]
}
}
},
"3050891f5d003b57998824da9ce95e0fa23eba58": {
"identifier": "pts\/encode-mp3-1.7.4",
"title": "LAME MP3 Encoding",
"app_version": "3.100",
"description": "WAV To MP3",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 9.1639999999999997015720509807579219341278076171875,
"raw_values": [
9.166000000000000369482222595252096652984619140625,
9.1669999999999998152588887023739516735076904296875,
9.160000000000000142108547152020037174224853515625
],
"test_run_times": [
9.1699999999999999289457264239899814128875732421875,
9.1699999999999999289457264239899814128875732421875,
9.160000000000000142108547152020037174224853515625
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 12.660000000000000142108547152020037174224853515625,
"raw_values": [
12.8733589649199995363915149937383830547332763671875,
12.4454810619350002554028833401389420032501220703125
],
"test_run_times": [
12.8699999999999992184029906638897955417633056640625,
12.4399999999999995026200849679298698902130126953125,
0.070000000000000006661338147750939242541790008544921875,
0.070000000000000006661338147750939242541790008544921875
]
}
}
},
"617870ab716342196c985657b9b14ba066caa04b": {
"identifier": "pts\/primesieve-1.7.0",
"title": "Primesieve",
"app_version": "7.4",
"description": "1e12 Prime Number Generation",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 8.644999999999999573674358543939888477325439453125,
"raw_values": [
8.592999999999999971578290569595992565155029296875,
8.6349999999999997868371792719699442386627197265625,
8.70700000000000073896444519050419330596923828125
],
"test_run_times": [
8.6500000000000003552713678800500929355621337890625,
8.699999999999999289457264239899814128875732421875,
8.769999999999999573674358543939888477325439453125
]
}
}
},
"7041b356c8b5f4cd92ecf820b29e093a9d0af53a": {
"identifier": "pts\/go-benchmark-1.1.4",
"title": "Go Benchmarks",
"arguments": "json",
"description": "Test: json",
"scale": "Nanoseconds\/Operation",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 1764387,
"raw_values": [
1775129,
1762919,
1755113
],
"test_run_times": [
10.4000000000000003552713678800500929355621337890625,
10.2799999999999993605115378159098327159881591796875,
10.1699999999999999289457264239899814128875732421875
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 7628100,
"raw_values": [
7659092,
7597108
],
"test_run_times": [
9.57000000000000028421709430404007434844970703125,
9.4000000000000003552713678800500929355621337890625,
0.070000000000000006661338147750939242541790008544921875,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"6c399a9a1e917499a04e7ad2ef2a1f307fd0464b": {
"identifier": "pts\/dacapobench-1.0.1",
"title": "DaCapo Benchmark",
"app_version": "9.12-MR1",
"arguments": "jython",
"description": "Java Test: Jython",
"scale": "msec",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 4364,
"raw_values": [
4392,
4351,
4349,
4362
],
"test_run_times": [
7.1699999999999999289457264239899814128875732421875,
6.9199999999999999289457264239899814128875732421875,
6.910000000000000142108547152020037174224853515625,
6.94000000000000039079850466805510222911834716796875
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 5167,
"raw_values": [
5082,
5251
],
"test_run_times": [
11.25,
10.160000000000000142108547152020037174224853515625,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"0a5fcb91439ff7d91351f9c8c7cb0161e684c6a1": {
"identifier": "pts\/go-benchmark-1.1.4",
"title": "Go Benchmarks",
"arguments": "http",
"description": "Test: http",
"scale": "Nanoseconds\/Operation",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 7360,
"raw_values": [
7300,
7459,
7321
],
"test_run_times": [
8.78999999999999914734871708787977695465087890625,
8.57000000000000028421709430404007434844970703125,
8.46000000000000085265128291212022304534912109375
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 9785,
"raw_values": [
9845,
9724
],
"test_run_times": [
11.5099999999999997868371792719699442386627197265625,
10.82000000000000028421709430404007434844970703125,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"f808601cbb8fb72f1b0e6763805770f06aa39b2d": {
"identifier": "system\/darktable-1.0.4",
"title": "Darktable",
"app_version": "2.4.1",
"arguments": "bench.SRW output.jpg --core -d opencl -d perf",
"description": "Test: Boat - Acceleration: OpenCL",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 0.38200000000000000621724893790087662637233734130859375,
"raw_values": [
0.38300000000000000710542735760100185871124267578125,
0.38200000000000000621724893790087662637233734130859375,
0.3810000000000000053290705182007513940334320068359375
],
"test_run_times": [
6.5999999999999996447286321199499070644378662109375,
3.1699999999999999289457264239899814128875732421875,
2
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 9.7200000000000006394884621840901672840118408203125,
"raw_values": [
19.062000000000001165290086646564304828643798828125,
0.37800000000000000266453525910037569701671600341796875
],
"test_run_times": [
29.190000000000001278976924368180334568023681640625,
3.430000000000000159872115546022541821002960205078125,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"9844cada39868470c62ab68a6464b249f79bdc81": {
"identifier": "pts\/dacapobench-1.0.1",
"title": "DaCapo Benchmark",
"app_version": "9.12-MR1",
"arguments": "h2",
"description": "Java Test: H2",
"scale": "msec",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 2784,
"raw_values": [
2705,
2791,
2874,
2765
],
"test_run_times": [
6.6500000000000003552713678800500929355621337890625,
6.660000000000000142108547152020037174224853515625,
6.92999999999999971578290569595992565155029296875,
6.730000000000000426325641456060111522674560546875
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 3591,
"raw_values": [
3836,
3346
],
"test_run_times": [
9.21000000000000085265128291212022304534912109375,
7.88999999999999968025576890795491635799407958984375,
0.070000000000000006661338147750939242541790008544921875,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"0cdfc6eea09327dc99a752fd0d9b763e3a56eeea": {
"identifier": "pts\/x264-2.6.1",
"title": "x264",
"app_version": "2019-12-17",
"description": "H.264 Video Encoding",
"scale": "Frames Per Second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 183.340000000000003410605131648480892181396484375,
"raw_values": [
178.19999999999998863131622783839702606201171875,
186.44999999999998863131622783839702606201171875,
185.3600000000000136424205265939235687255859375
],
"test_run_times": [
3.569999999999999840127884453977458178997039794921875,
3.4199999999999999289457264239899814128875732421875,
3.439999999999999946709294817992486059665679931640625
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 41.60000000000000142108547152020037174224853515625,
"raw_values": [
42.11999999999999744204615126363933086395263671875,
41.0799999999999982946974341757595539093017578125
],
"test_run_times": [
14.550000000000000710542735760100185871124267578125,
14.7799999999999993605115378159098327159881591796875,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"8ab5cc33552d96eaf0339bfa848f588594530579": {
"identifier": "pts\/svt-vp9-1.2.2",
"title": "SVT-VP9",
"app_version": "0.1",
"arguments": "-tune 2 -i Bosphorus_1920x1080_120fps_420_8bit_YUV.yuv -w 1920 -h 1080",
"description": "Tuning: VMAF Optimized - Input: Bosphorus 1080p",
"scale": "Frames Per Second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 250.1200000000000045474735088646411895751953125,
"raw_values": [
240,
254.669999999999987494447850622236728668212890625,
252.740000000000009094947017729282379150390625,
253.06000000000000227373675443232059478759765625
],
"test_run_times": [
3.359999999999999875655021241982467472553253173828125,
3.20999999999999996447286321199499070644378662109375,
3.2400000000000002131628207280300557613372802734375,
3.220000000000000195399252334027551114559173583984375
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 67.4599999999999937472239253111183643341064453125,
"raw_values": [
59.10000000000000142108547152020037174224853515625,
75.81999999999999317878973670303821563720703125
],
"test_run_times": [
10.9700000000000006394884621840901672840118408203125,
8.71000000000000085265128291212022304534912109375,
0.1000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"e31cbb65a2fe735d76d52231e2a63c5a6206b338": {
"identifier": "pts\/dav1d-1.6.0",
"title": "dav1d",
"app_version": "0.7.0",
"arguments": "-i summer_nature_1080p.ivf",
"description": "Video Input: Summer Nature 1080p",
"scale": "FPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 693.264374999999972715158946812152862548828125,
"raw_values": [
691.8700000000000045474735088646411895751953125,
692.05999999999994543031789362430572509765625,
695.8631249999999681676854379475116729736328125
],
"min_result": [
"691.28"
],
"max_result": [
"696.83"
],
"test_run_times": [
5.45000000000000017763568394002504646778106689453125,
5.3300000000000000710542735760100185871124267578125,
5.32000000000000028421709430404007434844970703125
]
}
}
},
"af9f382bed62c56b2fa199afad3d687f24b38345": {
"identifier": "pts\/tjbench-1.1.1",
"title": "libjpeg-turbo tjbench",
"app_version": "2.0.2",
"arguments": "decompression-throughput",
"description": "Test: Decompression Throughput",
"scale": "Megapixels\/sec",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 203,
"raw_values": [
203.313456000000002177330316044390201568603515625,
203.3878339999999980136635713279247283935546875,
203.322672000000011394149623811244964599609375
],
"test_run_times": [
6.29999999999999982236431605997495353221893310546875,
6.2400000000000002131628207280300557613372802734375,
6.25
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 191,
"raw_values": [
190.880412000000006855771061964333057403564453125,
191.686299999999988585841492749750614166259765625
],
"test_run_times": [
6.2400000000000002131628207280300557613372802734375,
6.160000000000000142108547152020037174224853515625,
0.070000000000000006661338147750939242541790008544921875,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"72592d0f6d40e912844af3c4d07e2e8a83971bf8": {
"identifier": "pts\/svt-vp9-1.2.2",
"title": "SVT-VP9",
"app_version": "0.1",
"arguments": "-tune 0 -i Bosphorus_1920x1080_120fps_420_8bit_YUV.yuv -w 1920 -h 1080",
"description": "Tuning: Visual Quality Optimized - Input: Bosphorus 1080p",
"scale": "Frames Per Second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 229.509999999999990905052982270717620849609375,
"raw_values": [
230.150000000000005684341886080801486968994140625,
228.05000000000001136868377216160297393798828125,
230.330000000000012505552149377763271331787109375
],
"test_run_times": [
3.430000000000000159872115546022541821002960205078125,
3.45999999999999996447286321199499070644378662109375,
3.399999999999999911182158029987476766109466552734375
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 84.9599999999999937472239253111183643341064453125,
"raw_values": [
88.2099999999999937472239253111183643341064453125,
81.7099999999999937472239253111183643341064453125
],
"test_run_times": [
7.3499999999999996447286321199499070644378662109375,
7.9900000000000002131628207280300557613372802734375,
0.0899999999999999966693309261245303787291049957275390625,
0.070000000000000006661338147750939242541790008544921875,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"519503d976403687edcab94cc1ecd6e6b55a8fee": {
"identifier": "pts\/svt-vp9-1.2.2",
"title": "SVT-VP9",
"app_version": "0.1",
"arguments": "-tune 1 -i Bosphorus_1920x1080_120fps_420_8bit_YUV.yuv -w 1920 -h 1080",
"description": "Tuning: PSNR\/SSIM Optimized - Input: Bosphorus 1080p",
"scale": "Frames Per Second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 255.93999999999999772626324556767940521240234375,
"raw_values": [
254.780000000000001136868377216160297393798828125,
255.969999999999998863131622783839702606201171875,
257.06999999999999317878973670303821563720703125
],
"test_run_times": [
3.20000000000000017763568394002504646778106689453125,
3.20999999999999996447286321199499070644378662109375,
3.180000000000000159872115546022541821002960205078125
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 88.7000000000000028421709430404007434844970703125,
"raw_values": [
80.06000000000000227373675443232059478759765625,
97.340000000000003410605131648480892181396484375
],
"test_run_times": [
8.2400000000000002131628207280300557613372802734375,
6.8499999999999996447286321199499070644378662109375,
0.08000000000000000166533453693773481063544750213623046875,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375,
0.08000000000000000166533453693773481063544750213623046875,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"ae97cf3570ea7bc012da1280adeb7ac75f5a96bf": {
"identifier": "system\/darktable-1.0.4",
"title": "Darktable",
"app_version": "2.4.1",
"arguments": "masskrug.NEF output.jpg --core -d opencl -d perf",
"description": "Test: Masskrug - Acceleration: OpenCL",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 0.457000000000000017319479184152442030608654022216796875,
"raw_values": [
0.455000000000000015543122344752191565930843353271484375,
0.459000000000000019095836023552692495286464691162109375,
0.456000000000000016431300764452316798269748687744140625
],
"test_run_times": [
2.1699999999999999289457264239899814128875732421875,
2.160000000000000142108547152020037174224853515625,
2.180000000000000159872115546022541821002960205078125
]
}
}
},
"9a754e25f4c14f5ded5f98b829dc76048b70b317": {
"identifier": "system\/darktable-1.0.4",
"title": "Darktable",
"app_version": "2.4.1",
"arguments": "server_room.NEF output.jpg --core -d opencl -d perf",
"description": "Test: Server Room - Acceleration: OpenCL",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 0.447000000000000008437694987151189707219600677490234375,
"raw_values": [
0.445000000000000006661338147750939242541790008544921875,
0.446000000000000007549516567451064474880695343017578125,
0.451000000000000011990408665951690636575222015380859375
],
"test_run_times": [
2.089999999999999857891452847979962825775146484375,
2.0800000000000000710542735760100185871124267578125,
2.0800000000000000710542735760100185871124267578125
]
}
}
},
"a35ff7205fdba27121cd1254916ec93054219845": {
"identifier": "system\/darktable-1.0.4",
"title": "Darktable",
"app_version": "2.4.1",
"arguments": "masskrug.NEF output.jpg --core -d perf --disable-opencl",
"description": "Test: Masskrug - Acceleration: CPU-only",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 0.457000000000000017319479184152442030608654022216796875,
"raw_values": [
0.455000000000000015543122344752191565930843353271484375,
0.457000000000000017319479184152442030608654022216796875,
0.458000000000000018207657603852567262947559356689453125
],
"test_run_times": [
1.7800000000000000266453525910037569701671600341796875,
1.770000000000000017763568394002504646778106689453125,
1.7600000000000000088817841970012523233890533447265625
]
}
}
},
"bbb6e18fd25a022da0242b6248fedfed5adbf9ed": {
"identifier": "system\/darktable-1.0.4",
"title": "Darktable",
"app_version": "2.4.1",
"arguments": "server_room.NEF output.jpg --core -d perf --disable-opencl",
"description": "Test: Server Room - Acceleration: CPU-only",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 0.464000000000000023536728122053318656980991363525390625,
"raw_values": [
0.464000000000000023536728122053318656980991363525390625,
0.463000000000000022648549702353193424642086029052734375,
0.465000000000000024424906541753443889319896697998046875
],
"test_run_times": [
1.6999999999999999555910790149937383830547332763671875,
1.6999999999999999555910790149937383830547332763671875,
1.6799999999999999378275106209912337362766265869140625
]
}
}
},
"d2033afa5c5efd8e1ffff64110967f6865101929": {
"identifier": "system\/darktable-1.0.4",
"title": "Darktable",
"app_version": "2.4.1",
"arguments": "server-rack.dng output.jpg --core -d opencl -d perf",
"description": "Test: Server Rack - Acceleration: OpenCL",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 0.25800000000000000710542735760100185871124267578125,
"raw_values": [
0.25700000000000000621724893790087662637233734130859375,
0.25900000000000000799360577730112709105014801025390625,
0.25700000000000000621724893790087662637233734130859375
],
"test_run_times": [
1.5900000000000000799360577730112709105014801025390625,
1.6100000000000000976996261670137755572795867919921875,
1.5900000000000000799360577730112709105014801025390625
]
}
}
},
"8a3c4a1de6da617cf2a53728f1d6114b2f6d988a": {
"identifier": "pts\/novabench-1.0.0",
"title": "Novabench",
"arguments": "--test=gpu",
"description": "Test: GPU",
"scale": "GFLOPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 15270,
"test_run_times": [
1.479999999999999982236431605997495353221893310546875,
1.4099999999999999200639422269887290894985198974609375,
1.3899999999999999023003738329862244427204132080078125
]
}
}
},
"25d0526027cecc95fa17eaa12b53f64f107ebb85": {
"identifier": "system\/darktable-1.0.4",
"title": "Darktable",
"app_version": "2.4.1",
"arguments": "bench.SRW output.jpg --core -d perf --disable-opencl",
"description": "Test: Boat - Acceleration: CPU-only",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 0.37600000000000000088817841970012523233890533447265625,
"raw_values": [
0.37399999999999999911182158029987476766109466552734375,
0.3770000000000000017763568394002504646778106689453125,
0.3770000000000000017763568394002504646778106689453125
],
"test_run_times": [
1.4199999999999999289457264239899814128875732421875,
1.4299999999999999378275106209912337362766265869140625,
1.4199999999999999289457264239899814128875732421875
]
}
}
},
"871aa52ad3a85df8dacb10edfe11aba496cf1349": {
"identifier": "system\/darktable-1.0.4",
"title": "Darktable",
"app_version": "2.4.1",
"arguments": "server-rack.dng output.jpg --core -d perf --disable-opencl",
"description": "Test: Server Rack - Acceleration: CPU-only",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 0.26100000000000000976996261670137755572795867919921875,
"raw_values": [
0.2600000000000000088817841970012523233890533447265625,
0.2600000000000000088817841970012523233890533447265625,
0.262000000000000010658141036401502788066864013671875
],
"test_run_times": [
1.20999999999999996447286321199499070644378662109375,
1.1799999999999999378275106209912337362766265869140625,
1.1999999999999999555910790149937383830547332763671875
]
}
}
},
"cc9f885dc8eb1c602bac8fd2bb116cf809a74e8d": {
"identifier": "pts\/novabench-1.0.0",
"title": "Novabench",
"arguments": "--test=ram",
"description": "Test: RAM",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 37631,
"raw_values": [
35841,
38986,
38065
],
"test_run_times": [
1.25,
0.7800000000000000266453525910037569701671600341796875,
0.7800000000000000266453525910037569701671600341796875
]
}
}
},
"1078cb1dacfa8617071f511bb1388a538b7907d7": {
"identifier": "pts\/novabench-1.0.0",
"title": "Novabench",
"arguments": "--test=ram",
"description": "Test: RAM",
"scale": "RAM Score",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 392,
"raw_values": [
386,
397,
394
],
"test_run_times": [
1.25,
0.7800000000000000266453525910037569701671600341796875,
0.7800000000000000266453525910037569701671600341796875
]
}
}
},
"18027f3e0f9621d1b733699e94a359174d2ef636": {
"identifier": "pts\/blake2-1.2.1",
"title": "BLAKE2",
"app_version": "20170307",
"scale": "Cycles Per Byte",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 9.019999999999999573674358543939888477325439453125,
"raw_values": [
9.07000000000000028421709430404007434844970703125,
8.96000000000000085265128291212022304534912109375,
9.019999999999999573674358543939888477325439453125
],
"test_run_times": [
0.75,
0.7399999999999999911182158029987476766109466552734375,
0.729999999999999982236431605997495353221893310546875
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 9.050000000000000710542735760100185871124267578125,
"raw_values": [
9.07000000000000028421709430404007434844970703125,
9.019999999999999573674358543939888477325439453125
],
"test_run_times": [
0.79000000000000003552713678800500929355621337890625,
0.7800000000000000266453525910037569701671600341796875,
0.059999999999999997779553950749686919152736663818359375,
0.059999999999999997779553950749686919152736663818359375
]
}
}
},
"f726c094606da03a54e516f51b34ac11e428f39a": {
"identifier": "pts\/java-scimark2-1.1.2",
"title": "Java SciMark",
"app_version": "2.0",
"arguments": "TEST_SOR",
"description": "Computational Test: Jacobi Successive Over-Relaxation",
"scale": "Mflops",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 1894,
"raw_values": [
1891.02542395320006107795052230358123779296875,
1892.883099571399952765204943716526031494140625,
1891.02542395320006107795052230358123779296875,
1899.41347100770008182735182344913482666015625
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 2028,
"raw_values": [
2022.519101107400047112605534493923187255859375,
2034.2623283607999837840907275676727294921875
]
}
}
},
"1095c544df0943ac83bee5cd91a8130ac5841575": {
"identifier": "pts\/java-scimark2-1.1.2",
"title": "Java SciMark",
"app_version": "2.0",
"arguments": "TEST_DENSE",
"description": "Computational Test: Dense LU Matrix Factorization",
"scale": "Mflops",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 5455,
"raw_values": [
5417.989330847600285778753459453582763671875,
5442.285417415099800564348697662353515625,
5526.2672075485997993382625281810760498046875,
5431.46052318910005851648747920989990234375
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 6492,
"raw_values": [
6461.2046915736000300967134535312652587890625,
6522.9422163725002974388189613819122314453125
]
}
}
},
"5a09c7e054ba1458261d30e1bc2145ab248d593e": {
"identifier": "pts\/java-scimark2-1.1.2",
"title": "Java SciMark",
"app_version": "2.0",
"arguments": "TEST_SPARSE",
"description": "Computational Test: Sparse Matrix Multiply",
"scale": "Mflops",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 2631,
"raw_values": [
2627.35142544790005558752454817295074462890625,
2635.93770640479988287552259862422943115234375,
2624.0641868093998709809966385364532470703125,
2637.92710731380020661163143813610076904296875
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 2563,
"raw_values": [
2557.50249851279977519880048930644989013671875,
2568.78014903209987096488475799560546875
]
}
}
},
"4fbd0bed179539c2f8218eb648714a753b5d01ad": {
"identifier": "pts\/java-scimark2-1.1.2",
"title": "Java SciMark",
"app_version": "2.0",
"arguments": "TEST_FFT",
"description": "Computational Test: Fast Fourier Transform",
"scale": "Mflops",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 1885,
"raw_values": [
1995.127995956399900023825466632843017578125,
1978.157827643600057854200713336467742919921875,
1584.233167168900081378524191677570343017578125,
1980.966179710899950805469416081905364990234375
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 1894,
"raw_values": [
1914.64392364080003972048871219158172607421875,
1872.50736120600004142033867537975311279296875
]
}
}
},
"3a869a480f4d3efb62355f669f01b0c9e328069d": {
"identifier": "pts\/java-scimark2-1.1.2",
"title": "Java SciMark",
"app_version": "2.0",
"arguments": "TEST_MONTE",
"description": "Computational Test: Monte Carlo",
"scale": "Mflops",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 1648,
"raw_values": [
1637.4256382134999512345530092716217041015625,
1651.910407573999918895424343645572662353515625,
1646.21209251760001279762946069240570068359375,
1654.455843058299933545640669763088226318359375
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 1704,
"raw_values": [
1720.0509940280999217065982520580291748046875,
1686.94708938129997477517463266849517822265625
]
}
}
},
"bedf6de77e99e1b6e2e1992e6d191275f2ed9719": {
"identifier": "pts\/blogbench-1.1.0",
"title": "BlogBench",
"app_version": "1.1",
"arguments": "WRITE",
"description": "Test: Write",
"scale": "Final Score",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 701,
"raw_values": [
710,
714,
678
]
},
"gogz-win81-00vm-ccta-cpu-massive-8jan2021": {
"value": 181,
"raw_values": [
186,
175
]
}
}
}
}
}