VMware testing on Microsoft Windows 10 Pro Build 18362 via the Phoronix Test Suite.
Compare your own system(s) to this result file with the
Phoronix Test Suite by running the command:
phoronix-test-suite benchmark 2101102-SP-2009246NE56
{
"title": "ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020",
"last_modified": "2021-01-10 15:51:01",
"description": "VMware testing on Microsoft Windows 10 Pro Build 18362 via the Phoronix Test Suite.",
"reference_id": "2009246-NE-THREADR1W39",
"systems": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"identifier": "ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020",
"hardware": {
"Processor": "AMD Ryzen Threadripper 3960X 24-Core @ 3.80GHz (24 Cores \/ 48 Threads)",
"Motherboard": "MSI Creator TRX40 (MS-7C59) (1.50 BIOS)",
"Memory": "4 x 16384 MB 1800MHz CMT64GX4M4K3600C16",
"Disk": "932GB Samsung SSD 860 EVO 1TB + 2 x 466GB Force MP600 + 4 x 477GB Samsung SSD 970 PRO 512GB + 57GB SanDisk Cruzer Blade USB + 477GB Samsung SSD 860 PRO 512GB + 10GB USB DISK CD USB + 932GB Force MP600 + 932GB JMicron Generic Disk + 932GB Inateck ASM1153E Disk + 954GB Inateck Disk + 954GB Samsung SSD 860 PRO 1TB + 1863GB Inateck ASM1153E Disk",
"Graphics": "AMD Radeon VII 4GB",
"Network": "Intel I211 Gigabit Connection + Intel Wi-Fi 6 AX200 160MHz"
},
"software": {
"OS": "Microsoft Windows 10 Enterprise LTSC Build 17763",
"Kernel": "10.0 (x86_64)",
"Display Driver": "27.20.1027.2005",
"OpenCL": "OpenCL 2.1 AMD-APP (3075.13)",
"Compiler": "GCC 8.3.0",
"File-System": "NTFS",
"Screen Resolution": "1920x1080"
},
"user": "SpingBasic",
"timestamp": "2020-09-22 01:01:05",
"data": {
"java": "OpenJDK Runtime Environment 18.9 (build 11.0.6+10-LTS)",
"python": "Python 3.7.9",
"security": "__user pointer sanitization: Disabled + Retpoline: Full + IBPB: Always + STIBP: Enabled"
}
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"identifier": "gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021",
"hardware": {
"Processor": "AMD Ryzen Threadripper 3960X 24-Core @ 3.80GHz (8 Cores)",
"Motherboard": "Intel 440BX",
"Memory": "1 x 16384 MB 0MHz VMW-16384MB",
"Disk": "512GB VMware Virtual S",
"Graphics": "VMware SVGA 3D",
"Network": "TAP-Windows V9 + Intel Dual Band Wireless-AC 8265 + Bluetooth Device (Personal Area )"
},
"software": {
"OS": "Microsoft Windows 10 Pro Build 18362",
"Kernel": "10.0 (x86_64)",
"Display Driver": "8.16.7.5",
"Compiler": "GCC 7.1.0",
"File-System": "NTFS",
"Screen Resolution": "919x968",
"System Layer": "VMware"
},
"user": "Chi",
"timestamp": "2021-01-09 02:31:01"
}
},
"results": {
"ffba90bc6e56b93d1e0102003a1dd9e41a8e227c": {
"identifier": "pts\/compress-7zip-1.7.1",
"title": "7-Zip Compression",
"app_version": "16.02",
"description": "Compress Speed Test",
"scale": "MIPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 106201,
"raw_values": [
100372,
114022,
99908,
100340,
99944,
115148,
112697,
115030,
99819,
99186,
99538,
116506,
101407,
102314,
116791
],
"test_run_times": [
48.02000000000000312638803734444081783294677734375,
44.07000000000000028421709430404007434844970703125,
48.43999999999999772626324556767940521240234375,
47.530000000000001136868377216160297393798828125,
48.11999999999999744204615126363933086395263671875,
44.92999999999999971578290569595992565155029296875,
45.469999999999998863131622783839702606201171875,
44.5,
48.21000000000000085265128291212022304534912109375,
47.6700000000000017053025658242404460906982421875,
48.13000000000000255795384873636066913604736328125,
42.969999999999998863131622783839702606201171875,
47.99000000000000198951966012828052043914794921875,
47.7999999999999971578290569595992565155029296875,
43.78999999999999914734871708787977695465087890625
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 26892,
"raw_values": [
25132,
28652
],
"test_run_times": [
28.07000000000000028421709430404007434844970703125,
28.07000000000000028421709430404007434844970703125,
0.1000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"44c9dca2ba5d3e14ac16e945e792a5f4d0dee1e6": {
"identifier": "pts\/apache-1.7.2",
"title": "Apache Benchmark",
"app_version": "2.4.29",
"description": "Static Web Page Serving",
"scale": "Requests Per Second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 6503.829999999999927240423858165740966796875,
"raw_values": [
6439.5399999999999636202119290828704833984375,
6537.239999999999781721271574497222900390625,
6534.6999999999998181010596454143524169921875
],
"test_run_times": [
155.479999999999989768184605054557323455810546875,
153.18000000000000682121026329696178436279296875,
153.25
]
}
}
},
"fc8b6de93b7723b344a80057c04367a883bac1ec": {
"identifier": "pts\/asmfish-1.1.1",
"title": "asmFish",
"app_version": "2018-07-23",
"description": "1024 Hash Memory, 26 Depth",
"scale": "Nodes\/second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 72340992,
"raw_values": [
71280014,
71889170,
73853791
],
"test_run_times": [
152.240000000000009094947017729282379150390625,
129.900000000000005684341886080801486968994140625,
136.530000000000001136868377216160297393798828125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 12354671,
"raw_values": [
12102663,
12606679
],
"test_run_times": [
231.6100000000000136424205265939235687255859375,
221.80000000000001136868377216160297393798828125,
0.309999999999999997779553950749686919152736663818359375,
0.11999999999999999555910790149937383830547332763671875
]
}
}
},
"18027f3e0f9621d1b733699e94a359174d2ef636": {
"identifier": "pts\/blake2-1.2.1",
"title": "BLAKE2",
"app_version": "20170307",
"scale": "Cycles Per Byte",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 9.019999999999999573674358543939888477325439453125,
"raw_values": [
9.07000000000000028421709430404007434844970703125,
8.96000000000000085265128291212022304534912109375,
9.019999999999999573674358543939888477325439453125
],
"test_run_times": [
0.75,
0.7399999999999999911182158029987476766109466552734375,
0.729999999999999982236431605997495353221893310546875
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 9.07000000000000028421709430404007434844970703125,
"raw_values": [
9.07000000000000028421709430404007434844970703125,
9.07000000000000028421709430404007434844970703125
],
"test_run_times": [
0.82999999999999996003197111349436454474925994873046875,
0.8000000000000000444089209850062616169452667236328125,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"e1f890b694d827aa7548fefbbe292cf60939acc7": {
"identifier": "pts\/blender-1.7.1",
"title": "Blender",
"app_version": "2.82",
"arguments": "-b ..\/bmw27_gpu.blend -o output.test -x 1 -F JPEG -f 1 OPENCL",
"description": "Blend File: BMW27 - Compute: OpenCL",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 378,
"raw_values": [
389.29000000000002046363078989088535308837890625,
372.18999999999999772626324556767940521240234375,
372.1200000000000045474735088646411895751953125
],
"test_run_times": [
412.32999999999998408384271897375583648681640625,
394.240000000000009094947017729282379150390625,
394.259999999999990905052982270717620849609375
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 414,
"raw_values": [
411.990000000000009094947017729282379150390625,
415.42000000000001591615728102624416351318359375
],
"test_run_times": [
416.8999999999999772626324556767940521240234375,
416.83999999999997498889570124447345733642578125,
0.1000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"da537b151d630d3406022e009f401b1b88c03aa6": {
"identifier": "pts\/blender-1.7.1",
"title": "Blender",
"app_version": "2.82",
"arguments": "-b ..\/bmw27_gpu.blend -o output.test -x 1 -F JPEG -f 1 NONE",
"description": "Blend File: BMW27 - Compute: CPU-Only",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 69.7699999999999960209606797434389591217041015625,
"raw_values": [
69.469999999999998863131622783839702606201171875,
69.3900000000000005684341886080801486968994140625,
70.4500000000000028421709430404007434844970703125
],
"test_run_times": [
91.4899999999999948840923025272786617279052734375,
91.340000000000003410605131648480892181396484375,
92.400000000000005684341886080801486968994140625
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 317.759999999999990905052982270717620849609375,
"raw_values": [
314.66000000000002501110429875552654266357421875,
320.8500000000000227373675443232059478759765625
],
"test_run_times": [
316.08999999999997498889570124447345733642578125,
322.3700000000000045474735088646411895751953125,
0.13000000000000000444089209850062616169452667236328125,
0.2200000000000000011102230246251565404236316680908203125
]
}
}
},
"ac988c8aa133da34c672fdc3f1ac8d4017b1c759": {
"identifier": "pts\/blender-1.7.1",
"title": "Blender",
"app_version": "2.82",
"arguments": "-b ..\/classroom_gpu.blend -o output.test -x 1 -F JPEG -f 1 OPENCL",
"description": "Blend File: Classroom - Compute: OpenCL",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 303,
"raw_values": [
299.93000000000000682121026329696178436279296875,
308.029999999999972715158946812152862548828125,
302.220000000000027284841053187847137451171875
],
"test_run_times": [
322.08999999999997498889570124447345733642578125,
330.220000000000027284841053187847137451171875,
324.3799999999999954525264911353588104248046875
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 903,
"raw_values": [
906.5900000000000318323145620524883270263671875,
899.3200000000000500222085975110530853271484375
],
"test_run_times": [
908.740000000000009094947017729282379150390625,
900.80999999999994543031789362430572509765625,
0.11999999999999999555910790149937383830547332763671875,
0.11999999999999999555910790149937383830547332763671875
]
}
}
},
"7bfe52f038c5b130d71757b1f9876e09fb4490b1": {
"identifier": "pts\/blender-1.7.1",
"title": "Blender",
"app_version": "2.82",
"arguments": "-b ..\/fishy_cat_gpu.blend -o output.test -x 1 -F JPEG -f 1 OPENCL",
"description": "Blend File: Fishy Cat - Compute: OpenCL",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 1209,
"raw_values": [
1224.549999999999954525264911353588104248046875,
1197.319999999999936335370875895023345947265625,
1205.339999999999918145476840436458587646484375
],
"test_run_times": [
1247.2100000000000363797880709171295166015625,
1262.009999999999990905052982270717620849609375,
1270
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 1193,
"raw_values": [
1199.90000000000009094947017729282379150390625,
1186.4600000000000363797880709171295166015625
],
"test_run_times": [
1202.200000000000045474735088646411895751953125,
1188.509999999999990905052982270717620849609375,
0.1000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"964cbe068efc3f8ad6e279d944e771f6d9d70e13": {
"identifier": "pts\/blender-1.7.1",
"title": "Blender",
"app_version": "2.82",
"arguments": "-b ..\/barbershop_interior_gpu.blend -o output.test -x 1 -F JPEG -f 1 OPENCL",
"description": "Blend File: Barbershop - Compute: OpenCL",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 614,
"raw_values": [
666.200000000000045474735088646411895751953125,
613.6799999999999499777914024889469146728515625,
601.8200000000000500222085975110530853271484375,
603.6299999999999954525264911353588104248046875,
616.4099999999999681676854379475116729736328125,
619.4600000000000363797880709171295166015625,
602.5700000000000500222085975110530853271484375,
602.3300000000000409272615797817707061767578125,
600.1499999999999772626324556767940521240234375
],
"test_run_times": [
737.950000000000045474735088646411895751953125,
684.8200000000000500222085975110530853271484375,
672.8200000000000500222085975110530853271484375,
674.48000000000001818989403545856475830078125,
687.1499999999999772626324556767940521240234375,
690.44000000000005456968210637569427490234375,
673.200000000000045474735088646411895751953125,
673.4199999999999590727384202182292938232421875,
671.1000000000000227373675443232059478759765625
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 1748,
"raw_values": [
1737.75,
1759.09999999999990905052982270717620849609375
],
"test_run_times": [
1761.76999999999998181010596454143524169921875,
1780.84999999999990905052982270717620849609375,
0.200000000000000011102230246251565404236316680908203125,
0.1000000000000000055511151231257827021181583404541015625
]
}
}
},
"9a743b764f90241ba42dc43a928cd9361eaed2c9": {
"identifier": "pts\/blender-1.7.1",
"title": "Blender",
"app_version": "2.82",
"arguments": "-b ..\/classroom_gpu.blend -o output.test -x 1 -F JPEG -f 1 NONE",
"description": "Blend File: Classroom - Compute: CPU-Only",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 188,
"raw_values": [
188.969999999999998863131622783839702606201171875,
186.6200000000000045474735088646411895751953125,
188.979999999999989768184605054557323455810546875
],
"test_run_times": [
253.06999999999999317878973670303821563720703125,
208.669999999999987494447850622236728668212890625,
211.020000000000010231815394945442676544189453125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 887,
"raw_values": [
896.51999999999998181010596454143524169921875,
877.4199999999999590727384202182292938232421875
],
"test_run_times": [
902.2899999999999636202119290828704833984375,
879.2899999999999636202119290828704833984375,
0.1000000000000000055511151231257827021181583404541015625,
0.1000000000000000055511151231257827021181583404541015625
]
}
}
},
"ea8b3eb302364079e21158085fac0b5f9d83b2f0": {
"identifier": "pts\/blender-1.7.1",
"title": "Blender",
"app_version": "2.82",
"arguments": "-b ..\/fishy_cat_gpu.blend -o output.test -x 1 -F JPEG -f 1 NONE",
"description": "Blend File: Fishy Cat - Compute: CPU-Only",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 104,
"raw_values": [
104.5499999999999971578290569595992565155029296875,
102.8299999999999982946974341757595539093017578125,
106.099999999999994315658113919198513031005859375
],
"test_run_times": [
127.06999999999999317878973670303821563720703125,
125.31999999999999317878973670303821563720703125,
128.56000000000000227373675443232059478759765625
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 529,
"raw_values": [
491.73000000000001818989403545856475830078125,
567.0800000000000409272615797817707061767578125
],
"test_run_times": [
494.1100000000000136424205265939235687255859375,
572.259999999999990905052982270717620849609375,
0.11000000000000000055511151231257827021181583404541015625,
0.1000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.1000000000000000055511151231257827021181583404541015625
]
}
}
},
"c9883de3b5843918d50bb37d38bd72266ba8f906": {
"identifier": "pts\/blender-1.7.1",
"title": "Blender",
"app_version": "2.82",
"arguments": "-b ..\/barbershop_interior_gpu.blend -o output.test -x 1 -F JPEG -f 1 NONE",
"description": "Blend File: Barbershop - Compute: CPU-Only",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 303,
"raw_values": [
300.69999999999998863131622783839702606201171875,
303.529999999999972715158946812152862548828125,
304.41000000000002501110429875552654266357421875
],
"test_run_times": [
330.240000000000009094947017729282379150390625,
332.240000000000009094947017729282379150390625,
333.279999999999972715158946812152862548828125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 1693,
"raw_values": [
1714.799999999999954525264911353588104248046875,
1670.48000000000001818989403545856475830078125
],
"test_run_times": [
1751.6300000000001091393642127513885498046875,
1699.579999999999927240423858165740966796875,
0.11999999999999999555910790149937383830547332763671875,
0.1000000000000000055511151231257827021181583404541015625
]
}
}
},
"b399a6747a6ba64f46124b5671250713d00f6fcb": {
"identifier": "pts\/blender-1.7.1",
"title": "Blender",
"app_version": "2.82",
"arguments": "-b ..\/pavillon_barcelone_gpu.blend -o output.test -x 1 -F JPEG -f 1 OPENCL",
"description": "Blend File: Pabellon Barcelona - Compute: OpenCL",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 1211,
"raw_values": [
1215.140000000000100044417195022106170654296875,
1203.799999999999954525264911353588104248046875,
1214.779999999999972715158946812152862548828125
],
"test_run_times": [
1237.640000000000100044417195022106170654296875,
1268.30999999999994543031789362430572509765625,
1279.359999999999899955582804977893829345703125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 2002,
"raw_values": [
1999.450000000000045474735088646411895751953125,
2005.0399999999999636202119290828704833984375
],
"test_run_times": [
2001.930000000000063664629124104976654052734375,
2007.05999999999994543031789362430572509765625,
0.1000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"76cbd4b69aae2d0dcb99e18956775552f2395171": {
"identifier": "pts\/blender-1.7.1",
"title": "Blender",
"app_version": "2.82",
"arguments": "-b ..\/pavillon_barcelone_gpu.blend -o output.test -x 1 -F JPEG -f 1 NONE",
"description": "Blend File: Pabellon Barcelona - Compute: CPU-Only",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 225,
"raw_values": [
224.020000000000010231815394945442676544189453125,
224.400000000000005684341886080801486968994140625,
226.81000000000000227373675443232059478759765625
],
"test_run_times": [
288.48000000000001818989403545856475830078125,
246.8700000000000045474735088646411895751953125,
249.25
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 1206,
"raw_values": [
1207.2899999999999636202119290828704833984375,
1203.8699999999998908606357872486114501953125
],
"test_run_times": [
1209.390000000000100044417195022106170654296875,
1207.1199999999998908606357872486114501953125,
0.4699999999999999733546474089962430298328399658203125,
0.440000000000000002220446049250313080847263336181640625
]
}
}
},
"d26eb6b47a8b6e5be191497d72172b008d41f9c1": {
"identifier": "pts\/blogbench-1.1.0",
"title": "BlogBench",
"app_version": "1.1",
"arguments": "READ",
"description": "Test: Read",
"scale": "Final Score",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 21043,
"raw_values": [
21318,
21014,
20796
],
"test_run_times": [
368.68999999999999772626324556767940521240234375,
321.54000000000002046363078989088535308837890625,
320.6000000000000227373675443232059478759765625
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 4289,
"raw_values": [
4784,
3794
],
"test_run_times": [
311.08999999999997498889570124447345733642578125,
310.990000000000009094947017729282379150390625,
0.11000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625,
0.11000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"bedf6de77e99e1b6e2e1992e6d191275f2ed9719": {
"identifier": "pts\/blogbench-1.1.0",
"title": "BlogBench",
"app_version": "1.1",
"arguments": "WRITE",
"description": "Test: Write",
"scale": "Final Score",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 701,
"raw_values": [
710,
714,
678
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 135,
"raw_values": [
141,
129
]
}
}
},
"32f6380f8c703d3d84bd215727a68137f54f31cb": {
"identifier": "pts\/c-ray-1.2.0",
"title": "C-Ray",
"app_version": "1.1",
"description": "Total Time - 4K, 16 Rays Per Pixel",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 22.385000000000001563194018672220408916473388671875,
"raw_values": [
22.5150000000000005684341886080801486968994140625,
22.34100000000000108002495835535228252410888671875,
22.29899999999999948840923025272786617279052734375
],
"test_run_times": [
24.910000000000000142108547152020037174224853515625,
24.690000000000001278976924368180334568023681640625,
24.629999999999999005240169935859739780426025390625
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 90.590000000000003410605131648480892181396484375,
"raw_values": [
90.7530000000000001136868377216160297393798828125,
90.4249999999999971578290569595992565155029296875
],
"test_run_times": [
93.280000000000001136868377216160297393798828125,
92.93999999999999772626324556767940521240234375,
0.1700000000000000122124532708767219446599483489990234375,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"52ee8b1c2f74fac119a6f4d3ec07a99d9a0a5cde": {
"identifier": "pts\/cachebench-1.1.2",
"title": "CacheBench",
"arguments": "-r",
"description": "Test: Read",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 10689,
"raw_values": [
10674.497442800000499119050800800323486328125,
10694.83450413299942738376557826995849609375,
10698.55232353300016256980597972869873046875
],
"min_result": [
"10642.59"
],
"max_result": [
"10718.22"
],
"test_run_times": [
125.090000000000003410605131648480892181396484375,
125.0799999999999982946974341757595539093017578125,
125.0799999999999982946974341757595539093017578125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 10266,
"raw_values": [
10275.9238425709991133771836757659912109375,
10256.967487380999955348670482635498046875
],
"test_run_times": [
125.18000000000000682121026329696178436279296875,
125.159999999999996589394868351519107818603515625,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"286caffe46f313fa6981b18b518b8f2ce68ade60": {
"identifier": "pts\/cachebench-1.1.2",
"title": "CacheBench",
"arguments": "-w",
"description": "Test: Write",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 31500,
"raw_values": [
31529.67021100000056321732699871063232421875,
31436.36740833300063968636095523834228515625,
31534.79451213300126255489885807037353515625
],
"min_result": [
"30621.71"
],
"max_result": [
"32079.18"
],
"test_run_times": [
125.0799999999999982946974341757595539093017578125,
125.0799999999999982946974341757595539093017578125,
125.0799999999999982946974341757595539093017578125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 14963,
"raw_values": [
14944.98527390500021283514797687530517578125,
14981.72496581000086735002696514129638671875
],
"test_run_times": [
125.159999999999996589394868351519107818603515625,
125.1700000000000017053025658242404460906982421875,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"f8b4f2829bbd82baf42a96ae5d0db82122541a34": {
"identifier": "pts\/cachebench-1.1.2",
"title": "CacheBench",
"arguments": "-b",
"description": "Test: Read \/ Modify \/ Write",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 31607,
"raw_values": [
31587.03443219999826396815478801727294921875,
31611.32414913300090120173990726470947265625,
31623.080722667000372894108295440673828125
],
"min_result": [
"31171.78"
],
"max_result": [
"31896.35"
],
"test_run_times": [
125.0799999999999982946974341757595539093017578125,
125.0799999999999982946974341757595539093017578125,
125.0799999999999982946974341757595539093017578125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 56817,
"raw_values": [
56700.704642523996881209313869476318359375,
56933.618235904999892227351665496826171875
],
"test_run_times": [
125.18999999999999772626324556767940521240234375,
125.2300000000000039790393202565610408782958984375,
0.1000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"2d198da79dae8a42c45d979302d1803b4c52bdd5": {
"identifier": "pts\/v-ray-1.2.1",
"title": "Chaos Group V-RAY",
"app_version": "4.10.07",
"arguments": "-m vray",
"description": "Mode: CPU",
"scale": "Ksamples",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 34538,
"raw_values": [
34100,
34940,
34573
],
"test_run_times": [
72.099999999999994315658113919198513031005859375,
71.7000000000000028421709430404007434844970703125,
71.780000000000001136868377216160297393798828125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 6189,
"raw_values": [
6073,
6304
],
"test_run_times": [
91.0100000000000051159076974727213382720947265625,
86.469999999999998863131622783839702606201171875,
0.1000000000000000055511151231257827021181583404541015625,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"07d46f075f5bd9e04f1eb7212d0c6c9fc50099af": {
"identifier": "pts\/cpuminer-opt-1.3.0",
"title": "Cpuminer-Opt",
"app_version": "3.8.8.1",
"arguments": "-a m7m",
"description": "Algorithm: m7m",
"scale": "kH\/s - Hash Speed",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 1101.30999999999994543031789362430572509765625,
"raw_values": [
1122.25,
1087.40000000000009094947017729282379150390625,
1094.279999999999972715158946812152862548828125
],
"test_run_times": [
35.8599999999999994315658113919198513031005859375,
35.75,
36.14999999999999857891452847979962825775146484375
]
}
}
},
"9fb6b1f63bd87c00364226a39e5d2bb685cd0d83": {
"identifier": "pts\/cpuminer-opt-1.3.0",
"title": "Cpuminer-Opt",
"app_version": "3.8.8.1",
"arguments": "-a deep",
"description": "Algorithm: deep",
"scale": "kH\/s - Hash Speed",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 19220.889999999999417923390865325927734375,
"raw_values": [
21610,
6291.4600000000000363797880709171295166015625,
22100,
21480,
22030,
21520,
41.93999999999999772626324556767940521240234375,
21390,
22070,
21650,
21960,
21740,
21470,
21170,
21790
],
"test_run_times": [
34.86999999999999744204615126363933086395263671875,
35.32000000000000028421709430404007434844970703125,
34.99000000000000198951966012828052043914794921875,
34.99000000000000198951966012828052043914794921875,
34.86999999999999744204615126363933086395263671875,
34.659999999999996589394868351519107818603515625,
35.42999999999999971578290569595992565155029296875,
34.77000000000000312638803734444081783294677734375,
35,
35.1400000000000005684341886080801486968994140625,
34.99000000000000198951966012828052043914794921875,
34.5499999999999971578290569595992565155029296875,
35.43999999999999772626324556767940521240234375
]
}
}
},
"fd8a26d206b1ff7e664749268590d8ee61acc1ad": {
"identifier": "pts\/cpuminer-opt-1.3.0",
"title": "Cpuminer-Opt",
"app_version": "3.8.8.1",
"arguments": "-a lbry",
"description": "Algorithm: lbry",
"scale": "kH\/s - Hash Speed",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 66793,
"raw_values": [
68200,
67250,
64930
],
"test_run_times": [
30.809999999999998721023075631819665431976318359375,
30.8299999999999982946974341757595539093017578125,
31.120000000000000994759830064140260219573974609375
]
}
}
},
"c75ad335f359edae4899c29f49cde43aeb58e789": {
"identifier": "pts\/cpuminer-opt-1.3.0",
"title": "Cpuminer-Opt",
"app_version": "3.8.8.1",
"arguments": "-a skein",
"description": "Algorithm: skein",
"scale": "kH\/s - Hash Speed",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 72514.289999999993597157299518585205078125,
"raw_values": [
77490,
2621.4499999999998181010596454143524169921875,
76860,
80840,
78870,
78940,
78610,
76870,
77420,
78400,
80920,
82330
],
"test_run_times": [
30.969999999999998863131622783839702606201171875,
31.120000000000000994759830064140260219573974609375,
31.199999999999999289457264239899814128875732421875,
30.989999999999998436805981327779591083526611328125,
30.980000000000000426325641456060111522674560546875,
30.989999999999998436805981327779591083526611328125,
30.75,
31.14999999999999857891452847979962825775146484375,
30.660000000000000142108547152020037174224853515625,
31.300000000000000710542735760100185871124267578125,
31.03999999999999914734871708787977695465087890625
]
}
}
},
"58eea68af3edce96bd553fc212c717811fc393b7": {
"identifier": "pts\/cpuminer-opt-1.3.0",
"title": "Cpuminer-Opt",
"app_version": "3.8.8.1",
"arguments": "-a myr-gr",
"description": "Algorithm: myr-gr",
"scale": "kH\/s - Hash Speed",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 7686.97999999999956344254314899444580078125,
"raw_values": [
8768.379999999999199644662439823150634765625,
7645.7200000000002546585164964199066162109375,
7566.399999999999636202119290828704833984375,
7579.510000000000218278728425502777099609375,
7629.6800000000002910383045673370361328125,
7654.7200000000002546585164964199066162109375,
7593.850000000000363797880709171295166015625,
7573.65999999999985448084771633148193359375,
7622.3199999999997089616954326629638671875,
7609.2899999999999636202119290828704833984375,
7543.4499999999998181010596454143524169921875,
7535.7100000000000363797880709171295166015625,
7665.27000000000043655745685100555419921875,
7629.760000000000218278728425502777099609375
],
"test_run_times": [
32.0799999999999982946974341757595539093017578125,
31.489999999999998436805981327779591083526611328125,
32.47999999999999687361196265555918216705322265625,
31.620000000000000994759830064140260219573974609375,
32.3599999999999994315658113919198513031005859375,
31.60000000000000142108547152020037174224853515625,
32.3599999999999994315658113919198513031005859375,
31.730000000000000426325641456060111522674560546875,
32.219999999999998863131622783839702606201171875,
31.440000000000001278976924368180334568023681640625,
31.530000000000001136868377216160297393798828125,
31.35000000000000142108547152020037174224853515625,
31.620000000000000994759830064140260219573974609375,
31.550000000000000710542735760100185871124267578125
]
}
}
},
"055f9243aefb2de6118ed0c85a05c6ca0ede0071": {
"identifier": "pts\/cpuminer-opt-1.3.0",
"title": "Cpuminer-Opt",
"app_version": "3.8.8.1",
"arguments": "-a sha256t",
"description": "Algorithm: sha256t",
"scale": "kH\/s - Hash Speed",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 168613,
"raw_values": [
169900,
168850,
167090
],
"test_run_times": [
30.4200000000000017053025658242404460906982421875,
30.870000000000000994759830064140260219573974609375,
30.989999999999998436805981327779591083526611328125
]
}
}
},
"6b1815d46eb146239522c2b1521de3ae4512fb73": {
"identifier": "pts\/crafty-1.4.4",
"title": "Crafty",
"app_version": "25.2",
"description": "Elapsed Time",
"scale": "Nodes Per Second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 8204260,
"raw_values": [
8190463,
8219677,
8202639
],
"test_run_times": [
23.92999999999999971578290569595992565155029296875,
23.949999999999999289457264239899814128875732421875,
24.010000000000001563194018672220408916473388671875
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 7382312,
"raw_values": [
7339209,
7425414
],
"test_run_times": [
27.089999999999999857891452847979962825775146484375,
26.940000000000001278976924368180334568023681640625,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"9844cada39868470c62ab68a6464b249f79bdc81": {
"identifier": "pts\/dacapobench-1.0.1",
"title": "DaCapo Benchmark",
"app_version": "9.12-MR1",
"arguments": "h2",
"description": "Java Test: H2",
"scale": "msec",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 2784,
"raw_values": [
2705,
2791,
2874,
2765
],
"test_run_times": [
6.6500000000000003552713678800500929355621337890625,
6.660000000000000142108547152020037174224853515625,
6.92999999999999971578290569595992565155029296875,
6.730000000000000426325641456060111522674560546875
]
}
}
},
"6c399a9a1e917499a04e7ad2ef2a1f307fd0464b": {
"identifier": "pts\/dacapobench-1.0.1",
"title": "DaCapo Benchmark",
"app_version": "9.12-MR1",
"arguments": "jython",
"description": "Java Test: Jython",
"scale": "msec",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 4364,
"raw_values": [
4392,
4351,
4349,
4362
],
"test_run_times": [
7.1699999999999999289457264239899814128875732421875,
6.9199999999999999289457264239899814128875732421875,
6.910000000000000142108547152020037174224853515625,
6.94000000000000039079850466805510222911834716796875
]
}
}
},
"f7d9afeee0ad6d3ac57fa2ea86ed8a7dc9b45df4": {
"identifier": "pts\/dacapobench-1.0.1",
"title": "DaCapo Benchmark",
"app_version": "9.12-MR1",
"arguments": "tradesoap",
"description": "Java Test: Tradesoap",
"scale": "msec",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 6411,
"raw_values": [
6571,
6363,
6239,
6469
],
"test_run_times": [
16.120000000000000994759830064140260219573974609375,
15.8499999999999996447286321199499070644378662109375,
15.6899999999999995026200849679298698902130126953125,
15.9399999999999995026200849679298698902130126953125
]
}
}
},
"542e395bccf53724e8470b4fcc194dfecbb3512f": {
"identifier": "pts\/dacapobench-1.0.1",
"title": "DaCapo Benchmark",
"app_version": "9.12-MR1",
"arguments": "tradebeans",
"description": "Java Test: Tradebeans",
"scale": "msec",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 3198,
"raw_values": [
3178,
3174,
3230,
3209
],
"test_run_times": [
12.6300000000000007815970093361102044582366943359375,
12.6400000000000005684341886080801486968994140625,
12.71000000000000085265128291212022304534912109375,
12.78999999999999914734871708787977695465087890625
]
}
}
},
"f808601cbb8fb72f1b0e6763805770f06aa39b2d": {
"identifier": "system\/darktable-1.0.4",
"title": "Darktable",
"app_version": "2.4.1",
"arguments": "bench.SRW output.jpg --core -d opencl -d perf",
"description": "Test: Boat - Acceleration: OpenCL",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 0.38200000000000000621724893790087662637233734130859375,
"raw_values": [
0.38300000000000000710542735760100185871124267578125,
0.38200000000000000621724893790087662637233734130859375,
0.3810000000000000053290705182007513940334320068359375
],
"test_run_times": [
6.5999999999999996447286321199499070644378662109375,
3.1699999999999999289457264239899814128875732421875,
2
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 0.61999999999999999555910790149937383830547332763671875,
"raw_values": [
0.64800000000000002042810365310288034379482269287109375,
0.58199999999999996180832795289461500942707061767578125
],
"test_run_times": [
6.11000000000000031974423109204508364200592041015625,
1.939999999999999946709294817992486059665679931640625,
0.1000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"25d0526027cecc95fa17eaa12b53f64f107ebb85": {
"identifier": "system\/darktable-1.0.4",
"title": "Darktable",
"app_version": "2.4.1",
"arguments": "bench.SRW output.jpg --core -d perf --disable-opencl",
"description": "Test: Boat - Acceleration: CPU-only",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 0.37600000000000000088817841970012523233890533447265625,
"raw_values": [
0.37399999999999999911182158029987476766109466552734375,
0.3770000000000000017763568394002504646778106689453125,
0.3770000000000000017763568394002504646778106689453125
],
"test_run_times": [
1.4199999999999999289457264239899814128875732421875,
1.4299999999999999378275106209912337362766265869140625,
1.4199999999999999289457264239899814128875732421875
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 0.59999999999999997779553950749686919152736663818359375,
"raw_values": [
0.6350000000000000088817841970012523233890533447265625,
0.5620000000000000550670620214077644050121307373046875
],
"test_run_times": [
2.1699999999999999289457264239899814128875732421875,
1.9499999999999999555910790149937383830547332763671875,
0.1000000000000000055511151231257827021181583404541015625,
0.08000000000000000166533453693773481063544750213623046875,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"ae97cf3570ea7bc012da1280adeb7ac75f5a96bf": {
"identifier": "system\/darktable-1.0.4",
"title": "Darktable",
"app_version": "2.4.1",
"arguments": "masskrug.NEF output.jpg --core -d opencl -d perf",
"description": "Test: Masskrug - Acceleration: OpenCL",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 0.457000000000000017319479184152442030608654022216796875,
"raw_values": [
0.455000000000000015543122344752191565930843353271484375,
0.459000000000000019095836023552692495286464691162109375,
0.456000000000000016431300764452316798269748687744140625
],
"test_run_times": [
2.1699999999999999289457264239899814128875732421875,
2.160000000000000142108547152020037174224853515625,
2.180000000000000159872115546022541821002960205078125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 0.8000000000000000444089209850062616169452667236328125,
"raw_values": [
0.721999999999999975131004248396493494510650634765625,
0.8810000000000000053290705182007513940334320068359375
],
"test_run_times": [
2.70999999999999996447286321199499070644378662109375,
2.70000000000000017763568394002504646778106689453125,
0.1000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625,
0.1000000000000000055511151231257827021181583404541015625,
0.11000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"a35ff7205fdba27121cd1254916ec93054219845": {
"identifier": "system\/darktable-1.0.4",
"title": "Darktable",
"app_version": "2.4.1",
"arguments": "masskrug.NEF output.jpg --core -d perf --disable-opencl",
"description": "Test: Masskrug - Acceleration: CPU-only",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 0.457000000000000017319479184152442030608654022216796875,
"raw_values": [
0.455000000000000015543122344752191565930843353271484375,
0.457000000000000017319479184152442030608654022216796875,
0.458000000000000018207657603852567262947559356689453125
],
"test_run_times": [
1.7800000000000000266453525910037569701671600341796875,
1.770000000000000017763568394002504646778106689453125,
1.7600000000000000088817841970012523233890533447265625
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 0.729999999999999982236431605997495353221893310546875,
"raw_values": [
0.7399999999999999911182158029987476766109466552734375,
0.71899999999999997246646898929611779749393463134765625
],
"test_run_times": [
2.479999999999999982236431605997495353221893310546875,
2.430000000000000159872115546022541821002960205078125,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"d2033afa5c5efd8e1ffff64110967f6865101929": {
"identifier": "system\/darktable-1.0.4",
"title": "Darktable",
"app_version": "2.4.1",
"arguments": "server-rack.dng output.jpg --core -d opencl -d perf",
"description": "Test: Server Rack - Acceleration: OpenCL",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 0.25800000000000000710542735760100185871124267578125,
"raw_values": [
0.25700000000000000621724893790087662637233734130859375,
0.25900000000000000799360577730112709105014801025390625,
0.25700000000000000621724893790087662637233734130859375
],
"test_run_times": [
1.5900000000000000799360577730112709105014801025390625,
1.6100000000000000976996261670137755572795867919921875,
1.5900000000000000799360577730112709105014801025390625
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 0.409999999999999975575093458246556110680103302001953125,
"raw_values": [
0.37600000000000000088817841970012523233890533447265625,
0.441000000000000003108624468950438313186168670654296875
],
"test_run_times": [
1.95999999999999996447286321199499070644378662109375,
1.75,
0.0899999999999999966693309261245303787291049957275390625,
0.1000000000000000055511151231257827021181583404541015625,
0.1000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625,
0.1000000000000000055511151231257827021181583404541015625
]
}
}
},
"9a754e25f4c14f5ded5f98b829dc76048b70b317": {
"identifier": "system\/darktable-1.0.4",
"title": "Darktable",
"app_version": "2.4.1",
"arguments": "server_room.NEF output.jpg --core -d opencl -d perf",
"description": "Test: Server Room - Acceleration: OpenCL",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 0.447000000000000008437694987151189707219600677490234375,
"raw_values": [
0.445000000000000006661338147750939242541790008544921875,
0.446000000000000007549516567451064474880695343017578125,
0.451000000000000011990408665951690636575222015380859375
],
"test_run_times": [
2.089999999999999857891452847979962825775146484375,
2.0800000000000000710542735760100185871124267578125,
2.0800000000000000710542735760100185871124267578125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 0.84999999999999997779553950749686919152736663818359375,
"raw_values": [
0.737999999999999989341858963598497211933135986328125,
0.9619999999999999662492200513952411711215972900390625
],
"test_run_times": [
2.62000000000000010658141036401502788066864013671875,
2.660000000000000142108547152020037174224853515625,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"871aa52ad3a85df8dacb10edfe11aba496cf1349": {
"identifier": "system\/darktable-1.0.4",
"title": "Darktable",
"app_version": "2.4.1",
"arguments": "server-rack.dng output.jpg --core -d perf --disable-opencl",
"description": "Test: Server Rack - Acceleration: CPU-only",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 0.26100000000000000976996261670137755572795867919921875,
"raw_values": [
0.2600000000000000088817841970012523233890533447265625,
0.2600000000000000088817841970012523233890533447265625,
0.262000000000000010658141036401502788066864013671875
],
"test_run_times": [
1.20999999999999996447286321199499070644378662109375,
1.1799999999999999378275106209912337362766265869140625,
1.1999999999999999555910790149937383830547332763671875
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 0.36999999999999999555910790149937383830547332763671875,
"raw_values": [
0.324000000000000010214051826551440171897411346435546875,
0.407999999999999973798736618846305646002292633056640625
],
"test_run_times": [
1.5900000000000000799360577730112709105014801025390625,
1.649999999999999911182158029987476766109466552734375,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.1000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"bbb6e18fd25a022da0242b6248fedfed5adbf9ed": {
"identifier": "system\/darktable-1.0.4",
"title": "Darktable",
"app_version": "2.4.1",
"arguments": "server_room.NEF output.jpg --core -d perf --disable-opencl",
"description": "Test: Server Room - Acceleration: CPU-only",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 0.464000000000000023536728122053318656980991363525390625,
"raw_values": [
0.464000000000000023536728122053318656980991363525390625,
0.463000000000000022648549702353193424642086029052734375,
0.465000000000000024424906541753443889319896697998046875
],
"test_run_times": [
1.6999999999999999555910790149937383830547332763671875,
1.6999999999999999555910790149937383830547332763671875,
1.6799999999999999378275106209912337362766265869140625
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 0.82999999999999996003197111349436454474925994873046875,
"raw_values": [
0.7720000000000000195399252334027551114559173583984375,
0.8930000000000000159872115546022541821002960205078125
],
"test_run_times": [
2.390000000000000124344978758017532527446746826171875,
2.479999999999999982236431605997495353221893310546875,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.1000000000000000055511151231257827021181583404541015625,
0.1000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"8a6ed830709f850982b3fef035ed22d9d2ff9c8f": {
"identifier": "pts\/dav1d-1.6.0",
"title": "dav1d",
"app_version": "0.7.0",
"arguments": "-i chimera_8b_1080p.ivf",
"description": "Video Input: Chimera 1080p",
"scale": "FPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 622.3899999999999863575794734060764312744140625,
"raw_values": [
622.1475000000000363797880709171295166015625,
622.904999999999972715158946812152862548828125,
622.1174999999999499777914024889469146728515625
],
"min_result": [
"621.97"
],
"max_result": [
"623.06"
],
"test_run_times": [
14.75,
14.5099999999999997868371792719699442386627197265625,
14.519999999999999573674358543939888477325439453125
]
}
}
},
"440cd1be88623872ff0e6ee9547b1e6a6becc0bb": {
"identifier": "pts\/dav1d-1.6.0",
"title": "dav1d",
"app_version": "0.7.0",
"arguments": "-i summer_nature_4k.ivf",
"description": "Video Input: Summer Nature 4K",
"scale": "FPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 190.912920000000013942553778178989887237548828125,
"raw_values": [
190.923750000000012505552149377763271331787109375,
190.789999999999992041921359486877918243408203125,
191.025000000000005684341886080801486968994140625
],
"min_result": [
"190.51"
],
"max_result": [
"191.36"
],
"test_run_times": [
19.269999999999999573674358543939888477325439453125,
19.089999999999999857891452847979962825775146484375,
19.07000000000000028421709430404007434844970703125
]
}
}
},
"e31cbb65a2fe735d76d52231e2a63c5a6206b338": {
"identifier": "pts\/dav1d-1.6.0",
"title": "dav1d",
"app_version": "0.7.0",
"arguments": "-i summer_nature_1080p.ivf",
"description": "Video Input: Summer Nature 1080p",
"scale": "FPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 693.264374999999972715158946812152862548828125,
"raw_values": [
691.8700000000000045474735088646411895751953125,
692.05999999999994543031789362430572509765625,
695.8631249999999681676854379475116729736328125
],
"min_result": [
"691.28"
],
"max_result": [
"696.83"
],
"test_run_times": [
5.45000000000000017763568394002504646778106689453125,
5.3300000000000000710542735760100185871124267578125,
5.32000000000000028421709430404007434844970703125
]
}
}
},
"04b0d506572ee7ad24b2ddcb33b3b632ddf2f47d": {
"identifier": "pts\/dav1d-1.6.0",
"title": "dav1d",
"app_version": "0.7.0",
"arguments": "-i chimera_10b_1080p.ivf",
"description": "Video Input: Chimera 1080p 10-bit",
"scale": "FPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 102.1889579999999995152393239550292491912841796875,
"raw_values": [
100.4899999999999948840923025272786617279052734375,
103.4893749999999954525264911353588104248046875,
102.587500000000005684341886080801486968994140625
],
"min_result": [
"100.46"
],
"max_result": [
"103.51"
],
"test_run_times": [
89.2999999999999971578290569595992565155029296875,
86.4800000000000039790393202565610408782958984375,
87.2300000000000039790393202565610408782958984375
]
}
}
},
"cc5f7480baf5f1901713e506466f0fbe4342712d": {
"identifier": "pts\/ebizzy-1.0.4",
"title": "ebizzy",
"app_version": "0.3",
"scale": "Records\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 5682,
"raw_values": [
5695,
5674,
5676
],
"test_run_times": [
25.440000000000001278976924368180334568023681640625,
24.17999999999999971578290569595992565155029296875,
25.219999999999998863131622783839702606201171875
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 2307,
"raw_values": [
2390,
2224
],
"test_run_times": [
20.8299999999999982946974341757595539093017578125,
20.660000000000000142108547152020037174224853515625,
0.1000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"fbcbf269b8c6e8f3ff08d0d5bc80c365ae4ad893": {
"identifier": "pts\/encode-flac-1.6.0",
"title": "FLAC Audio Encoding",
"app_version": "1.3.2",
"description": "WAV To FLAC",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 8.0570000000000003836930773104541003704071044921875,
"raw_values": [
8.1440000000000001278976924368180334568023681640625,
8.0359999999999995878852132591418921947479248046875,
8.0299999999999993605115378159098327159881591796875,
8.044000000000000483169060316868126392364501953125,
8.0310000000000005826450433232821524143218994140625
],
"test_run_times": [
8.1400000000000005684341886080801486968994140625,
8.03999999999999914734871708787977695465087890625,
8.0299999999999993605115378159098327159881591796875,
8.03999999999999914734871708787977695465087890625,
8.0299999999999993605115378159098327159881591796875
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 14.1699999999999999289457264239899814128875732421875,
"raw_values": [
9.751441001892100501891036401502788066864013671875,
18.584253072738999179591701249592006206512451171875
],
"test_run_times": [
9.75,
18.5799999999999982946974341757595539093017578125,
0.190000000000000002220446049250313080847263336181640625,
0.289999999999999980015985556747182272374629974365234375,
0.2300000000000000099920072216264088638126850128173828125,
0.2800000000000000266453525910037569701671600341796875,
0.2399999999999999911182158029987476766109466552734375,
0.1700000000000000122124532708767219446599483489990234375
]
}
}
},
"0a5fcb91439ff7d91351f9c8c7cb0161e684c6a1": {
"identifier": "pts\/go-benchmark-1.1.4",
"title": "Go Benchmarks",
"arguments": "http",
"description": "Test: http",
"scale": "Nanoseconds\/Operation",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 7360,
"raw_values": [
7300,
7459,
7321
],
"test_run_times": [
8.78999999999999914734871708787977695465087890625,
8.57000000000000028421709430404007434844970703125,
8.46000000000000085265128291212022304534912109375
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 10015,
"raw_values": [
10736,
9294
],
"test_run_times": [
12.410000000000000142108547152020037174224853515625,
10.4700000000000006394884621840901672840118408203125,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"7041b356c8b5f4cd92ecf820b29e093a9d0af53a": {
"identifier": "pts\/go-benchmark-1.1.4",
"title": "Go Benchmarks",
"arguments": "json",
"description": "Test: json",
"scale": "Nanoseconds\/Operation",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 1764387,
"raw_values": [
1775129,
1762919,
1755113
],
"test_run_times": [
10.4000000000000003552713678800500929355621337890625,
10.2799999999999993605115378159098327159881591796875,
10.1699999999999999289457264239899814128875732421875
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 7576512,
"raw_values": [
7582912,
7570111
],
"test_run_times": [
9.480000000000000426325641456060111522674560546875,
9.339999999999999857891452847979962825775146484375,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"bf4fff796527fd58f3aee1decf5986622a81a8db": {
"identifier": "pts\/go-benchmark-1.1.4",
"title": "Go Benchmarks",
"arguments": "build",
"description": "Test: build",
"scale": "Nanoseconds\/Operation",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 29396637323,
"raw_values": [
37940556500,
28315805500,
28212221400,
28397371000,
28562503900,
28770671600,
28629558400,
28445910400,
28611043100,
29335126500,
29129961200,
28911784800,
28893770900
],
"test_run_times": [
38.030000000000001136868377216160297393798828125,
28.39999999999999857891452847979962825775146484375,
28.280000000000001136868377216160297393798828125,
28.489999999999998436805981327779591083526611328125,
28.6400000000000005684341886080801486968994140625,
28.85000000000000142108547152020037174224853515625,
28.699999999999999289457264239899814128875732421875,
28.530000000000001136868377216160297393798828125,
28.690000000000001278976924368180334568023681640625,
29.4200000000000017053025658242404460906982421875,
29.21000000000000085265128291212022304534912109375,
28.989999999999998436805981327779591083526611328125,
28.96000000000000085265128291212022304534912109375
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 26780406350,
"raw_values": [
27675836400,
25884976300
],
"test_run_times": [
27.89999999999999857891452847979962825775146484375,
26.030000000000001136868377216160297393798828125,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"631d0ddd2f71400980ee09a97a6c7dfb5442e5a3": {
"identifier": "pts\/go-benchmark-1.1.4",
"title": "Go Benchmarks",
"arguments": "garbage",
"description": "Test: garbage",
"scale": "Nanoseconds\/Operation",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 702173,
"raw_values": [
705117,
700464,
700939
],
"test_run_times": [
20.300000000000000710542735760100185871124267578125,
20.0799999999999982946974341757595539093017578125,
20.1099999999999994315658113919198513031005859375
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 2095389,
"raw_values": [
2096993,
2093784
],
"test_run_times": [
12.57000000000000028421709430404007434844970703125,
12.269999999999999573674358543939888477325439453125,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"e9218196b66940de67c64189b01914d2cc2fb0e5": {
"identifier": "pts\/graphics-magick-2.0.1",
"title": "GraphicsMagick",
"app_version": "1.3.33",
"arguments": "-swirl 90",
"description": "Operation: Swirl",
"scale": "Iterations Per Minute",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 545,
"raw_values": [
541,
547,
547
],
"test_run_times": [
60.13000000000000255795384873636066913604736328125,
60.0799999999999982946974341757595539093017578125,
60.07000000000000028421709430404007434844970703125
]
}
}
},
"48c19a54c4daa4cf389fbfef01f5a1ec99df189d": {
"identifier": "pts\/graphics-magick-2.0.1",
"title": "GraphicsMagick",
"app_version": "1.3.33",
"arguments": "-rotate 90",
"description": "Operation: Rotate",
"scale": "Iterations Per Minute",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 353,
"raw_values": [
356,
350,
352
],
"test_run_times": [
60.159999999999996589394868351519107818603515625,
60.2000000000000028421709430404007434844970703125,
60.090000000000003410605131648480892181396484375
]
}
}
},
"d5a9a5e26f759aa7fe88ee6cfe06ffafbc2f83e8": {
"identifier": "pts\/graphics-magick-2.0.1",
"title": "GraphicsMagick",
"app_version": "1.3.33",
"arguments": "-sharpen 0x2.0",
"description": "Operation: Sharpen",
"scale": "Iterations Per Minute",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 136,
"raw_values": [
135,
136,
136
],
"test_run_times": [
60.18999999999999772626324556767940521240234375,
60.25,
60.0799999999999982946974341757595539093017578125
]
}
}
},
"2932bde0f723a15670f91da170ab533913725f85": {
"identifier": "pts\/graphics-magick-2.0.1",
"title": "GraphicsMagick",
"app_version": "1.3.33",
"arguments": "-enhance",
"description": "Operation: Enhanced",
"scale": "Iterations Per Minute",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 381,
"raw_values": [
384,
377,
381
],
"test_run_times": [
60.21000000000000085265128291212022304534912109375,
60.21000000000000085265128291212022304534912109375,
60.090000000000003410605131648480892181396484375
]
}
}
},
"216ef4bdc8cdeeef23a07753dd79d116198ba190": {
"identifier": "pts\/graphics-magick-2.0.1",
"title": "GraphicsMagick",
"app_version": "1.3.33",
"arguments": "-resize 50%",
"description": "Operation: Resizing",
"scale": "Iterations Per Minute",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 561,
"raw_values": [
560,
562,
562
],
"test_run_times": [
60.090000000000003410605131648480892181396484375,
60.1700000000000017053025658242404460906982421875,
60.090000000000003410605131648480892181396484375
]
}
}
},
"b29f4a46364bc8c91b0300c9af2da1a2d5a9993c": {
"identifier": "pts\/graphics-magick-2.0.1",
"title": "GraphicsMagick",
"app_version": "1.3.33",
"arguments": "-operator all Noise-Gaussian 30%",
"description": "Operation: Noise-Gaussian",
"scale": "Iterations Per Minute",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 262,
"raw_values": [
264,
260,
261
],
"test_run_times": [
60.17999999999999971578290569595992565155029296875,
60.18999999999999772626324556767940521240234375,
60.06000000000000227373675443232059478759765625
]
}
}
},
"accedc6e67d43152e34f7e88c587d92267c7b6a9": {
"identifier": "pts\/graphics-magick-2.0.1",
"title": "GraphicsMagick",
"app_version": "1.3.33",
"arguments": "-colorspace HWB",
"description": "Operation: HWB Color Space",
"scale": "Iterations Per Minute",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 452,
"raw_values": [
448,
454,
454
],
"test_run_times": [
60.1400000000000005684341886080801486968994140625,
60.0799999999999982946974341757595539093017578125,
60.11999999999999744204615126363933086395263671875
]
}
}
},
"a405928c66376da5d339299827bb08a752c1990b": {
"identifier": "pts\/java-scimark2-1.1.2",
"title": "Java SciMark",
"app_version": "2.0",
"arguments": "TEST_COMPOSITE",
"description": "Computational Test: Composite",
"scale": "Mflops",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 2702.30999999999994543031789362430572509765625,
"raw_values": [
2713.7839628837000418570823967456817626953125,
2720.23489172180006789858452975749969482421875,
2654.360415599499901873059570789337158203125,
2720.84462485599988212925381958484649658203125
],
"test_run_times": [
28.629999999999999005240169935859739780426025390625,
28.5799999999999982946974341757595539093017578125,
29.92999999999999971578290569595992565155029296875,
28.559999999999998721023075631819665431976318359375
]
}
}
},
"3a869a480f4d3efb62355f669f01b0c9e328069d": {
"identifier": "pts\/java-scimark2-1.1.2",
"title": "Java SciMark",
"app_version": "2.0",
"arguments": "TEST_MONTE",
"description": "Computational Test: Monte Carlo",
"scale": "Mflops",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 1647.5,
"raw_values": [
1637.4256382134999512345530092716217041015625,
1651.910407573999918895424343645572662353515625,
1646.21209251760001279762946069240570068359375,
1654.455843058299933545640669763088226318359375
]
}
}
},
"4fbd0bed179539c2f8218eb648714a753b5d01ad": {
"identifier": "pts\/java-scimark2-1.1.2",
"title": "Java SciMark",
"app_version": "2.0",
"arguments": "TEST_FFT",
"description": "Computational Test: Fast Fourier Transform",
"scale": "Mflops",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 1884.6199999999998908606357872486114501953125,
"raw_values": [
1995.127995956399900023825466632843017578125,
1978.157827643600057854200713336467742919921875,
1584.233167168900081378524191677570343017578125,
1980.966179710899950805469416081905364990234375
]
}
}
},
"5a09c7e054ba1458261d30e1bc2145ab248d593e": {
"identifier": "pts\/java-scimark2-1.1.2",
"title": "Java SciMark",
"app_version": "2.0",
"arguments": "TEST_SPARSE",
"description": "Computational Test: Sparse Matrix Multiply",
"scale": "Mflops",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 2631.32000000000016370904631912708282470703125,
"raw_values": [
2627.35142544790005558752454817295074462890625,
2635.93770640479988287552259862422943115234375,
2624.0641868093998709809966385364532470703125,
2637.92710731380020661163143813610076904296875
]
}
}
},
"1095c544df0943ac83bee5cd91a8130ac5841575": {
"identifier": "pts\/java-scimark2-1.1.2",
"title": "Java SciMark",
"app_version": "2.0",
"arguments": "TEST_DENSE",
"description": "Computational Test: Dense LU Matrix Factorization",
"scale": "Mflops",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 5454.5,
"raw_values": [
5417.989330847600285778753459453582763671875,
5442.285417415099800564348697662353515625,
5526.2672075485997993382625281810760498046875,
5431.46052318910005851648747920989990234375
]
}
}
},
"f726c094606da03a54e516f51b34ac11e428f39a": {
"identifier": "pts\/java-scimark2-1.1.2",
"title": "Java SciMark",
"app_version": "2.0",
"arguments": "TEST_SOR",
"description": "Computational Test: Jacobi Successive Over-Relaxation",
"scale": "Mflops",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 1893.589999999999918145476840436458587646484375,
"raw_values": [
1891.02542395320006107795052230358123779296875,
1892.883099571399952765204943716526031494140625,
1891.02542395320006107795052230358123779296875,
1899.41347100770008182735182344913482666015625
]
}
}
},
"d17be9fede4ecead5b0d4491becef37b02e8bfd8": {
"identifier": "pts\/john-the-ripper-1.7.2",
"title": "John The Ripper",
"app_version": "1.9.0-jumbo-1",
"arguments": "--format=bcrypt",
"description": "Test: Blowfish",
"scale": "Real C\/S",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 34003,
"raw_values": [
33694,
36054,
33571,
33526,
33596,
33575
],
"test_run_times": [
52.85000000000000142108547152020037174224853515625,
49.3299999999999982946974341757595539093017578125,
52.93999999999999772626324556767940521240234375,
53.02000000000000312638803734444081783294677734375,
52.909999999999996589394868351519107818603515625,
52.93999999999999772626324556767940521240234375
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 10097,
"raw_values": [
11017,
9177
],
"test_run_times": [
54.1700000000000017053025658242404460906982421875,
32.4500000000000028421709430404007434844970703125,
0.0899999999999999966693309261245303787291049957275390625,
0.1000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"7cfeee9e921038188a30cf66762a05c8877fa8fc": {
"identifier": "pts\/john-the-ripper-1.7.2",
"title": "John The Ripper",
"app_version": "1.9.0-jumbo-1",
"arguments": "--format=md5crypt",
"description": "Test: MD5",
"scale": "Real C\/S",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 1659667,
"raw_values": [
1662000,
1664000,
1653000
],
"test_run_times": [
92.9800000000000039790393202565610408782958984375,
92.68000000000000682121026329696178436279296875,
92.9800000000000039790393202565610408782958984375
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 383499,
"raw_values": [
400481,
366516
],
"test_run_times": [
96.4599999999999937472239253111183643341064453125,
64.840000000000003410605131648480892181396484375,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.1000000000000000055511151231257827021181583404541015625,
0.1000000000000000055511151231257827021181583404541015625,
0.11000000000000000055511151231257827021181583404541015625
]
}
}
},
"3050891f5d003b57998824da9ce95e0fa23eba58": {
"identifier": "pts\/encode-mp3-1.7.4",
"title": "LAME MP3 Encoding",
"app_version": "3.100",
"description": "WAV To MP3",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 9.1639999999999997015720509807579219341278076171875,
"raw_values": [
9.166000000000000369482222595252096652984619140625,
9.1669999999999998152588887023739516735076904296875,
9.160000000000000142108547152020037174224853515625
],
"test_run_times": [
9.1699999999999999289457264239899814128875732421875,
9.1699999999999999289457264239899814128875732421875,
9.160000000000000142108547152020037174224853515625
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 12.339999999999999857891452847979962825775146484375,
"raw_values": [
13.0431101322170004408462773426435887813568115234375,
11.63102912902800056826890795491635799407958984375
],
"test_run_times": [
13.03999999999999914734871708787977695465087890625,
11.6300000000000007815970093361102044582366943359375,
0.1600000000000000033306690738754696212708950042724609375,
0.2399999999999999911182158029987476766109466552734375,
0.25,
0.1000000000000000055511151231257827021181583404541015625,
0.1700000000000000122124532708767219446599483489990234375
]
}
}
},
"35040347f411be86c49ef7e27c1fae3e43e8daaf": {
"identifier": "pts\/lczero-1.4.0",
"title": "LeelaChessZero",
"app_version": "0.25",
"arguments": "-b blas",
"description": "Backend: BLAS",
"scale": "Nodes Per Second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 671,
"raw_values": [
696,
682,
638,
667,
655,
692,
676,
664
],
"test_run_times": [
357.259999999999990905052982270717620849609375,
355.51999999999998181010596454143524169921875,
354.970000000000027284841053187847137451171875,
355.43999999999999772626324556767940521240234375,
354.93999999999999772626324556767940521240234375,
355.67000000000001591615728102624416351318359375,
354.20999999999997953636921010911464691162109375,
354.8600000000000136424205265939235687255859375
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 829,
"raw_values": [
1060,
597
],
"test_run_times": [
356.94999999999998863131622783839702606201171875,
356.04000000000002046363078989088535308837890625,
0.1499999999999999944488848768742172978818416595458984375,
0.200000000000000011102230246251565404236316680908203125,
0.2200000000000000011102230246251565404236316680908203125,
0.2099999999999999922284388276239042170345783233642578125,
0.2399999999999999911182158029987476766109466552734375
]
}
}
},
"97814e1f0e82fc4b949e5dda62f5c57ef7d9feba": {
"identifier": "pts\/lczero-1.4.0",
"title": "LeelaChessZero",
"app_version": "0.25",
"arguments": "-b eigen",
"description": "Backend: Eigen",
"scale": "Nodes Per Second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 545,
"raw_values": [
562,
541,
533
],
"test_run_times": [
358.06000000000000227373675443232059478759765625,
360.5,
356.6100000000000136424205265939235687255859375
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 439,
"raw_values": [
429,
449
],
"test_run_times": [
362.6200000000000045474735088646411895751953125,
364.57999999999998408384271897375583648681640625,
0.1700000000000000122124532708767219446599483489990234375,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"819d7d01ceccc0b0066bf5ba95f9d1258c7b2015": {
"identifier": "pts\/lczero-1.4.0",
"title": "LeelaChessZero",
"app_version": "0.25",
"arguments": "-b random",
"description": "Backend: Random",
"scale": "Nodes Per Second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 126764,
"raw_values": [
128571,
123181,
128541
],
"test_run_times": [
342.56000000000000227373675443232059478759765625,
342.56000000000000227373675443232059478759765625,
342.529999999999972715158946812152862548828125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 124746,
"raw_values": [
118141,
131351
],
"test_run_times": [
343.54000000000002046363078989088535308837890625,
343.05000000000001136868377216160297393798828125,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"af9f382bed62c56b2fa199afad3d687f24b38345": {
"identifier": "pts\/tjbench-1.1.1",
"title": "libjpeg-turbo tjbench",
"app_version": "2.0.2",
"arguments": "decompression-throughput",
"description": "Test: Decompression Throughput",
"scale": "Megapixels\/sec",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 203.3413209999999935462255962193012237548828125,
"raw_values": [
203.313456000000002177330316044390201568603515625,
203.3878339999999980136635713279247283935546875,
203.322672000000011394149623811244964599609375
],
"test_run_times": [
6.29999999999999982236431605997495353221893310546875,
6.2400000000000002131628207280300557613372802734375,
6.25
]
}
}
},
"a97fff0392e0dc8b06e7492f95d75e65766f23e6": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-exz,0",
"description": "Test: XZ 0 - Process: Compression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 38,
"test_run_times": [
41.85000000000000142108547152020037174224853515625,
41.57000000000000028421709430404007434844970703125,
41.50999999999999801048033987171947956085205078125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 31,
"raw_values": [
32,
30
],
"test_run_times": [
70.0199999999999960209606797434389591217041015625,
75.0199999999999960209606797434389591217041015625,
0.13000000000000000444089209850062616169452667236328125,
0.11000000000000000055511151231257827021181583404541015625,
0.11000000000000000055511151231257827021181583404541015625,
0.13000000000000000444089209850062616169452667236328125,
0.13000000000000000444089209850062616169452667236328125
]
}
}
},
"1ab5b4ed40c087b99cdeeacac2864e3fc9e296c3": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-exz,0",
"description": "Test: XZ 0 - Process: Decompression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 121,
"test_run_times": [
41.85000000000000142108547152020037174224853515625,
41.57000000000000028421709430404007434844970703125,
41.50999999999999801048033987171947956085205078125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 106,
"raw_values": [
96,
115
],
"test_run_times": [
70.0199999999999960209606797434389591217041015625,
75.0199999999999960209606797434389591217041015625,
0.13000000000000000444089209850062616169452667236328125,
0.11000000000000000055511151231257827021181583404541015625,
0.11000000000000000055511151231257827021181583404541015625,
0.13000000000000000444089209850062616169452667236328125,
0.13000000000000000444089209850062616169452667236328125
]
}
}
},
"32c63cd3d6f9367ae720a9b9afc31d7ebf02dc32": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-ezstd,1",
"description": "Test: Zstd 1 - Process: Compression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 495,
"raw_values": [
505,
452,
456,
483,
475,
491,
509,
506,
510,
507,
508,
500,
509,
508,
507
],
"test_run_times": [
23.35000000000000142108547152020037174224853515625,
22.1400000000000005684341886080801486968994140625,
21.969999999999998863131622783839702606201171875,
23.3299999999999982946974341757595539093017578125,
23.559999999999998721023075631819665431976318359375,
23.1700000000000017053025658242404460906982421875,
22.949999999999999289457264239899814128875732421875,
22.879999999999999005240169935859739780426025390625,
22.969999999999998863131622783839702606201171875,
22.78999999999999914734871708787977695465087890625,
23.050000000000000710542735760100185871124267578125,
24.17999999999999971578290569595992565155029296875,
22.739999999999998436805981327779591083526611328125,
23.8900000000000005684341886080801486968994140625,
24.019999999999999573674358543939888477325439453125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 484,
"test_run_times": [
26.0799999999999982946974341757595539093017578125,
23.760000000000001563194018672220408916473388671875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"6f249b3604d44d4d3f31896bb0653b59dd533a28": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-ezstd,1",
"description": "Test: Zstd 1 - Process: Decompression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 829,
"raw_values": [
824,
833,
835,
821,
831,
832,
809,
842,
812,
832,
828,
839,
825,
840,
839
],
"test_run_times": [
23.35000000000000142108547152020037174224853515625,
22.1400000000000005684341886080801486968994140625,
21.969999999999998863131622783839702606201171875,
23.3299999999999982946974341757595539093017578125,
23.559999999999998721023075631819665431976318359375,
23.1700000000000017053025658242404460906982421875,
22.949999999999999289457264239899814128875732421875,
22.879999999999999005240169935859739780426025390625,
22.969999999999998863131622783839702606201171875,
22.78999999999999914734871708787977695465087890625,
23.050000000000000710542735760100185871124267578125,
24.17999999999999971578290569595992565155029296875,
22.739999999999998436805981327779591083526611328125,
23.8900000000000005684341886080801486968994140625,
24.019999999999999573674358543939888477325439453125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 737,
"raw_values": [
675,
798
],
"test_run_times": [
26.0799999999999982946974341757595539093017578125,
23.760000000000001563194018672220408916473388671875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"25c9b4e1e91e600c606af13ee9cfca164a8525fc": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-ezstd,8",
"description": "Test: Zstd 8 - Process: Compression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 99,
"test_run_times": [
30.879999999999999005240169935859739780426025390625,
31.8599999999999994315658113919198513031005859375,
31.469999999999998863131622783839702606201171875
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 66,
"raw_values": [
73,
59
],
"test_run_times": [
25.6400000000000005684341886080801486968994140625,
28.9200000000000017053025658242404460906982421875,
0.08000000000000000166533453693773481063544750213623046875,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"edb61ba9199c6f5934fef86679c889a5f7bfdd5f": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-ezstd,8",
"description": "Test: Zstd 8 - Process: Decompression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 846,
"raw_values": [
840,
843,
855
],
"test_run_times": [
30.879999999999999005240169935859739780426025390625,
31.8599999999999994315658113919198513031005859375,
31.469999999999998863131622783839702606201171875
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 795,
"raw_values": [
825,
764
],
"test_run_times": [
25.6400000000000005684341886080801486968994140625,
28.9200000000000017053025658242404460906982421875,
0.08000000000000000166533453693773481063544750213623046875,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"87df8d5b801ad3bd53e12a1906a9873b94369183": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-ecrush,0",
"description": "Test: Crush 0 - Process: Compression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 107,
"raw_values": [
115,
101,
116,
114,
119,
93,
106,
117,
116,
88,
102,
100,
119,
100,
93
],
"test_run_times": [
29.280000000000001136868377216160297393798828125,
30.78999999999999914734871708787977695465087890625,
28.3900000000000005684341886080801486968994140625,
28.769999999999999573674358543939888477325439453125,
28.739999999999998436805981327779591083526611328125,
22.309999999999998721023075631819665431976318359375,
30.21000000000000085265128291212022304534912109375,
28.260000000000001563194018672220408916473388671875,
28.730000000000000426325641456060111522674560546875,
22.809999999999998721023075631819665431976318359375,
30.6400000000000005684341886080801486968994140625,
30.980000000000000426325641456060111522674560546875,
28.53999999999999914734871708787977695465087890625,
31,
22.239999999999998436805981327779591083526611328125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 62,
"raw_values": [
57,
66
],
"test_run_times": [
31.280000000000001136868377216160297393798828125,
26.78999999999999914734871708787977695465087890625,
0.11999999999999999555910790149937383830547332763671875,
0.13000000000000000444089209850062616169452667236328125,
0.13000000000000000444089209850062616169452667236328125,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"128f8170b40bf9f91c0f3e4e77094829ac809e06": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-ecrush,0",
"description": "Test: Crush 0 - Process: Decompression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 467,
"raw_values": [
468,
468,
467,
467,
468,
466,
466,
468,
468,
467,
468,
467,
467,
467,
468
],
"test_run_times": [
29.280000000000001136868377216160297393798828125,
30.78999999999999914734871708787977695465087890625,
28.3900000000000005684341886080801486968994140625,
28.769999999999999573674358543939888477325439453125,
28.739999999999998436805981327779591083526611328125,
22.309999999999998721023075631819665431976318359375,
30.21000000000000085265128291212022304534912109375,
28.260000000000001563194018672220408916473388671875,
28.730000000000000426325641456060111522674560546875,
22.809999999999998721023075631819665431976318359375,
30.6400000000000005684341886080801486968994140625,
30.980000000000000426325641456060111522674560546875,
28.53999999999999914734871708787977695465087890625,
31,
22.239999999999998436805981327779591083526611328125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 432,
"raw_values": [
450,
413
],
"test_run_times": [
31.280000000000001136868377216160297393798828125,
26.78999999999999914734871708787977695465087890625,
0.11999999999999999555910790149937383830547332763671875,
0.13000000000000000444089209850062616169452667236328125,
0.13000000000000000444089209850062616169452667236328125,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"f38647cf6335ade425c0eac6f6df2d7613680603": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-ebrotli,0",
"description": "Test: Brotli 0 - Process: Compression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 399,
"raw_values": [
268,
410,
414,
409,
413,
408,
411,
414,
412,
411,
412,
408
],
"test_run_times": [
22.519999999999999573674358543939888477325439453125,
23.82000000000000028421709430404007434844970703125,
23.550000000000000710542735760100185871124267578125,
23.82000000000000028421709430404007434844970703125,
23.440000000000001278976924368180334568023681640625,
23.8299999999999982946974341757595539093017578125,
23.85000000000000142108547152020037174224853515625,
23.35000000000000142108547152020037174224853515625,
23.53999999999999914734871708787977695465087890625,
23.5,
23.8599999999999994315658113919198513031005859375,
23.780000000000001136868377216160297393798828125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 359,
"raw_values": [
338,
380
],
"test_run_times": [
24.6700000000000017053025658242404460906982421875,
23.989999999999998436805981327779591083526611328125,
0.11999999999999999555910790149937383830547332763671875,
0.11999999999999999555910790149937383830547332763671875,
0.11999999999999999555910790149937383830547332763671875,
0.11000000000000000055511151231257827021181583404541015625,
0.11000000000000000055511151231257827021181583404541015625
]
}
}
},
"3f13c4be7c50b339286084a06826f68c6931739b": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-ebrotli,0",
"description": "Test: Brotli 0 - Process: Decompression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 580,
"raw_values": [
579,
581,
581,
579,
579,
578,
581,
581,
580,
579,
579,
580
],
"test_run_times": [
22.519999999999999573674358543939888477325439453125,
23.82000000000000028421709430404007434844970703125,
23.550000000000000710542735760100185871124267578125,
23.82000000000000028421709430404007434844970703125,
23.440000000000001278976924368180334568023681640625,
23.8299999999999982946974341757595539093017578125,
23.85000000000000142108547152020037174224853515625,
23.35000000000000142108547152020037174224853515625,
23.53999999999999914734871708787977695465087890625,
23.5,
23.8599999999999994315658113919198513031005859375,
23.780000000000001136868377216160297393798828125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 551,
"raw_values": [
550,
551
],
"test_run_times": [
24.6700000000000017053025658242404460906982421875,
23.989999999999998436805981327779591083526611328125,
0.11999999999999999555910790149937383830547332763671875,
0.11999999999999999555910790149937383830547332763671875,
0.11999999999999999555910790149937383830547332763671875,
0.11000000000000000055511151231257827021181583404541015625,
0.11000000000000000055511151231257827021181583404541015625
]
}
}
},
"f7bde4847668cfa74d587c36950b7ec182656eba": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-ebrotli,2",
"description": "Test: Brotli 2 - Process: Compression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 186,
"raw_values": [
192,
181,
173,
169,
191,
187,
187,
191,
190,
191,
183,
190,
188,
191,
191
],
"test_run_times": [
26.89999999999999857891452847979962825775146484375,
23.230000000000000426325641456060111522674560546875,
23.32000000000000028421709430404007434844970703125,
23.6099999999999994315658113919198513031005859375,
27.480000000000000426325641456060111522674560546875,
22.550000000000000710542735760100185871124267578125,
22.3299999999999982946974341757595539093017578125,
27.030000000000001136868377216160297393798828125,
27.410000000000000142108547152020037174224853515625,
27.519999999999999573674358543939888477325439453125,
22.780000000000001136868377216160297393798828125,
22.219999999999998863131622783839702606201171875,
22.190000000000001278976924368180334568023681640625,
27.050000000000000710542735760100185871124267578125,
27.32000000000000028421709430404007434844970703125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 176,
"raw_values": [
179,
173
],
"test_run_times": [
24.410000000000000142108547152020037174224853515625,
25.379999999999999005240169935859739780426025390625,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"b3275b0d8c80227f7222f198278a8d6e1564d846": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-ebrotli,2",
"description": "Test: Brotli 2 - Process: Decompression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 670,
"raw_values": [
672,
670,
666,
657,
670,
670,
670,
672,
669,
670,
670,
671,
671,
671,
674
],
"test_run_times": [
26.89999999999999857891452847979962825775146484375,
23.230000000000000426325641456060111522674560546875,
23.32000000000000028421709430404007434844970703125,
23.6099999999999994315658113919198513031005859375,
27.480000000000000426325641456060111522674560546875,
22.550000000000000710542735760100185871124267578125,
22.3299999999999982946974341757595539093017578125,
27.030000000000001136868377216160297393798828125,
27.410000000000000142108547152020037174224853515625,
27.519999999999999573674358543939888477325439453125,
22.780000000000001136868377216160297393798828125,
22.219999999999998863131622783839702606201171875,
22.190000000000001278976924368180334568023681640625,
27.050000000000000710542735760100185871124267578125,
27.32000000000000028421709430404007434844970703125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 607,
"raw_values": [
577,
636
],
"test_run_times": [
24.410000000000000142108547152020037174224853515625,
25.379999999999999005240169935859739780426025390625,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"a481174911c170e0ad2a765ad48b6e015cc7b36e": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-elibdeflate,1",
"description": "Test: Libdeflate 1 - Process: Compression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 253,
"test_run_times": [
23.370000000000000994759830064140260219573974609375,
22.199999999999999289457264239899814128875732421875,
23.300000000000000710542735760100185871124267578125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 234,
"raw_values": [
243,
224
],
"test_run_times": [
24.410000000000000142108547152020037174224853515625,
25.190000000000001278976924368180334568023681640625,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"32c31601b81a0078b59cf31c4b9a5edfdb2d6ac7": {
"identifier": "pts\/lzbench-1.1.0",
"title": "lzbench",
"app_version": "1.8",
"arguments": "-elibdeflate,1",
"description": "Test: Libdeflate 1 - Process: Decompression",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 1227,
"raw_values": [
1226,
1231,
1225
],
"test_run_times": [
23.370000000000000994759830064140260219573974609375,
22.199999999999999289457264239899814128875732421875,
23.300000000000000710542735760100185871124267578125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 1195,
"raw_values": [
1199,
1190
],
"test_run_times": [
24.410000000000000142108547152020037174224853515625,
25.190000000000001278976924368180334568023681640625,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"3d6c0a94e5b200281e3eaf197b04f2628760514d": {
"identifier": "pts\/m-queens-1.1.0",
"title": "m-queens",
"app_version": "1.2",
"description": "Time To Solve",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 25.57600000000000051159076974727213382720947265625,
"raw_values": [
25.428000000000000824229573481716215610504150390625,
25.620999999999998664179656771011650562286376953125,
25.678999999999998493649400188587605953216552734375
],
"test_run_times": [
25.42999999999999971578290569595992565155029296875,
25.620000000000000994759830064140260219573974609375,
25.67999999999999971578290569595992565155029296875
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 119.7099999999999937472239253111183643341064453125,
"raw_values": [
114.3618979454000026407811674289405345916748046875,
125.05224490166000350654940120875835418701171875
],
"test_run_times": [
114.3599999999999994315658113919198513031005859375,
125.0499999999999971578290569595992565155029296875,
0.1000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"2f9e749dc7b6b26e2e18e63972e940a1aa5ad80c": {
"identifier": "pts\/minion-1.5.2",
"title": "Minion",
"app_version": "1.8",
"arguments": "benchmarks\/graceful\/k6p2_table.minion",
"description": "Benchmark: Graceful",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 49.10900000000000176214598468504846096038818359375,
"raw_values": [
49.0870000000000032969182939268648624420166015625,
48.98599999999999710098563809879124164581298828125,
49.2530000000000001136868377216160297393798828125
],
"test_run_times": [
49.17999999999999971578290569595992565155029296875,
49.06000000000000227373675443232059478759765625,
49.31000000000000227373675443232059478759765625
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 55.2999999999999971578290569595992565155029296875,
"raw_values": [
55.5799999999999982946974341757595539093017578125,
55.01599999999999823785401531495153903961181640625
],
"test_run_times": [
58.96000000000000085265128291212022304534912109375,
55.13000000000000255795384873636066913604736328125,
0.1000000000000000055511151231257827021181583404541015625,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"2b540121ed8dc4275bde294a726e7059480f5985": {
"identifier": "pts\/minion-1.5.2",
"title": "Minion",
"app_version": "1.8",
"arguments": "benchmarks\/solitaire\/solitaire_benchmark_8.minion",
"description": "Benchmark: Solitaire",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 48.143000000000000682121026329696178436279296875,
"raw_values": [
48.3250000000000028421709430404007434844970703125,
48.06000000000000227373675443232059478759765625,
48.0439999999999969304553815163671970367431640625
],
"test_run_times": [
48.409999999999996589394868351519107818603515625,
48.14999999999999857891452847979962825775146484375,
48.11999999999999744204615126363933086395263671875
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 55.530000000000001136868377216160297393798828125,
"raw_values": [
54.11200000000000187583282240666449069976806640625,
56.93999999999999772626324556767940521240234375
],
"test_run_times": [
54.25,
57.07000000000000028421709430404007434844970703125,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"83c4f37052ad8eff43bd22c660b254c2d166f665": {
"identifier": "pts\/minion-1.5.2",
"title": "Minion",
"app_version": "1.8",
"arguments": "benchmarks\/Quasigroup\/qg-watchelement-7-10.minion",
"description": "Benchmark: Quasigroup",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 112,
"raw_values": [
112.3610000000000042064129956997931003570556640625,
112.1979999999999932924765744246542453765869140625,
112.1929999999999978399500832892954349517822265625
],
"test_run_times": [
112.43000000000000682121026329696178436279296875,
112.2600000000000051159076974727213382720947265625,
112.25
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 120,
"raw_values": [
119.98799999999999954525264911353588104248046875,
120.0720000000000027284841053187847137451171875
],
"test_run_times": [
120.18999999999999772626324556767940521240234375,
120.219999999999998863131622783839702606201171875,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"2b19a6308da71d23f6fdf23cb294745a8899d8b1": {
"identifier": "pts\/nginx-1.2.2",
"title": "NGINX Benchmark",
"app_version": "1.9.9",
"description": "Static Web Page Serving",
"scale": "Requests Per Second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 731,
"raw_values": [
721.5399999999999636202119290828704833984375,
734.6200000000000045474735088646411895751953125,
737.1100000000000136424205265939235687255859375
],
"test_run_times": [
2787.55999999999994543031789362430572509765625,
2738.13999999999987267074175179004669189453125,
2728.94000000000005456968210637569427490234375
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 673,
"raw_values": [
680.25,
665.1100000000000136424205265939235687255859375
],
"test_run_times": [
2957.3000000000001818989403545856475830078125,
3023.40000000000009094947017729282379150390625,
0.1000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"a0a6ca760a440139eff96952ffeb95422093a58a": {
"identifier": "pts\/novabench-1.0.0",
"title": "Novabench",
"arguments": "--test=cpu",
"description": "Test: CPU",
"scale": "CPU Score",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 633,
"raw_values": [
523,
666,
597,
688,
647,
661,
670,
573,
623,
606,
642,
661,
645,
650,
650
],
"test_run_times": [
18.199999999999999289457264239899814128875732421875,
18.0799999999999982946974341757595539093017578125,
18.089999999999999857891452847979962825775146484375,
18.089999999999999857891452847979962825775146484375,
18.089999999999999857891452847979962825775146484375,
18.089999999999999857891452847979962825775146484375,
18.10000000000000142108547152020037174224853515625,
18.0799999999999982946974341757595539093017578125,
18.1099999999999994315658113919198513031005859375,
18.10000000000000142108547152020037174224853515625,
18.0799999999999982946974341757595539093017578125,
18.10000000000000142108547152020037174224853515625,
18.089999999999999857891452847979962825775146484375,
18.089999999999999857891452847979962825775146484375,
18.089999999999999857891452847979962825775146484375
]
}
}
},
"8a3c4a1de6da617cf2a53728f1d6114b2f6d988a": {
"identifier": "pts\/novabench-1.0.0",
"title": "Novabench",
"arguments": "--test=gpu",
"description": "Test: GPU",
"scale": "GFLOPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 15270,
"test_run_times": [
1.479999999999999982236431605997495353221893310546875,
1.4099999999999999200639422269887290894985198974609375,
1.3899999999999999023003738329862244427204132080078125
]
}
}
},
"1078cb1dacfa8617071f511bb1388a538b7907d7": {
"identifier": "pts\/novabench-1.0.0",
"title": "Novabench",
"arguments": "--test=ram",
"description": "Test: RAM",
"scale": "RAM Score",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 392,
"raw_values": [
386,
397,
394
],
"test_run_times": [
1.25,
0.7800000000000000266453525910037569701671600341796875,
0.7800000000000000266453525910037569701671600341796875
]
}
}
},
"cc9f885dc8eb1c602bac8fd2bb116cf809a74e8d": {
"identifier": "pts\/novabench-1.0.0",
"title": "Novabench",
"arguments": "--test=ram",
"description": "Test: RAM",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 37631,
"raw_values": [
35841,
38986,
38065
],
"test_run_times": [
1.25,
0.7800000000000000266453525910037569701671600341796875,
0.7800000000000000266453525910037569701671600341796875
]
}
}
},
"946f5e0360cf178c425ab9080de3597a148b632b": {
"identifier": "pts\/phpbench-1.1.5",
"title": "PHPBench",
"app_version": "0.8.1",
"description": "PHP Benchmark Suite",
"scale": "Score",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 382488,
"raw_values": [
383614,
381915,
381934
],
"test_run_times": [
26.199999999999999289457264239899814128875732421875,
26.260000000000001563194018672220408916473388671875,
26.280000000000001136868377216160297393798828125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 234849,
"raw_values": [
234253,
235445
],
"test_run_times": [
43.219999999999998863131622783839702606201171875,
42.68999999999999772626324556767940521240234375,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"ee40faab64e98100a54ff22c56593430d4219d8f": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "ON_DISK NORMAL_LOAD READ_ONLY",
"description": "Scaling: On-Disk - Test: Normal Load - Mode: Read Only",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 389451,
"raw_values": [
390212.47147400001995265483856201171875,
391526.1676309999893419444561004638671875,
386614.6295800000079907476902008056640625
],
"test_run_times": [
4079.829999999999927240423858165740966796875,
4000.920000000000072759576141834259033203125,
3976.88999999999987267074175179004669189453125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 71369,
"raw_values": [
65400.2898320000022067688405513763427734375,
77337.240380999995977617800235748291015625
],
"test_run_times": [
4253.3800000000001091393642127513885498046875,
3566.73000000000001818989403545856475830078125,
0.11999999999999999555910790149937383830547332763671875,
0.11999999999999999555910790149937383830547332763671875,
0.0899999999999999966693309261245303787291049957275390625,
0.11000000000000000055511151231257827021181583404541015625,
0.11000000000000000055511151231257827021181583404541015625
]
}
}
},
"4f92e19442e6e76294f4b303bf8059b6b49216b5": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "ON_DISK NORMAL_LOAD READ_WRITE",
"description": "Scaling: On-Disk - Test: Normal Load - Mode: Read Write",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 1308,
"raw_values": [
1315.034769999999980427674017846584320068359375,
1302.586645000000089567038230597972869873046875,
1304.958378999999922598362900316715240478515625
],
"test_run_times": [
3980.1300000000001091393642127513885498046875,
4625.739999999999781721271574497222900390625,
4661.1899999999995998223312199115753173828125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 955,
"raw_values": [
852.125009999999974752427078783512115478515625,
1058.516030000000000654836185276508331298828125
],
"test_run_times": [
3430.57000000000016370904631912708282470703125,
3534.760000000000218278728425502777099609375,
0.299999999999999988897769753748434595763683319091796875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"fa028f5ee807c5bf67a72f64e2358dde1f0d0ea0": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "ON_DISK SINGLE_THREAD READ_ONLY",
"description": "Scaling: On-Disk - Test: Single Thread - Mode: Read Only",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 23456,
"raw_values": [
21497.85195200000089243985712528228759765625,
23847.321423000001232139766216278076171875,
25022.66853099999934784136712551116943359375
],
"test_run_times": [
4576.920000000000072759576141834259033203125,
4818.9499999999998181010596454143524169921875,
4839.670000000000072759576141834259033203125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 4635,
"raw_values": [
4242.2333779999999023857526481151580810546875,
5028.5553209999998216517269611358642578125
],
"test_run_times": [
3136.09999999999990905052982270717620849609375,
3237.329999999999927240423858165740966796875,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"19ea2443be9cb6407a151d0d773471a430064db7": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "MOSTLY_CACHE NORMAL_LOAD READ_ONLY",
"description": "Scaling: Mostly RAM - Test: Normal Load - Mode: Read Only",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 130419,
"raw_values": [
130186.705474000002141110599040985107421875,
130279.8973450000048615038394927978515625,
130791.614038999992772005498409271240234375
],
"test_run_times": [
3333.94000000000005456968210637569427490234375,
3357.5399999999999636202119290828704833984375,
3296.4499999999998181010596454143524169921875
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 17827,
"raw_values": [
17627.31603099999847472645342350006103515625,
18026.01708100000178092159330844879150390625
],
"test_run_times": [
2981.23000000000001818989403545856475830078125,
2613.23000000000001818989403545856475830078125,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"f5424b42195c71e9b11259bd62716f63ff3206b2": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "ON_DISK SINGLE_THREAD READ_WRITE",
"description": "Scaling: On-Disk - Test: Single Thread - Mode: Read Write",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 42.20312899999999700639818911440670490264892578125,
"raw_values": [
42.30132900000000262252797256223857402801513671875,
42.08528199999999941383066470734775066375732421875,
42.22277700000000066893335315398871898651123046875
],
"test_run_times": [
4506.72999999999956344254314899444580078125,
4647.6800000000002910383045673370361328125,
4623.0500000000001818989403545856475830078125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 113.8599999999999994315658113919198513031005859375,
"raw_values": [
126.6763600000000025147528504021465778350830078125,
101.0370549999999951751306070946156978607177734375
],
"test_run_times": [
2793.579999999999927240423858165740966796875,
2881.2899999999999636202119290828704833984375,
0.1000000000000000055511151231257827021181583404541015625,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.1000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"d7314229cd48bfb4f238b2c179d7b33c6776ac0e": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "BUFFER_TEST NORMAL_LOAD READ_ONLY",
"description": "Scaling: Buffer Test - Test: Normal Load - Mode: Read Only",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 144452,
"raw_values": [
144164.27153200001339428126811981201171875,
144647.66216099998564459383487701416015625,
144544.79310199999599717557430267333984375
],
"test_run_times": [
112.2000000000000028421709430404007434844970703125,
112.8299999999999982946974341757595539093017578125,
112.9800000000000039790393202565610408782958984375
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 45406,
"raw_values": [
44702.5805820000023231841623783111572265625,
46108.7984300000025541521608829498291015625
],
"test_run_times": [
103.18000000000000682121026329696178436279296875,
103.4599999999999937472239253111183643341064453125,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"38ffcce4263977c87ee9483e7777e7dc0813194a": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "MOSTLY_CACHE NORMAL_LOAD READ_WRITE",
"description": "Scaling: Mostly RAM - Test: Normal Load - Mode: Read Write",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 12878,
"raw_values": [
13104.69198300000061863102018833160400390625,
13068.005279999999402207322418689727783203125,
12462.700570000000880099833011627197265625
],
"test_run_times": [
3407.67999999999983629095368087291717529296875,
3424.51999999999998181010596454143524169921875,
3336.84999999999990905052982270717620849609375
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 2371,
"raw_values": [
2380.7361190000001442967914044857025146484375,
2361.75140400000009321956895291805267333984375
],
"test_run_times": [
2642.36000000000012732925824820995330810546875,
2663.71999999999979991116560995578765869140625,
0.13000000000000000444089209850062616169452667236328125,
0.1000000000000000055511151231257827021181583404541015625
]
}
}
},
"7c62964246ec3d489ef68d9dd7de35b2a6057f19": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "BUFFER_TEST NORMAL_LOAD READ_WRITE",
"description": "Scaling: Buffer Test - Test: Normal Load - Mode: Read Write",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 37125,
"raw_values": [
36688.6589670000030309893190860748291015625,
37022.0126200000013341195881366729736328125,
37663.3456829999995534308254718780517578125
],
"test_run_times": [
115.840000000000003410605131648480892181396484375,
114.7399999999999948840923025272786617279052734375,
116.099999999999994315658113919198513031005859375
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 4266,
"raw_values": [
4810.808810000000448781065642833709716796875,
3720.83629299999984141322784125804901123046875
],
"test_run_times": [
113.5499999999999971578290569595992565155029296875,
112.090000000000003410605131648480892181396484375,
0.11000000000000000055511151231257827021181583404541015625,
0.1000000000000000055511151231257827021181583404541015625,
0.11000000000000000055511151231257827021181583404541015625,
0.13000000000000000444089209850062616169452667236328125,
0.11999999999999999555910790149937383830547332763671875
]
}
}
},
"05cf268a1bd590ed44b0290394dc409a0705e797": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "MOSTLY_CACHE SINGLE_THREAD READ_ONLY",
"description": "Scaling: Mostly RAM - Test: Single Thread - Mode: Read Only",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 3776,
"raw_values": [
3798.8768730000001596636138856410980224609375,
3716.22207699999989927164278924465179443359375,
3812.78796000000011190422810614109039306640625
],
"test_run_times": [
3371.01999999999998181010596454143524169921875,
3344.94000000000005456968210637569427490234375,
3347.73000000000001818989403545856475830078125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 1228,
"raw_values": [
1166.272752000000082261976785957813262939453125,
1290.49162200000000666477717459201812744140625
],
"test_run_times": [
2964.48000000000001818989403545856475830078125,
3135.25,
0.1000000000000000055511151231257827021181583404541015625,
0.1000000000000000055511151231257827021181583404541015625,
0.1000000000000000055511151231257827021181583404541015625,
0.1000000000000000055511151231257827021181583404541015625,
0.1000000000000000055511151231257827021181583404541015625
]
}
}
},
"11649e745d454fd5d98aa6ad135c43296d8b2a0e": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "ON_DISK HEAVY_CONTENTION READ_ONLY",
"description": "Scaling: On-Disk - Test: Heavy Contention - Mode: Read Only",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 388300,
"raw_values": [
382171.41986000002361834049224853515625,
388620.913816999993287026882171630859375,
394107.4018049999722279608249664306640625
],
"test_run_times": [
4617.4600000000000363797880709171295166015625,
4607.9399999999995998223312199115753173828125,
4522.670000000000072759576141834259033203125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 68793,
"raw_values": [
74298.5830100000021047890186309814453125,
63288.168772999997599981725215911865234375
],
"test_run_times": [
3708.01999999999998181010596454143524169921875,
3509.3000000000001818989403545856475830078125,
0.14000000000000001332267629550187848508358001708984375,
0.1600000000000000033306690738754696212708950042724609375,
0.13000000000000000444089209850062616169452667236328125,
0.11000000000000000055511151231257827021181583404541015625,
0.11999999999999999555910790149937383830547332763671875
]
}
}
},
"5db7d6938ebc18fdd26330fa9b532afcca9b820b": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "BUFFER_TEST SINGLE_THREAD READ_ONLY",
"description": "Scaling: Buffer Test - Test: Single Thread - Mode: Read Only",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 14115,
"raw_values": [
14758.66267499999958090484142303466796875,
14762.03015299999970011413097381591796875,
13744.04439099999945028685033321380615234375,
13168.020868999999947845935821533203125,
14717.908695000000079744495451450347900390625,
14724.74614199999996344558894634246826171875,
13445.537173999999140505678951740264892578125,
13071.207488000000012107193470001220703125,
14791.44098999999914667569100856781005859375,
14818.76752800000031129457056522369384765625,
13658.59610400000019581057131290435791015625,
13090.9794089999995776452124118804931640625,
13619.0522899999996297992765903472900390625,
14559.79774600000018835999071598052978515625,
14801.248642000000472762621939182281494140625
],
"test_run_times": [
113.4200000000000017053025658242404460906982421875,
112.6200000000000045474735088646411895751953125,
111.340000000000003410605131648480892181396484375,
111.650000000000005684341886080801486968994140625,
111.8900000000000005684341886080801486968994140625,
111.9200000000000017053025658242404460906982421875,
112.1400000000000005684341886080801486968994140625,
112.31000000000000227373675443232059478759765625,
112.18999999999999772626324556767940521240234375,
112.2600000000000051159076974727213382720947265625,
111.5,
113.2699999999999960209606797434389591217041015625,
113.0400000000000062527760746888816356658935546875,
111.2600000000000051159076974727213382720947265625,
111.280000000000001136868377216160297393798828125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 3487,
"raw_values": [
3342.679097000000183470547199249267578125,
3630.91158099999984187888912856578826904296875
],
"test_run_times": [
135.590000000000003410605131648480892181396484375,
115.0799999999999982946974341757595539093017578125,
0.11000000000000000055511151231257827021181583404541015625,
0.11000000000000000055511151231257827021181583404541015625,
0.11000000000000000055511151231257827021181583404541015625,
0.1000000000000000055511151231257827021181583404541015625,
0.1000000000000000055511151231257827021181583404541015625
]
}
}
},
"4c35c7ef81c117cd77c00e1a2fca916073e5176b": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "MOSTLY_CACHE SINGLE_THREAD READ_WRITE",
"description": "Scaling: Mostly RAM - Test: Single Thread - Mode: Read Write",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 1339,
"raw_values": [
1297.797956000000112908310256898403167724609375,
1352.552028000000063912011682987213134765625,
1367.66870100000005550100468099117279052734375
],
"test_run_times": [
3335.7899999999999636202119290828704833984375,
3411.38999999999987267074175179004669189453125,
3331.03000000000020008883439004421234130859375
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 287,
"raw_values": [
219.508024000000006026311893947422504425048828125,
354.71288199999997914346749894320964813232421875
],
"test_run_times": [
3814.239999999999781721271574497222900390625,
3935.920000000000072759576141834259033203125,
0.11000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625,
0.1000000000000000055511151231257827021181583404541015625,
0.11999999999999999555910790149937383830547332763671875,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"d6ea74c179ada8147bd542d1e364b05c2763f25c": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "ON_DISK HEAVY_CONTENTION READ_WRITE",
"description": "Scaling: On-Disk - Test: Heavy Contention - Mode: Read Write",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 1273,
"raw_values": [
1265.537861999999904583091847598552703857421875,
1281.586939000000029409420676529407501220703125,
1272.48312699999996766564436256885528564453125
],
"test_run_times": [
4475.899999999999636202119290828704833984375,
4486.02000000000043655745685100555419921875,
4498.399999999999636202119290828704833984375
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 937,
"raw_values": [
928.80929900000000998261384665966033935546875,
945.7713390000000117652234621345996856689453125
],
"test_run_times": [
3350.76999999999998181010596454143524169921875,
3411.1300000000001091393642127513885498046875,
0.11000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"48c6fecaafbf03ab4a1904bf55026f5006decc10": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "BUFFER_TEST SINGLE_THREAD READ_WRITE",
"description": "Scaling: Buffer Test - Test: Single Thread - Mode: Read Write",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 2016,
"raw_values": [
2004.528317000000015468685887753963470458984375,
2012.307238000000097599695436656475067138671875,
2030.393013999999993757228367030620574951171875
],
"test_run_times": [
114.530000000000001136868377216160297393798828125,
113.9500000000000028421709430404007434844970703125,
112.18000000000000682121026329696178436279296875
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 515,
"raw_values": [
549.7179640000000517829903401434421539306640625,
480.77694200000001956141204573214054107666015625
],
"test_run_times": [
109.8599999999999994315658113919198513031005859375,
106.8700000000000045474735088646411895751953125,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.1000000000000000055511151231257827021181583404541015625,
0.1000000000000000055511151231257827021181583404541015625
]
}
}
},
"df2c810c84611ca3583a60283ccf36c34b2dbb5e": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "MOSTLY_CACHE HEAVY_CONTENTION READ_ONLY",
"description": "Scaling: Mostly RAM - Test: Heavy Contention - Mode: Read Only",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 126996,
"raw_values": [
125126.335852000003797002136707305908203125,
129299.16046899999491870403289794921875,
126561.61941600000136531889438629150390625
],
"test_run_times": [
3381.829999999999927240423858165740966796875,
3418.55999999999994543031789362430572509765625,
3372.65999999999985448084771633148193359375
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 14202,
"raw_values": [
13943.59066499999971711076796054840087890625,
14460.514440000000831787474453449249267578125
],
"test_run_times": [
3036.94000000000005456968210637569427490234375,
3006.40000000000009094947017729282379150390625,
0.11000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"d088577125c72342c951cbeda831e10f4f40a72f": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "BUFFER_TEST HEAVY_CONTENTION READ_ONLY",
"description": "Scaling: Buffer Test - Test: Heavy Contention - Mode: Read Only",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 143212,
"raw_values": [
142410.24336900000344030559062957763671875,
142658.97163700000965036451816558837890625,
144568.00670599998557008802890777587890625
],
"test_run_times": [
112.18999999999999772626324556767940521240234375,
111.780000000000001136868377216160297393798828125,
112.849999999999994315658113919198513031005859375
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 35516,
"raw_values": [
35494.572037999998428858816623687744140625,
35538.170282999999471940100193023681640625
],
"test_run_times": [
104.400000000000005684341886080801486968994140625,
109.349999999999994315658113919198513031005859375,
0.1000000000000000055511151231257827021181583404541015625,
0.1000000000000000055511151231257827021181583404541015625
]
}
}
},
"7fda2c8031e473865874c39b8ba37efcccc384b0": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "MOSTLY_CACHE HEAVY_CONTENTION READ_WRITE",
"description": "Scaling: Mostly RAM - Test: Heavy Contention - Mode: Read Write",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 12277,
"raw_values": [
9483.0749250000008032657206058502197265625,
13520.292347000000518164597451686859130859375,
13014.86703699999998207204043865203857421875,
10312.590436000000408967025578022003173828125,
12732.59068900000056601129472255706787109375,
12866.281246000000464846380054950714111328125,
12623.7093349999995552934706211090087890625,
13017.466303999999581719748675823211669921875,
12922.369280000000799191184341907501220703125
],
"test_run_times": [
3355.670000000000072759576141834259033203125,
3361.69000000000005456968210637569427490234375,
3365.09999999999990905052982270717620849609375,
3447.53000000000020008883439004421234130859375,
3417.2899999999999636202119290828704833984375,
3402.25,
3334.9600000000000363797880709171295166015625,
3381.38999999999987267074175179004669189453125,
3400.239999999999781721271574497222900390625
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 2223,
"raw_values": [
2193.47582999999985986505635082721710205078125,
2252.46886399999993955134414136409759521484375
],
"test_run_times": [
2963.23000000000001818989403545856475830078125,
3060.90000000000009094947017729282379150390625,
0.1000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"f562bc9c757669fd20f1aa832e22b6648a239e2b": {
"identifier": "pts\/pgbench-1.9.1",
"title": "PostgreSQL pgbench",
"app_version": "12.0",
"arguments": "BUFFER_TEST HEAVY_CONTENTION READ_WRITE",
"description": "Scaling: Buffer Test - Test: Heavy Contention - Mode: Read Write",
"scale": "TPS",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 36148,
"raw_values": [
35851.4786890000032144598662853240966796875,
36723.38049199999659322202205657958984375,
35869.879702999998698942363262176513671875
],
"test_run_times": [
114.909999999999996589394868351519107818603515625,
115.7999999999999971578290569595992565155029296875,
115.0199999999999960209606797434389591217041015625
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 3645,
"raw_values": [
3546.72665300000016941339708864688873291015625,
3743.49384099999997488339431583881378173828125
],
"test_run_times": [
116.1099999999999994315658113919198513031005859375,
116.93999999999999772626324556767940521240234375,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"617870ab716342196c985657b9b14ba066caa04b": {
"identifier": "pts\/primesieve-1.7.0",
"title": "Primesieve",
"app_version": "7.4",
"description": "1e12 Prime Number Generation",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 8.644999999999999573674358543939888477325439453125,
"raw_values": [
8.592999999999999971578290569595992565155029296875,
8.6349999999999997868371792719699442386627197265625,
8.70700000000000073896444519050419330596923828125
],
"test_run_times": [
8.6500000000000003552713678800500929355621337890625,
8.699999999999999289457264239899814128875732421875,
8.769999999999999573674358543939888477325439453125
]
}
}
},
"0f845ac1a17cd256254d014ce39a409e29f136ab": {
"identifier": "pts\/renaissance-1.1.1",
"title": "Renaissance",
"app_version": "0.10.0",
"arguments": "dec-tree",
"description": "Test: Random Forest",
"scale": "ms",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 2090.60500000000001818989403545856475830078125,
"raw_values": [
2052.64499999999998181010596454143524169921875,
2060.73500000000012732925824820995330810546875,
2068.07900000000017826096154749393463134765625,
2141.05499999999983629095368087291717529296875,
2130.510000000000218278728425502777099609375
],
"test_run_times": [
91.3900000000000005684341886080801486968994140625,
91.06999999999999317878973670303821563720703125,
90.9200000000000017053025658242404460906982421875,
93.3599999999999994315658113919198513031005859375,
93.7399999999999948840923025272786617279052734375
]
}
}
},
"f0635bf087c8c22b967aa795a85e5a48a3da1376": {
"identifier": "pts\/renaissance-1.1.1",
"title": "Renaissance",
"app_version": "0.10.0",
"arguments": "als",
"description": "Test: Apache Spark ALS",
"scale": "ms",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 2042.3440000000000509317032992839813232421875,
"raw_values": [
2022.2899999999999636202119290828704833984375,
2043.138999999999896317603997886180877685546875,
2025.488000000000056388671509921550750732421875,
2061.90000000000009094947017729282379150390625,
2058.902000000000043655745685100555419921875
],
"test_run_times": [
68.400000000000005684341886080801486968994140625,
69.2900000000000062527760746888816356658935546875,
70.2399999999999948840923025272786617279052734375,
69.9200000000000017053025658242404460906982421875,
69.6200000000000045474735088646411895751953125
]
}
}
},
"48ca4e09e0763c37e3112e8028ef870a625c1da5": {
"identifier": "pts\/renaissance-1.1.1",
"title": "Renaissance",
"app_version": "0.10.0",
"arguments": "naive-bayes",
"description": "Test: Apache Spark Bayes",
"scale": "ms",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 680.5800000000000409272615797817707061767578125,
"raw_values": [
807.7569999999999481588019989430904388427734375,
798.7730000000000245563569478690624237060546875,
179,
797.597999999999956344254314899444580078125,
793.4370000000000118234311230480670928955078125,
807.5890000000000554791768081486225128173828125,
804.249000000000023646862246096134185791015625,
800.078999999999950887286104261875152587890625,
806.1670000000000300133251585066318511962890625,
815.0489999999999781721271574497222900390625,
812.0240000000000009094947017729282379150390625,
178.08699999999998908606357872486114501953125,
815.3229999999999790816218592226505279541015625,
822.115000000000009094947017729282379150390625,
182.955000000000012505552149377763271331787109375,
808.25199999999995270627550780773162841796875,
804.9740000000000463842297904193401336669921875,
806.9950000000000045474735088646411895751953125,
811.56899999999995998223312199115753173828125,
799.133000000000038198777474462985992431640625,
180.8990000000000009094947017729282379150390625,
798.7530000000000427462509833276271820068359375,
181.951999999999998181010596454143524169921875,
805.76800000000002910383045673370361328125,
796.008000000000038198777474462985992431640625
],
"test_run_times": [
36.1099999999999994315658113919198513031005859375,
35.75,
18.480000000000000426325641456060111522674560546875,
36.159999999999996589394868351519107818603515625,
35.63000000000000255795384873636066913604736328125,
35.8299999999999982946974341757595539093017578125,
36.219999999999998863131622783839702606201171875,
36.25999999999999801048033987171947956085205078125,
35.86999999999999744204615126363933086395263671875,
36.32000000000000028421709430404007434844970703125,
36.27000000000000312638803734444081783294677734375,
18.300000000000000710542735760100185871124267578125,
36.530000000000001136868377216160297393798828125,
36.18999999999999772626324556767940521240234375,
18.059999999999998721023075631819665431976318359375,
36.2000000000000028421709430404007434844970703125,
36.00999999999999801048033987171947956085205078125,
36.17999999999999971578290569595992565155029296875,
36.28999999999999914734871708787977695465087890625,
35.74000000000000198951966012828052043914794921875,
18.440000000000001278976924368180334568023681640625,
36.17999999999999971578290569595992565155029296875,
16.67999999999999971578290569595992565155029296875,
36.10000000000000142108547152020037174224853515625,
35.719999999999998863131622783839702606201171875
]
}
}
},
"a29989100c7301c7213b8f55d7200be7bebbc7fc": {
"identifier": "pts\/renaissance-1.1.1",
"title": "Renaissance",
"app_version": "0.10.0",
"arguments": "reactors",
"description": "Test: Savina Reactors.IO",
"scale": "ms",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 13140.992000000000189174897968769073486328125,
"raw_values": [
13284.6800000000002910383045673370361328125,
12295.9830000000001746229827404022216796875,
12437.686999999999898136593401432037353515625,
12797.264999999999417923390865325927734375,
16341.906000000000858562998473644256591796875,
12259.188000000000101863406598567962646484375,
15176.010000000000218278728425502777099609375,
13138.314000000000305590219795703887939453125,
13292.153000000000247382558882236480712890625,
13126.815000000000509317032992839813232421875,
11510.9680000000007566995918750762939453125,
12466.435999999999694409780204296112060546875,
12573.53900000000066938810050487518310546875,
10756.89700000000084401108324527740478515625,
13009.923000000000683940015733242034912109375,
14579.07300000000032014213502407073974609375,
15503.30700000000069849193096160888671875,
12281.290999999999257852323353290557861328125,
12169.81899999999950523488223552703857421875,
13818.576999999999316059984266757965087890625
],
"test_run_times": [
140,
123.5799999999999982946974341757595539093017578125,
120.7300000000000039790393202565610408782958984375,
139.740000000000009094947017729282379150390625,
149.5,
138.849999999999994315658113919198513031005859375,
148.259999999999990905052982270717620849609375,
139.75,
127.31000000000000227373675443232059478759765625,
135.280000000000001136868377216160297393798828125,
122.9500000000000028421709430404007434844970703125,
129.280000000000001136868377216160297393798828125,
135.81000000000000227373675443232059478759765625,
128.659999999999996589394868351519107818603515625,
137.740000000000009094947017729282379150390625,
139.19999999999998863131622783839702606201171875,
153.18999999999999772626324556767940521240234375,
126.2600000000000051159076974727213382720947265625,
137.830000000000012505552149377763271331787109375,
140.539999999999992041921359486877918243408203125
]
}
}
},
"173fffdad56a6293614ce6fbd99dd45c3743dd33": {
"identifier": "pts\/renaissance-1.1.1",
"title": "Renaissance",
"app_version": "0.10.0",
"arguments": "page-rank",
"description": "Test: Apache Spark PageRank",
"scale": "ms",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 3372.72100000000000363797880709171295166015625,
"raw_values": [
3433.44900000000006912159733474254608154296875,
3638.78400000000010550138540565967559814453125,
3785.20800000000008367351256310939788818359375,
3523.25,
3408.25300000000015643308870494365692138671875,
3388.45299999999997453414835035800933837890625,
3384.35899999999992360244505107402801513671875,
3422.009000000000014551915228366851806640625,
3022.12399999999979627318680286407470703125,
3351.8890000000001236912794411182403564453125,
3219.7730000000001382431946694850921630859375,
3328.78299999999990177457220852375030517578125,
3664.08800000000019281287677586078643798828125,
3225.58599999999978535925038158893585205078125,
3434.35399999999981446308083832263946533203125,
3308.72899999999981446308083832263946533203125,
3419.90900000000010550138540565967559814453125,
2997.56500000000005456968210637569427490234375,
3311.57299999999986539478413760662078857421875,
3678.13599999999996725819073617458343505859375,
3487.670000000000072759576141834259033203125,
3277.46700000000009822542779147624969482421875,
3118.052999999999883584678173065185546875,
3066.3429999999998472048901021480560302734375,
3422.22600000000011277734301984310150146484375
],
"test_run_times": [
88.93999999999999772626324556767940521240234375,
92.030000000000001136868377216160297393798828125,
93.1700000000000017053025658242404460906982421875,
90.349999999999994315658113919198513031005859375,
90.1099999999999994315658113919198513031005859375,
83.9800000000000039790393202565610408782958984375,
90.849999999999994315658113919198513031005859375,
89.280000000000001136868377216160297393798828125,
84.25,
88.5400000000000062527760746888816356658935546875,
87.969999999999998863131622783839702606201171875,
85.1099999999999994315658113919198513031005859375,
90.2600000000000051159076974727213382720947265625,
88.4599999999999937472239253111183643341064453125,
89.06999999999999317878973670303821563720703125,
88.1700000000000017053025658242404460906982421875,
89.06000000000000227373675443232059478759765625,
81.349999999999994315658113919198513031005859375,
88.93000000000000682121026329696178436279296875,
91.409999999999996589394868351519107818603515625,
89.31999999999999317878973670303821563720703125,
87.2000000000000028421709430404007434844970703125,
84.5400000000000062527760746888816356658935546875,
81.5499999999999971578290569595992565155029296875,
89.9500000000000028421709430404007434844970703125
]
}
}
},
"0a32a78f934d3c6f13e09c0f9aa99954b9a9ffc7": {
"identifier": "pts\/renaissance-1.1.1",
"title": "Renaissance",
"app_version": "0.10.0",
"arguments": "finagle-http",
"description": "Test: Twitter HTTP Requests",
"scale": "ms",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 4299.30199999999967985786497592926025390625,
"raw_values": [
4291.7200000000002546585164964199066162109375,
4243.5410000000001673470251262187957763671875,
4239.7349999999996725819073617458343505859375,
4328.1899999999995998223312199115753173828125,
4393.3249999999998181010596454143524169921875
],
"test_run_times": [
57.47999999999999687361196265555918216705322265625,
55.89999999999999857891452847979962825775146484375,
55.97999999999999687361196265555918216705322265625,
58.030000000000001136868377216160297393798828125,
57.35000000000000142108547152020037174224853515625
]
}
}
},
"fa60fd3ebf878d0448239bc9d7aaac894b28f0a8": {
"identifier": "pts\/renaissance-1.1.1",
"title": "Renaissance",
"app_version": "0.10.0",
"arguments": "db-shootout",
"description": "Test: In-Memory Database Shootout",
"scale": "ms",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 10996.26399999999921419657766819000244140625,
"raw_values": [
10945.0210000000006402842700481414794921875,
11118.51399999999921419657766819000244140625,
11054.468999999999141437001526355743408203125,
10806.78299999999944702722132205963134765625,
11056.533999999999650754034519195556640625
],
"test_run_times": [
179.94999999999998863131622783839702606201171875,
185.979999999999989768184605054557323455810546875,
186.1100000000000136424205265939235687255859375,
181.219999999999998863131622783839702606201171875,
187.1100000000000136424205265939235687255859375
]
}
}
},
"94b48082d6e5461c9bc3ee3a0f96cb51327bc7ab": {
"identifier": "pts\/renaissance-1.1.1",
"title": "Renaissance",
"app_version": "0.10.0",
"arguments": "akka-uct",
"description": "Test: Akka Unbalanced Cobwebbed Tree",
"scale": "ms",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 10986.720999999999548890627920627593994140625,
"raw_values": [
10719.1630000000004656612873077392578125,
11533.85599999999976716935634613037109375,
10969.89700000000084401108324527740478515625,
11676.35299999999915598891675472259521484375,
11295.75300000000061118043959140777587890625,
10675.3199999999997089616954326629638671875,
10541.69399999999950523488223552703857421875,
11036.40999999999985448084771633148193359375,
11031.3389999999999417923390865325927734375,
11047.671000000000276486389338970184326171875,
10693.012000000000625732354819774627685546875,
11366.072000000000116415321826934814453125,
11061.96299999999973806552588939666748046875,
10598.618000000000392901711165904998779296875,
10553.69800000000032014213502407073974609375
],
"test_run_times": [
267,
269.8799999999999954525264911353588104248046875,
266.66000000000002501110429875552654266357421875,
267.720000000000027284841053187847137451171875,
277.70999999999997953636921010911464691162109375,
261.509999999999990905052982270717620849609375,
262.69999999999998863131622783839702606201171875,
269.18999999999999772626324556767940521240234375,
270.529999999999972715158946812152862548828125,
268.41000000000002501110429875552654266357421875,
264.33999999999997498889570124447345733642578125,
272.3799999999999954525264911353588104248046875,
265.1299999999999954525264911353588104248046875,
263.66000000000002501110429875552654266357421875,
262.68999999999999772626324556767940521240234375
]
}
}
},
"c5bfa87584fc49593a2851c8dbab5eac0b0bb10b": {
"identifier": "pts\/renaissance-1.1.1",
"title": "Renaissance",
"app_version": "0.10.0",
"arguments": "future-genetic",
"description": "Test: Genetic Algorithm Using Jenetics + Futures",
"scale": "ms",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 1094.24800000000004729372449219226837158203125,
"raw_values": [
1101.548000000000001818989403545856475830078125,
1113.2629999999999199644662439823150634765625,
1098.642000000000052750692702829837799072265625,
1072.953999999999950887286104261875152587890625,
1084.83300000000008367351256310939788818359375
],
"test_run_times": [
56.68999999999999772626324556767940521240234375,
56.52000000000000312638803734444081783294677734375,
56.1099999999999994315658113919198513031005859375,
54.909999999999996589394868351519107818603515625,
55.72999999999999687361196265555918216705322265625
]
}
}
},
"46a591eee486b9dbaf87a9ef7f3f9554435f353d": {
"identifier": "pts\/stockfish-1.1.1",
"title": "Stockfish",
"app_version": "9",
"description": "Total Time",
"scale": "Nodes Per Second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 69222386,
"raw_values": [
69317222,
70747396,
67602540
],
"test_run_times": [
60.8599999999999994315658113919198513031005859375,
56.25,
58.81000000000000227373675443232059478759765625
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 11705849,
"raw_values": [
12763123,
10648575
],
"test_run_times": [
109.900000000000005684341886080801486968994140625,
120.56000000000000227373675443232059478759765625,
0.14000000000000001332267629550187848508358001708984375,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.1499999999999999944488848768742172978818416595458984375,
0.11000000000000000055511151231257827021181583404541015625
]
}
}
},
"6102c5c5a72a9c836f5047b0e1f6704e0865349f": {
"identifier": "pts\/svt-av1-2.2.1",
"title": "SVT-AV1",
"app_version": "0.8",
"arguments": "-enc-mode 8 -n 320 -i Bosphorus_1920x1080_120fps_420_8bit_YUV.yuv -w 1920 -h 1080",
"description": "Encoder Mode: Enc Mode 8 - Input: 1080p",
"scale": "Frames Per Second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 71.102000000000003865352482534945011138916015625,
"raw_values": [
69.33899999999999863575794734060764312744140625,
71.748999999999995225152815692126750946044921875,
72.2180000000000035242919693700969219207763671875
],
"test_run_times": [
8.53999999999999914734871708787977695465087890625,
7.87999999999999989341858963598497211933135986328125,
8.42999999999999971578290569595992565155029296875
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 15.3499999999999996447286321199499070644378662109375,
"raw_values": [
11.7309999999999998721023075631819665431976318359375,
18.9729999999999989768184605054557323455810546875
],
"test_run_times": [
31.67999999999999971578290569595992565155029296875,
18.260000000000001563194018672220408916473388671875,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"9dba57b02628dcad5b62fa47d8ab13ed936d7880": {
"identifier": "pts\/svt-hevc-1.1.0",
"title": "SVT-HEVC",
"app_version": "1.4.1",
"description": "1080p 8-bit YUV To HEVC Video Encode",
"scale": "Frames Per Second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 131.18999999999999772626324556767940521240234375,
"raw_values": [
130.659999999999996589394868351519107818603515625,
133.18999999999999772626324556767940521240234375,
129.729999999999989768184605054557323455810546875
],
"test_run_times": [
5.20999999999999996447286321199499070644378662109375,
5.12000000000000010658141036401502788066864013671875,
5.21999999999999975131004248396493494510650634765625
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 32.67999999999999971578290569595992565155029296875,
"raw_values": [
27.78999999999999914734871708787977695465087890625,
37.56000000000000227373675443232059478759765625
],
"test_run_times": [
23.10000000000000142108547152020037174224853515625,
16.78999999999999914734871708787977695465087890625,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"8ab5cc33552d96eaf0339bfa848f588594530579": {
"identifier": "pts\/svt-vp9-1.2.2",
"title": "SVT-VP9",
"app_version": "0.1",
"arguments": "-tune 2 -i Bosphorus_1920x1080_120fps_420_8bit_YUV.yuv -w 1920 -h 1080",
"description": "Tuning: VMAF Optimized - Input: Bosphorus 1080p",
"scale": "Frames Per Second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 250.1200000000000045474735088646411895751953125,
"raw_values": [
240,
254.669999999999987494447850622236728668212890625,
252.740000000000009094947017729282379150390625,
253.06000000000000227373675443232059478759765625
],
"test_run_times": [
3.359999999999999875655021241982467472553253173828125,
3.20999999999999996447286321199499070644378662109375,
3.2400000000000002131628207280300557613372802734375,
3.220000000000000195399252334027551114559173583984375
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 12.1899999999999995026200849679298698902130126953125,
"raw_values": [
5.769999999999999573674358543939888477325439453125,
18.6099999999999994315658113919198513031005859375
],
"test_run_times": [
104.8900000000000005684341886080801486968994140625,
33.969999999999998863131622783839702606201171875,
1.0500000000000000444089209850062616169452667236328125,
0.2600000000000000088817841970012523233890533447265625,
0.419999999999999984456877655247808434069156646728515625,
0.330000000000000015543122344752191565930843353271484375,
0.63000000000000000444089209850062616169452667236328125
]
}
}
},
"519503d976403687edcab94cc1ecd6e6b55a8fee": {
"identifier": "pts\/svt-vp9-1.2.2",
"title": "SVT-VP9",
"app_version": "0.1",
"arguments": "-tune 1 -i Bosphorus_1920x1080_120fps_420_8bit_YUV.yuv -w 1920 -h 1080",
"description": "Tuning: PSNR\/SSIM Optimized - Input: Bosphorus 1080p",
"scale": "Frames Per Second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 255.93999999999999772626324556767940521240234375,
"raw_values": [
254.780000000000001136868377216160297393798828125,
255.969999999999998863131622783839702606201171875,
257.06999999999999317878973670303821563720703125
],
"test_run_times": [
3.20000000000000017763568394002504646778106689453125,
3.20999999999999996447286321199499070644378662109375,
3.180000000000000159872115546022541821002960205078125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 82.31999999999999317878973670303821563720703125,
"raw_values": [
39.27000000000000312638803734444081783294677734375,
125.3700000000000045474735088646411895751953125
],
"test_run_times": [
73.969999999999998863131622783839702606201171875,
5.4000000000000003552713678800500929355621337890625,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.1000000000000000055511151231257827021181583404541015625
]
}
}
},
"72592d0f6d40e912844af3c4d07e2e8a83971bf8": {
"identifier": "pts\/svt-vp9-1.2.2",
"title": "SVT-VP9",
"app_version": "0.1",
"arguments": "-tune 0 -i Bosphorus_1920x1080_120fps_420_8bit_YUV.yuv -w 1920 -h 1080",
"description": "Tuning: Visual Quality Optimized - Input: Bosphorus 1080p",
"scale": "Frames Per Second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 229.509999999999990905052982270717620849609375,
"raw_values": [
230.150000000000005684341886080801486968994140625,
228.05000000000001136868377216160297393798828125,
230.330000000000012505552149377763271331787109375
],
"test_run_times": [
3.430000000000000159872115546022541821002960205078125,
3.45999999999999996447286321199499070644378662109375,
3.399999999999999911182158029987476766109466552734375
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 92.4200000000000017053025658242404460906982421875,
"raw_values": [
88.3299999999999982946974341757595539093017578125,
96.5100000000000051159076974727213382720947265625
],
"test_run_times": [
7.36000000000000031974423109204508364200592041015625,
6.769999999999999573674358543939888477325439453125,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"61f7806218ca246f1cf151f32a69e40ea45376b6": {
"identifier": "pts\/t-test1-1.0.1",
"title": "t-test1",
"app_version": "2017-01-13",
"arguments": "1",
"description": "Threads: 1",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 23.33500000000000085265128291212022304534912109375,
"raw_values": [
23.870999999999998664179656771011650562286376953125,
23.251000000000001222133505507372319698333740234375,
22.8840000000000003410605131648480892181396484375
],
"test_run_times": [
23.870000000000000994759830064140260219573974609375,
23.25,
22.879999999999999005240169935859739780426025390625
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 30.25,
"raw_values": [
30.79219794273399912754030083306133747100830078125,
29.713466882706001115366234444081783294677734375
],
"test_run_times": [
30.78999999999999914734871708787977695465087890625,
29.71000000000000085265128291212022304534912109375,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"5ea36db146ece8d78e82e6ec6526ac2bdae1f6db": {
"identifier": "pts\/t-test1-1.0.1",
"title": "t-test1",
"app_version": "2017-01-13",
"arguments": "2",
"description": "Threads: 2",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 130,
"raw_values": [
126.7339999999999946567186270840466022491455078125,
133.039999999999992041921359486877918243408203125,
129.9370000000000118234311230480670928955078125
],
"test_run_times": [
126.7300000000000039790393202565610408782958984375,
133.039999999999992041921359486877918243408203125,
129.93999999999999772626324556767940521240234375
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 3243,
"raw_values": [
3258.80092000959984943619929254055023193359375,
3227.60273194310002509155310690402984619140625
],
"test_run_times": [
3258.8000000000001818989403545856475830078125,
3227.59999999999990905052982270717620849609375,
0.11000000000000000055511151231257827021181583404541015625,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"317be2daeb2913752106421a5f105f2a4216240d": {
"identifier": "system\/tesseract-ocr-1.0.1",
"title": "Tesseract OCR",
"app_version": "v4.0.0.20181030",
"description": "Time To OCR 7 Images",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 53.88799999999999812416717759333550930023193359375,
"raw_values": [
53.46600000000000108002495835535228252410888671875,
54.11800000000000210320649784989655017852783203125,
54.07900000000000062527760746888816356658935546875
],
"test_run_times": [
53.469999999999998863131622783839702606201171875,
54.11999999999999744204615126363933086395263671875,
54.0799999999999982946974341757595539093017578125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 381.470000000000027284841053187847137451171875,
"raw_values": [
378.2477350235000130851403810083866119384765625,
384.6873660087600228507653810083866119384765625
],
"test_run_times": [
378.240000000000009094947017729282379150390625,
384.68000000000000682121026329696178436279296875,
0.190000000000000002220446049250313080847263336181640625,
0.13000000000000000444089209850062616169452667236328125
]
}
}
},
"62662fe4970af15356a2d2c6262ade3c5b300d37": {
"identifier": "pts\/mrbayes-1.4.0",
"title": "Timed MrBayes Analysis",
"app_version": "3.2.7",
"description": "Primate Phylogeny Analysis",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 401,
"raw_values": [
400.27899999999999636202119290828704833984375,
401.26699999999999590727384202182292938232421875,
400.09300000000001773514668457210063934326171875
],
"test_run_times": [
400.279999999999972715158946812152862548828125,
401.26999999999998181010596454143524169921875,
400.08999999999997498889570124447345733642578125
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 420,
"raw_values": [
422.0241589546200202676118351519107818603515625,
418.1404390335100060838158242404460906982421875
],
"test_run_times": [
422.01999999999998181010596454143524169921875,
418.1399999999999863575794734060764312744140625,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
},
"0cdfc6eea09327dc99a752fd0d9b763e3a56eeea": {
"identifier": "pts\/x264-2.6.1",
"title": "x264",
"app_version": "2019-12-17",
"description": "H.264 Video Encoding",
"scale": "Frames Per Second",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 183.340000000000003410605131648480892181396484375,
"raw_values": [
178.19999999999998863131622783839702606201171875,
186.44999999999998863131622783839702606201171875,
185.3600000000000136424205265939235687255859375
],
"test_run_times": [
3.569999999999999840127884453977458178997039794921875,
3.4199999999999999289457264239899814128875732421875,
3.439999999999999946709294817992486059665679931640625
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 43.56000000000000227373675443232059478759765625,
"raw_values": [
31.14999999999999857891452847979962825775146484375,
55.969999999999998863131622783839702606201171875
],
"test_run_times": [
19.71000000000000085265128291212022304534912109375,
10.8900000000000005684341886080801486968994140625,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875,
0.0899999999999999966693309261245303787291049957275390625,
0.1000000000000000055511151231257827021181583404541015625,
0.0899999999999999966693309261245303787291049957275390625
]
}
}
},
"87114e45f9a85f8303291ae0736e9ffa7f5f9dcf": {
"identifier": "pts\/compress-xz-1.1.0",
"title": "XZ Compression",
"app_version": "5.2.4",
"description": "Compressing ubuntu-16.04.3-server-i386.img, Compression Level 9",
"scale": "Seconds",
"proportion": "LIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 23.245000000000000994759830064140260219573974609375,
"raw_values": [
23.434999999999998721023075631819665431976318359375,
23.106999999999999317878973670303821563720703125,
23.193000000000001392663762089796364307403564453125
],
"test_run_times": [
23.42999999999999971578290569595992565155029296875,
23.1099999999999994315658113919198513031005859375,
23.190000000000001278976924368180334568023681640625
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 71.719999999999998863131622783839702606201171875,
"raw_values": [
71.9542999267579972411112976260483264923095703125,
71.485126972198003159064683131873607635498046875
],
"test_run_times": [
71.9500000000000028421709430404007434844970703125,
71.4800000000000039790393202565610408782958984375,
0.1000000000000000055511151231257827021181583404541015625,
0.1000000000000000055511151231257827021181583404541015625
]
}
}
},
"07cf1bee84d2966a28d7285e8311bec9427a245e": {
"identifier": "pts\/compress-zstd-1.2.0",
"title": "Zstd Compression",
"app_version": "1.4.5",
"arguments": "-b19",
"description": "Compression Level: 19",
"scale": "MB\/s",
"proportion": "HIB",
"display_format": "BAR_GRAPH",
"results": {
"ThreadR1-win10-ent-CCTA-CPU-Massive-21sep2020": {
"value": 40.7000000000000028421709430404007434844970703125,
"raw_values": [
40.89999999999999857891452847979962825775146484375,
40.39999999999999857891452847979962825775146484375,
40.7999999999999971578290569595992565155029296875
],
"test_run_times": [
69.599999999999994315658113919198513031005859375,
68.909999999999996589394868351519107818603515625,
69.599999999999994315658113919198513031005859375
]
},
"gnsft-win10-pro3-vm1-ccta-cpu-massive-8jan2021": {
"value": 16.35000000000000142108547152020037174224853515625,
"raw_values": [
11.0999999999999996447286321199499070644378662109375,
21.60000000000000142108547152020037174224853515625
],
"test_run_times": [
315.8999999999999772626324556767940521240234375,
238.719999999999998863131622783839702606201171875,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.0899999999999999966693309261245303787291049957275390625,
0.08000000000000000166533453693773481063544750213623046875,
0.08000000000000000166533453693773481063544750213623046875
]
}
}
}
}
}