spark svt, "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 2000 - Broadcast Inner Join Test Time", Lower Results Are Better "A",38.315052829683 "B",39.539676900953 "C",39.500209517777 "D",40.69029532373 "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 2000 - Inner Join Test Time", Lower Results Are Better "A",37.297719012946 "B",42.991230763495 "C",41.549481585622 "D",40.050418391824 "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 2000 - Repartition Test Time", Lower Results Are Better "A",34.28675962612 "B",36.685799844563 "C",34.329040184617 "D",37.230797365308 "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 2000 - Group By Test Time", Lower Results Are Better "A",45.701707974076 "B",46.884314518422 "C",43.424181077629 "D",45.958881795406 "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 2000 - Calculate Pi Benchmark Using Dataframe", Lower Results Are Better "A",6.0337099060416 "B",6.4092136323452 "C",6.0902453102171 "D",6.0242148637772 "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 2000 - Calculate Pi Benchmark", Lower Results Are Better "A",80.613989766687 "B",80.460167039186 "C",80.009097661823 "D",79.765531349927 "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 2000 - SHA-512 Benchmark Time", Lower Results Are Better "A",63.539483282715 "B",68.375926636159 "C",64.995817430317 "D",63.318618033081 "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 500 - Broadcast Inner Join Test Time", Lower Results Are Better "A",35.718209017068 "B",36.8903035447 "C",36.39296592772 "D",36.08638240397 "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 500 - Inner Join Test Time", Lower Results Are Better "A",35.655435044318 "B",35.864789601415 "C",37.654130730778 "D",36.938864368945 "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 500 - Repartition Test Time", Lower Results Are Better "A",35.697137672454 "B",34.164871565998 "C",35.638177663088 "D",34.349215596914 "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 500 - Group By Test Time", Lower Results Are Better "A",46.759768888354 "B",45.466531600803 "C",44.611227218062 "D",45.729973420501 "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 500 - Calculate Pi Benchmark Using Dataframe", Lower Results Are Better "A",6.0682108476758 "B",6.2600716240704 "C",6.1791776940227 "D",6.0353014990687 "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 500 - Calculate Pi Benchmark", Lower Results Are Better "A",80.685102038085 "B",81.195966094732 "C",80.432669304311 "D",80.066376529634 "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 500 - SHA-512 Benchmark Time", Lower Results Are Better "A",63.080436076969 "B",65.059325214475 "C",66.570126160979 "D",70.383650764823 "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 100 - Broadcast Inner Join Test Time", Lower Results Are Better "A",39.405277136713 "B",35.037506222725 "C",37.292689856142 "D",36.86741906777 "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 100 - Inner Join Test Time", Lower Results Are Better "A",38.222173400223 "B",36.852343700826 "C",36.676735404879 "D",37.702602669597 "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 100 - Repartition Test Time", Lower Results Are Better "A",37.612585153431 "B",33.431621715426 "C",34.683440189809 "D",36.69283612445 "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 100 - Group By Test Time", Lower Results Are Better "A",42.581896807998 "B",42.585700139403 "C",38.981267377734 "D",40.389442633837 "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 100 - Calculate Pi Benchmark Using Dataframe", Lower Results Are Better "A",6.0551611818373 "B",6.0887408554554 "C",5.898299947381 "D",6.1095719896257 "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 100 - Calculate Pi Benchmark", Lower Results Are Better "A",80.394511118531 "B",80.878542620689 "C",79.709967862815 "D",79.724962204695 "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 100 - SHA-512 Benchmark Time", Lower Results Are Better "A",79.478820774704 "B",61.734795160592 "C",60.380136072636 "D",71.765881694853 "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 1000 - Broadcast Inner Join Test Time", Lower Results Are Better "A",37.356575518847 "B",36.627149529755 "C",36.116306856275 "D",35.584342874587 "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 1000 - Inner Join Test Time", Lower Results Are Better "A",37.915543567389 "B",37.540734805167 "C",36.422617346048 "D",36.823922809213 "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 1000 - Repartition Test Time", Lower Results Are Better "A",35.822943214327 "B",35.269792396575 "C",34.773762021214 "D",34.937112510204 "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 1000 - Group By Test Time", Lower Results Are Better "A",45.874635223299 "B",43.701886277646 "C",41.484559092671 "D",42.189775951207 "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 1000 - Calculate Pi Benchmark Using Dataframe", Lower Results Are Better "A",6.3332592695951 "B",6.0360098406672 "C",6.2364275790751 "D",6.2560946382582 "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 1000 - Calculate Pi Benchmark", Lower Results Are Better "A",80.958371009678 "B",80.78790506348 "C",79.912269670516 "D",80.393130585551 "Apache Spark 3.3 - Row Count: 40000000 - Partitions: 1000 - SHA-512 Benchmark Time", Lower Results Are Better "A",71.674975372851 "B",63.029607411474 "C",63.357327248901 "D",65.187269687653 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 2000 - Broadcast Inner Join Test Time", Lower Results Are Better "A",21.603905517608 "B",22.239170137793 "C",21.416465699673 "D",22.394690196961 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 2000 - Inner Join Test Time", Lower Results Are Better "A",23.016819532961 "B",22.879152499139 "C",22.146266959608 "D",23.340843912214 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 2000 - Repartition Test Time", Lower Results Are Better "A",23.430894140154 "B",20.297288578004 "C",20.313922327012 "D",22.032922390848 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 2000 - Group By Test Time", Lower Results Are Better "A",23.474719312042 "B",23.705127395689 "C",22.785493545234 "D",22.54645948112 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 2000 - Calculate Pi Benchmark Using Dataframe", Lower Results Are Better "A",6.0953381098807 "B",6.0059632062912 "C",5.9197609610856 "D",6.0836590826511 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 2000 - Calculate Pi Benchmark", Lower Results Are Better "A",80.41997621581 "B",79.813857901841 "C",78.645136542618 "D",79.09362276271 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 2000 - SHA-512 Benchmark Time", Lower Results Are Better "A",39.284737557173 "B",39.188442289829 "C",36.551815953106 "D",37.405472487211 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 1000 - Broadcast Inner Join Test Time", Lower Results Are Better "A",20.165435411036 "B",19.469774994999 "C",19.447968862951 "D",19.770927663893 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 1000 - Inner Join Test Time", Lower Results Are Better "A",21.881266064942 "B",21.317617256194 "C",21.062580898404 "D",21.833729147911 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 1000 - Repartition Test Time", Lower Results Are Better "A",22.41803541407 "B",23.277137450874 "C",22.835460159928 "D",24.899873513728 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 1000 - Group By Test Time", Lower Results Are Better "A",21.56584809348 "B",22.621239624918 "C",23.777570314705 "D",19.232665196061 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 1000 - Calculate Pi Benchmark Using Dataframe", Lower Results Are Better "A",6.0413396954536 "B",6.091321092099 "C",6.1068482063711 "D",6.271550860256 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 1000 - Calculate Pi Benchmark", Lower Results Are Better "A",80.717050824314 "B",79.725453238934 "C",79.440891511738 "D",79.041680581868 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 1000 - SHA-512 Benchmark Time", Lower Results Are Better "A",38.633433971554 "B",41.147843506187 "C",35.63490916416 "D",36.370775222778 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 100 - Broadcast Inner Join Test Time", Lower Results Are Better "A",19.697398867458 "B",19.173225820065 "C",21.904143538326 "D",20.335954874754 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 100 - Inner Join Test Time", Lower Results Are Better "A",21.764107052237 "B",19.995737612247 "C",22.904757682234 "D",22.585571739823 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 100 - Repartition Test Time", Lower Results Are Better "A",23.156575795263 "B",20.078116618097 "C",21.411372870207 "D",19.570967875421 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 100 - Group By Test Time", Lower Results Are Better "A",21.053952220827 "B",20.288018546999 "C",18.532921612263 "D",18.981947921216 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 100 - Calculate Pi Benchmark Using Dataframe", Lower Results Are Better "A",6.0298572406173 "B",6.0339653715491 "C",5.9722554348409 "D",6.0195252671838 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 100 - Calculate Pi Benchmark", Lower Results Are Better "A",79.155712287873 "B",79.731256116182 "C",79.19144910574 "D",79.251065261662 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 100 - SHA-512 Benchmark Time", Lower Results Are Better "A",36.15040127933 "B",35.843670826405 "C",34.025679908693 "D",44.056317169219 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 500 - Broadcast Inner Join Test Time", Lower Results Are Better "A",18.500829365104 "B",18.124400619417 "C",20.07652611658 "D",18.794486802071 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 500 - Inner Join Test Time", Lower Results Are Better "A", "B",19.456130415201 "C",20.950773429126 "D",20.172776632011 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 500 - Repartition Test Time", Lower Results Are Better "A",19.180626787245 "B",23.494533833116 "C",22.412220384926 "D",19.200692374259 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 500 - Group By Test Time", Lower Results Are Better "A",21.823605563492 "B",19.574653789401 "C",20.642127964646 "D",21.588225759566 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 500 - Calculate Pi Benchmark Using Dataframe", Lower Results Are Better "A",6.2893201522529 "B",6.1652563139796 "C",6.0733442567289 "D",5.9068250618875 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 500 - Calculate Pi Benchmark", Lower Results Are Better "A",80.529722481966 "B",79.909820012748 "C",80.016460619867 "D",79.755643583834 "Apache Spark 3.3 - Row Count: 20000000 - Partitions: 500 - SHA-512 Benchmark Time", Lower Results Are Better "A",37.111184969544 "B",37.951387628913 "C",40.033498760313 "D",36.676244907081 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 2000 - Broadcast Inner Join Test Time", Lower Results Are Better "A",14.710479471833 "B",12.970103979111 "C",11.820764526725 "D",11.745072741061 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 2000 - Inner Join Test Time", Lower Results Are Better "A",17.077941592783 "B",14.952002875507 "C",15.383714895695 "D",14.546564303339 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 2000 - Repartition Test Time", Lower Results Are Better "A",13.866779241711 "B",14.258127775043 "C",13.463561218232 "D",13.259307671338 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 2000 - Group By Test Time", Lower Results Are Better "A",19.089781604707 "B",17.983034074306 "C",18.470326397568 "D",16.29902754724 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 2000 - Calculate Pi Benchmark Using Dataframe", Lower Results Are Better "A",6.1932056583464 "B",5.9340621195734 "C",6.1636243835092 "D",6.1835645027459 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 2000 - Calculate Pi Benchmark", Lower Results Are Better "A",80.73784404248 "B",80.665759969503 "C",79.914923511446 "D",80.06362458691 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 2000 - SHA-512 Benchmark Time", Lower Results Are Better "A",22.997755326331 "B",25.032268319279 "C",22.310313798487 "D",22.893547877669 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 1000 - Broadcast Inner Join Test Time", Lower Results Are Better "A",12.965712420642 "B",11.125276856124 "C",11.954779911786 "D",11.712664850056 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 1000 - Inner Join Test Time", Lower Results Are Better "A",13.867439482361 "B",13.759842433035 "C",14.396549869329 "D",13.187543336302 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 1000 - Repartition Test Time", Lower Results Are Better "A",15.373457409441 "B",13.645223516971 "C",14.852833334357 "D",14.343064095825 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 1000 - Group By Test Time", Lower Results Are Better "A",17.085118696094 "B",16.50211307779 "C",15.061644233763 "D",14.196643583477 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 1000 - Calculate Pi Benchmark Using Dataframe", Lower Results Are Better "A",6.1840057894588 "B",6.2846188396215 "C",6.4969418868423 "D",6.3717419058084 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 1000 - Calculate Pi Benchmark", Lower Results Are Better "A",80.864952377975 "B",80.05961817503 "C",79.266427919269 "D",79.837997347116 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 1000 - SHA-512 Benchmark Time", Lower Results Are Better "A",22.556307923049 "B",22.143037606031 "C",21.853141844273 "D",21.017762392759 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 100 - Broadcast Inner Join Test Time", Lower Results Are Better "A",11.665331747383 "B",10.563611589372 "C",11.115546371788 "D",10.930716358125 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 100 - Inner Join Test Time", Lower Results Are Better "A",13.509580601007 "B",11.793500136584 "C",13.108095884323 "D",13.345734227449 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 100 - Repartition Test Time", Lower Results Are Better "A",12.307953778654 "B",16.716047871858 "C",15.713994789869 "D",16.340572435409 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 100 - Group By Test Time", Lower Results Are Better "A",14.54570755735 "B",15.099411081523 "C",15.176151808351 "D",17.507809083909 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 100 - Calculate Pi Benchmark Using Dataframe", Lower Results Are Better "A",6.403102260083 "B",6.2749464027584 "C",6.3119843266904 "D",6.4333193972707 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 100 - Calculate Pi Benchmark", Lower Results Are Better "A",78.801551513374 "B",79.50064823404 "C",78.438637536019 "D",78.889388151467 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 100 - SHA-512 Benchmark Time", Lower Results Are Better "A",22.377060249448 "B",21.425002679229 "C",24.0651255548 "D",24.393461380154 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 500 - Broadcast Inner Join Test Time", Lower Results Are Better "A",12.278462275863 "B",11.231294848025 "C",12.292197629809 "D",10.812617436051 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 500 - Inner Join Test Time", Lower Results Are Better "A",12.555087905377 "B",12.166659213603 "C",14.147518336773 "D",13.234085656703 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 500 - Repartition Test Time", Lower Results Are Better "A",15.125815551728 "B",13.722000364214 "C",12.643913395703 "D",15.261011961848 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 500 - Group By Test Time", Lower Results Are Better "A",15.299718938768 "B",19.535005953163 "C",14.997071053833 "D",15.499287869781 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 500 - Calculate Pi Benchmark Using Dataframe", Lower Results Are Better "A",6.2350364997983 "B",6.1385622359812 "C",6.2242169752717 "D",6.5597854405642 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 500 - Calculate Pi Benchmark", Lower Results Are Better "A",79.677550718188 "B",79.441398814321 "C",79.266380991787 "D",79.223242860287 "Apache Spark 3.3 - Row Count: 10000000 - Partitions: 500 - SHA-512 Benchmark Time", Lower Results Are Better "A",21.912073552608 "B",22.513441499323 "C",22.000520613045 "D",21.952058464289 "SVT-AV1 1.2 - Encoder Mode: Preset 4 - Input: Bosphorus 4K", Higher Results Are Better "A", "B", "C", "D", "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 2000 - Broadcast Inner Join Test Time", Lower Results Are Better "A",5.618902053684 "B",7.742389626801 "C",4.5743801444769 "D",5.710038471967 "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 2000 - Inner Join Test Time", Lower Results Are Better "A",5.7390008568764 "B",5.8662004694343 "C",6.5207657441497 "D",5.7533586807549 "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 2000 - Repartition Test Time", Lower Results Are Better "A",6.1928059570491 "B",7.1873952820897 "C",5.7593834511936 "D",6.9742112495005 "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 2000 - Group By Test Time", Lower Results Are Better "A",10.054900940508 "B",10.774214066565 "C",10.133694924414 "D",10.014567747712 "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 2000 - Calculate Pi Benchmark Using Dataframe", Lower Results Are Better "A",6.5680596493185 "B",6.2013646177948 "C",6.2078235000372 "D",6.3213053941727 "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 2000 - Calculate Pi Benchmark", Lower Results Are Better "A",78.837156817317 "B",79.729975916445 "C",78.701980803162 "D",78.499752309173 "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 2000 - SHA-512 Benchmark Time", Lower Results Are Better "A",8.3132490813732 "B",8.6139228418469 "C",7.7104900367558 "D",8.56266637519 "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 1000 - Broadcast Inner Join Test Time", Lower Results Are Better "A",3.016429271549 "B",4.061107557267 "C",3.2494365386665 "D",3.277701638639 "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 1000 - Inner Join Test Time", Lower Results Are Better "A",4.530955504626 "B",5.0097604617476 "C",5.4698095656931 "D",4.3336564712226 "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 1000 - Repartition Test Time", Lower Results Are Better "A",5.4264044798911 "B",5.6844824589789 "C",5.2972730249166 "D",9.2187364138663 "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 1000 - Group By Test Time", Lower Results Are Better "A",9.9928023405373 "B",10.811930652708 "C",10.416560474783 "D",10.146551273763 "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 1000 - Calculate Pi Benchmark Using Dataframe", Lower Results Are Better "A",6.0849482528865 "B",6.3985552489758 "C",6.5412993915379 "D",6.1059712134302 "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 1000 - Calculate Pi Benchmark", Lower Results Are Better "A",78.879806503654 "B",78.83965587616 "C",78.555115543306 "D",78.623805981129 "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 1000 - SHA-512 Benchmark Time", Lower Results Are Better "A",7.7629738636315 "B",7.4345077648759 "C",8.7471137233078 "D",7.81341002509 "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 500 - Broadcast Inner Join Test Time", Lower Results Are Better "A",3.1634489260614 "B",3.5612530186772 "C",4.4745620302856 "D",3.110276132822 "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 500 - Inner Join Test Time", Lower Results Are Better "A",4.3996814712882 "B",4.5787084661424 "C",4.4311352707446 "D",5.6198441162705 "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 500 - Repartition Test Time", Lower Results Are Better "A",5.0908580198884 "B",5.9842699579895 "C",7.2901867553592 "D",5.8300270438194 "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 500 - Group By Test Time", Lower Results Are Better "A",9.848550606519 "B",9.2976200915873 "C",9.5318540744483 "D",9.9922301210463 "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 500 - Calculate Pi Benchmark Using Dataframe", Lower Results Are Better "A",6.0199620015919 "B",6.4172042421997 "C",6.2598686292768 "D",6.3010820932686 "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 500 - Calculate Pi Benchmark", Lower Results Are Better "A",78.340095706284 "B",79.10664922744 "C",78.535377927125 "D",78.644196435809 "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 500 - SHA-512 Benchmark Time", Lower Results Are Better "A",8.5651164241135 "B",7.5659930370748 "C",7.7093941345811 "D",7.820338472724 "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 100 - Broadcast Inner Join Test Time", Lower Results Are Better "A",3.0975380167365 "B",2.5627955012023 "C",2.0463072881103 "D",2.9121210202575 "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 100 - Inner Join Test Time", Lower Results Are Better "A",3.6529255695641 "B",5.0059627406299 "C",3.9693906977773 "D",3.5246142968535 "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 100 - Repartition Test Time", Lower Results Are Better "A",4.6412390395999 "B",3.8744418509305 "C",4.7655016705394 "D",7.9425828494132 "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 100 - Group By Test Time", Lower Results Are Better "A",9.2423383519053 "B",8.9342361576855 "C",9.2500656805933 "D",9.5622421614826 "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 100 - Calculate Pi Benchmark Using Dataframe", Lower Results Are Better "A",6.0538418702781 "B",6.3552062660456 "C",6.2623119987547 "D",6.2588018588722 "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 100 - Calculate Pi Benchmark", Lower Results Are Better "A",78.680011417717 "B",80.085283178836 "C",79.562627512962 "D",78.794594027102 "Apache Spark 3.3 - Row Count: 1000000 - Partitions: 100 - SHA-512 Benchmark Time", Lower Results Are Better "A",7.6243781037629 "B",7.1580970138311 "C",7.6932943612337 "D",7.5489463359118 "SVT-AV1 1.2 - Encoder Mode: Preset 4 - Input: Bosphorus 1080p", Higher Results Are Better "A", "B", "C", "D", "SVT-AV1 1.2 - Encoder Mode: Preset 8 - Input: Bosphorus 4K", Higher Results Are Better "A", "B", "C", "D", "SVT-AV1 1.2 - Encoder Mode: Preset 10 - Input: Bosphorus 4K", Higher Results Are Better "A", "B", "C", "D", "SVT-AV1 1.2 - Encoder Mode: Preset 12 - Input: Bosphorus 4K", Higher Results Are Better "A", "B", "C", "D", "SVT-AV1 1.2 - Encoder Mode: Preset 8 - Input: Bosphorus 1080p", Higher Results Are Better "A", "B", "C", "D", "SVT-AV1 1.2 - Encoder Mode: Preset 10 - Input: Bosphorus 1080p", Higher Results Are Better "A", "B", "C", "D", "SVT-AV1 1.2 - Encoder Mode: Preset 12 - Input: Bosphorus 1080p", Higher Results Are Better "A", "B", "C", "D",