MLPerf Inference 6.0: Software Gains & Broadening Competition Shake Things Up

MLCommons released MLPerf Inference v6.0 results, marking what the consortium describes as the most significant update to the benchmark suite to date.
The round introduced five new workloads, including a multimodal vision-language model, a text-to-video generation benchmark, and a new interactive scenario for the DeepSeek-R1 reasoning model.
Research Note: MLPerf Inference 4.0 Results
MLCommons released the results of its MLPerf Inference v4.0 benchmarks, which introduced two new workloads, Llama 2 and Stable Diffusion XL.
Since its inception in 2018, MLPerf has established itself as a crucial benchmark in the accelerator market. The benchmarks offer detailed comparisons across a variety of system configurations for specific use cases.
NVIDIA H100 Dominates New MLPerf v3.0 Benchmark Results

To know how a system performs across a range of AI workloads, you look at its MLPerf benchmark numbers. AI is rapidly evolving, with generative AI workloads becoming increasingly prominent, and MLPerf is evolving with the industry. Its new MLPerf Training v3.0 benchmark suite introduces new tests for recommendation engines and large language model (LLM) […]