Simplyblock, TTA, UBIX, IBM, WDC, and YanRong.
Check the results here:
mlcommons.org/benchmarks/s...
#MLPerf #AI #Storage #Benchmarking #MachineLearning #MLCommons
@mlcommons.org.bsky.social
MLCommons is an AI engineering consortium, built on a philosophy of open collaboration to improve AI systems. Through our collective engineering efforts, we continually measure and improve AI technologies' accuracy, safety, speed, and efficiency.
Simplyblock, TTA, UBIX, IBM, WDC, and YanRong.
Check the results here:
mlcommons.org/benchmarks/s...
#MLPerf #AI #Storage #Benchmarking #MachineLearning #MLCommons
5/ Congratulations and thanks to all submitters!
Alluxio, Argonne National Lab, DDN, ExponTech, FarmGPU, H3C, Hammerspace, HPE, JNIST/Huawei, Juicedata, Kingston, KIOXIA, Lightbits Labs, MangoBoost, Micron, Nutanix, Oracle, Quanta Cloud Technology, Samsung, Sandisk,
4/ The v2.0 submissions showcase a wide range of technical solutionsβlocal & object storage, in-storage accelerators, software-defined storage, block systems, and more. This diversity highlights the communityβs commitment to advancing AI infrastructure.
04.08.2025 17:36 β π 0 π 0 π¬ 1 π 03/ New in this round: checkpoint benchmarks, designed to reflect real-world practices in large-scale AI training systems. These benchmarks provide key data to help stakeholders optimize system reliability and efficiency at scale.
04.08.2025 17:36 β π 0 π 0 π¬ 1 π 0 v2.0 highlights:
- 200+ results
- 26 organizations
- 7 countries represented
- Benchmarked systems now support about 2x the accelerators vs v1.0
1/ MLCommons just released results for the MLPerf Storage v2.0 benchmarkβan industry-standard suite for measuring storage system performance in #ML workloads. This benchmark remains architecture-neutral, representative, and reproducible.
mlcommons.org/2025/08/mlpe...
MLPerf Client v1.0 is out! π
The new benchmark for LLMs on PCs and client systems is now availableβfeaturing expanded model support, new workload scenarios, and broad hardware integration.
Thank you to all submitters! #AMD, #Intel, @microsoft.com, #NVIDIA, #Qualcomm
mlcommons.org/2025/07/mlpe...
You can read more details here: mlcommons.org/2025/07/mlpe...
10.07.2025 19:11 β π 0 π 0 π¬ 0 π 0MLCommons just launched MLPerf Mobile on the Google Play Store! π±
Benchmark your Android deviceβs AI performance on real-world ML tasks with this free, open-source app.
Try it now: play.google.com/store/apps/d...
3/3
27.06.2025 19:07 β π 0 π 0 π¬ 0 π 03/3 @cam.ac.uk , @ox.ac.uk, University of Illinois Urbana-Champaign, and @ucsb.bsky.social.
Read more about the collaborative development of the Agentic Reliability Evaluation Standard and opportunities to participate: mlcommons.org/2025/06/ares...
2/3 Contributions from: Advai, AI Verify Foundation, @anthropic.com, @arize.bsky.social , @cohere.com , Google, Intel, LNE, Meta, @microsoft.com, NASSCOM, OpenAI, Patronus AI, @polymtl.bsky.social, Qualcomm, QuantumBlack - AI by McKinsey, Salesforce, Schmidt Sciences, @servicenow.bsky.social,
27.06.2025 19:07 β π 0 π 0 π¬ 1 π 0Today, MLCommons is announcing a new collaboration with contributors from across academia, civil society, and industry to co-develop an open agent reliability evaluation standard to operationalize trust in agentic deployments.
πhttps://mlcommons.org/2025/06/ares-announce/
1/3
We're all about acceleration! π
Watch @priya-kasimbeg.bsky.social & @fsschneider.bsky.social speedrun an explanation of the AlgoPerf benchmark, rules, and results all within a tight 5 minutes for our #ICLR2025 paper video on "Accelerating Neural Network Training". See you in Singapore!
Companies are deploying AI tools that haven't been pressure-tested, and it's already backfiring.
In her new op-ed, our President, Rebecca Weiss, breaks down how industry-led AI reliability standards can help executives avoid costly, high-profile failures.
π More: bit.ly/3FP0kjg
@fastcompany.com
Call for Submissions!
#MLCommons & @AVCConsortium are accepting submissions for the #MLPerf Automotive Benchmark Suite! Help drive fair comparisons & optimize AI systems in vehicles. Focus is on camera sensor perception.
π
Submissions close June 13th, 2025
Join: mlcommons.org/community/su...
4/ Read more and check out the full results here:
πhttps://mlcommons.org/2025/06/mlperf-training-v5-0-results/
#MLPerf #MLCommons #AI #MachineLearning #Benchmarking
3/ MLPerf Training v5.0 introduces the Llama 3.1 405B benchmark, our largest language model yet. We also saw big performance gains for Stable Diffusion and Llama 2.0 70B LoRAβAI training is getting faster and smarter.
04.06.2025 15:35 β π 0 π 0 π¬ 1 π 02/ Thank you to all 20 submitters for driving progress in AI benchmarking:
AMD, ASUSTeK, Cisco, CoreWeave, Dell, GigaComputing, Google Cloud, HPE, IBM, Krai, Lambda, Lenovo, MangoBoost, Nebius, NVIDIA, Oracle, QCT, SCITIX, Supermicro, TinyCorp.
1/ The MLPerf Training v5.0 results are hereβLetβs have a fresh look at the state of large-scale AI training! This round set a new record: 201 performance results from across the industry.
πhttps://mlcommons.org/2025/06/mlperf-training-v5-0-results/
Call for papers!
We are organising the 1st Workshop on Multilingual Data Quality Signals with @mlcommons.org and @eleutherai.bsky.social, held in tandem with @colmweb.org. Submit your research on multilingual data quality!
Submission deadline is 23 June, more info: wmdqs.org
MLCommons is partnering with Nasscom to develop globally recognized AI reliability benchmarks, including India-specific, Hindi-language evaluations. Together, we are advancing trustworthy AI.
π mlcommons.org/2025/05/nass...
#AIForAll #IndiaAI #ResponsibleAI #Nasscom #MLCommons
MLCommons' MLPerf Training suite has a new #pretraining #benchmark based on #Metaβs Llama 3.1 405B model. We use the same dataset with a bigger model and longer context, offering a more relevant and challenging measure for todayβs #AI systems. mlcommons.org/2025/05/trai...
05.05.2025 16:22 β π 0 π 0 π¬ 0 π 0As AI models grow, storage is key to #ML performance. MLCommons' @dkanter.bsky.social joins #Nutanixβs Tech Barometer podcast to explain why and how the #MLPerf #Storage #benchmark guides smarter #data #infrastructure for #AI.
Listen: www.nutanix.com/theforecastb...
#DataStorage #EnterpriseIT
3/ We want to thank all the participants: #Intel, #Microsoft,
#NVIDIA, #Qualcomm Technologies.
#MLPerf #MLCommons #Client
2/ This update broadens hardware compatibility and introduces improved device selection and updated software components, providing a transparent and standardized approach to measuring AI performance across next-generation platforms.
28.04.2025 15:12 β π 0 π 0 π¬ 1 π 01/ MLCommons announces the release of MLPerf Client v0.6, the first open benchmark to support NPU and GPU acceleration on consumer AI PCs.
Read more: mlcommons.org/2025/04/mlpe...
#MLCommons just released two new French prompt #datasets for #AILuminate:
πΉDemo set: 1,200+ prompts, free for AI safety testing
πΉPractice set: 12,000 prompts for deeper evaluation (on request)
Native speakers made both and are ready for #ModelBench. Details: mlcommons.org/2025/04/ailu...
#AI #AIRR
"Whatβs admirable about MLPerf is that everything is shared and benchmarks are open sourced. Results need to be reproducible β no mystery can remain. This openness allows for more dynamic comparisons beyond raw side-by-side speed, like performance..."
newsroom.intel.com/artificial-i...
We also want to thank the additional technical contributors: Pablo Gonzalez, MLCommons; Anandhu Sooraj, MLCommons; Arjun Suresh, AMD (formerly at MLCommons)
07.04.2025 21:32 β π 0 π 0 π¬ 0 π 0