Menu

Unveiling the Power of GPUs: A Dive into ML Commons Benchmarks

December 12, 2023

Unveiling the Power of GPUs: A Dive into ML Commons Benchmarks

By Ray DeMoss, Support Manager

In the dynamic realm of machine learning, staying ahead of the curve requires not only innovative algorithms, but also cutting-edge hardware. ML Commons, a consortium of industry leaders, has emerged as a pivotal player in this space. In this blog post, we'll explore the significance of ML Commons, with a special focus on its compute and inferencing benchmarks tailored for edge scenarios.

At its core, ML Commons is a collaborative effort among industry giants to establish standardized benchmarks for machine learning tasks. These benchmarks provide a yardstick for evaluating the performance of GPUs, crucial components in the ever-evolving landscape of artificial intelligence.

ML Commons' compute benchmarks delve into the raw processing power of GPUs. These assessments go beyond mere clock speeds and delve into the intricate dance of parallel computations. As we navigate the compute benchmarks, we'll uncover the nuanced capabilities that make GPUs indispensable for intensive tasks, from model training to complex simulations.

GPUs

The true litmus test for any GPU lies in its ability to perform real-time inferencing, especially in edge computing scenarios. ML Commons' inferencing benchmarks focus precisely on this aspect. Whether it's enabling autonomous vehicles to make split-second decisions, or powering smart cameras for 360-degree situational awareness, the edge benchmarks provide a comprehensive view of a GPU's efficiency in real-world applications.

Why the emphasis on edge computing? As our world becomes increasingly connected, the demand for on-the-spot decision-making is paramount. ML Commons recognizes this trend, and by emphasizing edge benchmarks, it addresses the unique challenges and opportunities presented by decentralized computing.

The benchmarks set by ML Commons are not just technical metrics; they signal a shift in the industry landscape. Companies involved in GPU development now have a standardized framework for showcasing their products' capabilities, fostering healthy competition, and ultimately driving innovation.

In the dynamic landscape of machine learning, ML Commons stands as a beacon of collaboration, setting industry standards for GPU benchmarks. As we explore the intricacies of compute and inferencing benchmarks tailored for the edge, the true significance of these metrics comes to light.

Enter One Stop Systems (OSS), a company committed to pushing the boundaries of computational efficiency. Behind the scenes, OSS' engineering collaborated with cTuning and the ML Commons Task Force on Automation and Reproducibility, utilizing the CLI tools from the Collective Mind (CM) project.

Benchmarking

In collaboration with the ML Commons task force, our Rigel Edge Supercomputer system has achieved remarkable results in the inference at the edge category, showcasing the prowess of our flagship server. The benchmarks highlight an impressive throughput of 14,000 samples per second in the BERT natural language model category, paving the way for a new benchmark category in the upcoming MLPerf Inference v3.1 release focused on GPT-J, a large language model.

Navigating the complex landscape of benchmarking required expertise and collaboration. The journey involved leveraging the power of ML Commons' task force tools, led by cTuning, and tapping into the community on their public Discord server.

Our engineering team emphasizes the multifaceted nature of interpreting benchmark results, considering factors such as the model, dataset, quality target, and nuances between single-stream (latency) performance and offline samples per second. Power requirements are also a crucial aspect, with potential exploration into power monitoring systems compatible with ML Perf CLI tools.

As we embark on this project, the collaboration with ML Commons not only reinforces OSS' commitment to excellence, but also positions us at the forefront of technological innovation. Rigel's outstanding performance in ML Commons benchmarks is a testament to our dedication to providing cutting-edge solutions for the challenges posed by modern machine learning applications.

Join us in this journey as we continue to explore the intersections of hardware and benchmarks, contributing to the ever-growing landscape of AI and computational efficiency.

Click the buttons below to share this blog post!

Return to the main Blog page




Leave a comment

Comments will be approved before showing up.


Also in One Stop Systems Blog

Composable Infrastructure:  Dynamically Changing IT Infrastructure
Composable Infrastructure: Dynamically Changing IT Infrastructure

May 01, 2024

The evolution of IT infrastructure spans several decades and is marked by significant advancements in computing technology, networking, storage, and management practices. Data Centers have historically relied on Converged or Hyper-Converged infrastructures when deploying their hardware which proved to limited in flexibility, efficiency, scalability, and support for the Artificial Intelligence / Machine Learning (AI/ML) modern workloads of today. 

Continue Reading

Edge Computing
The Four Types of Edge Computing

April 17, 2024

“Edge Computing” is a term which has been widely adopted by the tech sector. Dominant leaders in accelerated computing have designated “Edge” as one of their fastest-growing segments, with FY24 revenue projected to be nearly $100 billion. The boom in the market for Edge Computing has become so significant that it is increasingly common to see companies create their own edge-related spinoff terms such as ‘Rugged Edge’, ‘Edge AI’, ‘Extreme Edge’, and a whole slew of other new buzzwords. 

Continue Reading

Datalogging in Autonomous Military
Unveiling the Strategic Edge: Datalogging in Autonomous Military Vehicles

March 11, 2024

The landscape of modern warfare is undergoing a profound transformation with the integration of cutting-edge technologies, and at the forefront of this evolution are autonomous military vehicles. Datalogging, a seemingly inconspicuous yet indispensable technology, plays a pivotal role in shaping the capabilities and effectiveness of these autonomous marvels. In this blog post, we delve into the critical role of datalogging in autonomous military vehicles and its impact on the future of defense strategies.

Continue Reading

You are now leaving the OSS website