A PHP Error was encountered

Severity: Warning

Message: fopen(/var/cpanel/php/sessions/ea-php80/ci_session75aebd7bbdbab9e72b648bcb565ba45589ba6940): Failed to open stream: No space left on device

Filename: drivers/Session_files_driver.php

Line Number: 174

Backtrace:

File: /home/judfadzm/public_html/contenttechseries.com/application/controllers/News.php
Line: 23
Function: library

File: /home/judfadzm/public_html/contenttechseries.com/index.php
Line: 317
Function: require_once

A PHP Error was encountered

Severity: Warning

Message: session_start(): Failed to read session data: user (path: /var/cpanel/php/sessions/ea-php80)

Filename: Session/Session.php

Line Number: 143

Backtrace:

File: /home/judfadzm/public_html/contenttechseries.com/application/controllers/News.php
Line: 23
Function: library

File: /home/judfadzm/public_html/contenttechseries.com/index.php
Line: 317
Function: require_once

MLCommons Introduces MLPerf Client v0.5

A New Benchmark for Consumer AI Performance

MLCommons Introduces MLPerf Client v0.5

Press Contacts
Kelly Berschauer
kelly@mlcommons.org

MLCommons®, the leading open engineering consortium dedicated to advancing machine learning (ML), is excited to announce the public release of the MLPerf® Client v0.5 benchmark. This benchmark sets a new standard for evaluating consumer AI performance, enabling users, press, and the industry to measure how effectively laptops, desktops, and workstations can run cutting-edge large language models (LLMs).

A Collaborative Effort by Industry Leaders

MLPerf Client represents a collaboration among technology leaders, including AMD, Intel, Microsoft, NVIDIA, Qualcomm Technologies, Inc., and top PC OEMs. These stakeholders have pooled resources and expertise to create a standardized benchmark, offering new insight into performance on key consumer AI workloads.

“MLPerf Client is a pivotal step forward in measuring consumer AI PC performance, bringing together industry heavyweights to set a new standard for evaluating generative AI applications on personal computers,” said David Kanter, Head of MLPerf at MLCommons.

Key Features of the MLPerf Client v0.5 benchmark:

  • AI model: The benchmark’s tests are based on Meta's Llama 2 7B large language model, optimized for reduced memory and computational requirements via 4-bit integer quantization.
  • Tests and metrics: Includes four AI tasks—content generation, creative writing, and text summarization of two different document lengths—evaluated using familiar metrics like time-to-first-token (TTFT) and tokens-per-second (TPS).
  • Hardware optimization: Supports hardware-accelerated execution on integrated and discrete GPUs via two distinct paths: ONNX Runtime GenAI and Intel OpenVINO.
  • Platform support: This initial release supports Windows 11 on x86-64 systems, with future updates planned for Windows on Arm and macOS.
  • Freely accessible: The benchmark is freely downloadable from MLCommons.org, empowering anyone to measure AI performance on supported systems.

Future Development

While version 0.5 marks the benchmark's debut, MLCommons plans to expand its capabilities in future releases, including support for additional hardware acceleration paths and a broader set of test scenarios incorporating a range of AI models.

Availability

The MLPerf Client v0.5 benchmark is available for download now from MLCommons. See the website for additional details on the benchmark’s hardware and software support requirements.

About MLCommons

MLCommons is the world leader in building benchmarks for AI. It is an open engineering consortium with a mission to make AI better for everyone through benchmarks and data. The foundation for MLCommons began with the MLPerf benchmarks in 2018, which rapidly scaled as a set of industry metrics to measure machine learning performance and promote transparency of machine learning techniques. In collaboration with its 125+ members, global technology providers, academics, and researchers, MLCommons is focused on collaborative engineering work that builds tools for the entire AI industry through benchmarks and metrics, public datasets, and measurements for AI risk and reliability.

For more information and details on becoming a member, please visit MLCommons.org or contact participation@mlcommons.org.


Read Previous

TTC Global Joins The Valuable Directory

Read Next

Generational Group Advises ITsynch, LLC

Add Comment