|
SPECworkstation 4.0 Benchmark Measures Latest Workstation Hardware, Adds AI/ML WorkloadsBy Chandra Sakthivel, SPECwpc ChairI’m extremely excited that SPEC has announced the release of the SPECworkstation 4.0 benchmark, a major update to SPEC’s comprehensive tool designed to measure all key aspects of workstation performance. This significant upgrade from version 3.1 incorporates cutting-edge features to keep pace with the latest workstation hardware and the evolving demands of professional applications, including the increasing reliance on AI and machine learning (ML). Workstations are crucial for a range of industries, including architecture, engineering & construction (AEC), energy, life sciences, product design, media & entertainment, scientific research, and software development. The new SPECworkstation 4.0 benchmark provides a robust, real-world measure of CPU, graphics, accelerator, and disk performance, ensuring that professionals have the data they need to make informed decisions about their hardware investments. Developed by a consortium of semiconductor vendors and system manufacturers working as members of the non-profit SPEC organization, the benchmark is an unbiased, real-world, application-driven tool for measuring performance across a wide range of workstation-based workloads. Let’s look in detail at what the new SPECworkstation 4.0 benchmark offers, the importance of its support for AI/ML, and the benefits it brings to the professional computing community. What’s new in the SPECworkstation 4.0 benchmarkThe SPECworkstation 4.0 benchmark caters to the diverse needs of engineers, scientists, and developers who rely on workstation hardware for daily tasks. The benchmark includes real-world applications like Blender, Handbrake, LLVM and more, providing a comprehensive performance measure across various scenarios, and spanning seven different industry verticals, each focusing on specific use cases and subsystems critical to workstation users. Compared to the SPECworkstation 3.1 benchmark, the SPECworkstation 4.0 benchmark provides a more accurate, detailed, and relevant measure of the performance of today’s workstations, especially when it comes to AI/ML. Key new features of the benchmark include:
All-in on AI/ML and Data AnalyticsThe release of the SPECworkstation 4.0 benchmark holds significant value for the user community, particularly those wanting to take advantage of rapidly evolving AI technologies. As enterprises increasingly focus on building infrastructure to support AI-powered use cases, having a reliable benchmark is more critical than ever as more data scientists and software developers rely on workstations for data science, data analytics, and ML. Until now, there hasn’t been a comprehensive benchmark that accurately measures a workstation's performance across the entire data science pipeline – from data acquisition and cleaning to ETL processes and algorithm development. However, data scientists often use workstations to experiment with different algorithms and solutions before scaling to server-based model training. The workstation's performance at a system level is crucial, especially when dealing with terabytes of data. This new benchmark addresses these challenges by providing workloads designed to measure performance across the data science pipeline. It provides the first set of performance metrics specifically designed to measure data science and ONNX-based inference workloads on workstations and serves as a useful tool for estimating the performance of workstation hardware in real-world AI applications. ONNX Runtime (ONNX RT) is a high-performance inference engine that enables the deployment of ML models in various environments, including cloud, edge, and mobile devices. Developed by Microsoft, it supports models trained in multiple frameworks like PyTorch and TensorFlow, allowing seamless execution across different hardware platforms. ONNX RT is optimized for speed and efficiency and supports a wide variety of platforms, making it an attractive choice for running AI models in production and at scale. The SPECworkstation 4.0 benchmark evaluates inference latency and throughput performance across different quantization approaches and batch sizes, including FP32, FP16, and INT8. This reflects the industry's ongoing evolution towards more efficient AI models. By offering a comprehensive set of performance metrics, this benchmark helps scientists and developers understand the capabilities of their workstation hardware and enables them to make informed decisions to optimize their systems for better performance in real-world AI/ML applications. The benchmark also covers classical machine learning workloads by measuring performance with widely used libraries like XGBoost and Scikit-learn. This further ensures the benchmark is relevant for a wide range of AI/ML development tasks, from data pre-processing to algorithm design. Available for Immediate DownloadDownloading and installing the SPECworkstation 4.0 benchmark is simple and straightforward, and you can begin running tests within a few minutes. You can also check out the benchmark’s published results page as they become available to compare your performance with that of other systems, or to research a system that will best balance your performance and budget requirements. The SPECworkstation 4.0 benchmark is available from SPEC under a two-tiered pricing structure: free for the user community and $5,000 for sellers of computer-related products and services. SPEC/GWPG members receive benchmark licenses as a membership benefit. If you're a SPECworkstation user, I highly recommend that you download this latest release. |