The recent launch of a new tool aimed at assessing device capabilities in artificial intelligence is an exciting development. With the increasing prevalence of AI in technology, benchmarks that accurately reflect performance are more essential than ever. The latest addition to this domain is a tool from Primate Labs that offers users a comprehensive evaluation of their devices' AI performance through a range of tests.
The newly introduced Geekbench AI 1.0 platform measures the AI capabilities of various devices. This benchmarking suite, which is accessible for free on multiple platforms, provides developers with tools to effectively test their device's performance across different parameters. By conducting extensive tests on the CPU, GPU, and neural processing unit, it produces a clear performance score, allowing users to gauge their device's AI proficiency.
While announcing the launch, Primate Labs emphasized that this new suite applies a meticulous testing approach tailored for machine learning and deep learning tasks. It boasts the same cross-platform functionality and real-world relatability that users expect from their benchmarking tools.
The application is programmed to automatically execute ten diverse AI workloads, each involving the processing of three distinct data types. This thorough testing allows for a detailed analysis of on-device AI capabilities. Compatibility extends across a variety of operating systems, including Android, iOS, Linux, macOS, and Windows, making it suitable for smartphones, tablets, laptops, and desktops alike.
The previous version, initially called Geekbench ML, was rebranded to align with the industry's growing use of the term AI in discussions about these types of workloads. The app is designed to navigate the complexities involved in evaluating AI performance by taking into account not only the device’s workloads but also its hardware and chosen AI framework.
Geekbench AI evaluates devices for both speed and accuracy to ascertain any potential compromises between performance and efficiency. In addition to these benchmarks, metrics pertaining to datasets, frameworks, runtime, computer vision, natural language processing, and more are also included.
To further enhance user capabilities, Primate Labs has introduced an ML Benchmarks leaderboard to provide insights into the performance of the CPU, GPU, and NPU across various devices. This feature allows users to identify top-performing devices easily. To run this app, devices must meet specific software requirements, notably Android 12 or later.