Saturday, December 7, 2024

MinIO Fuels AI on Arm-Based Chipsets

Related stories

Alibaba Launches ‘Pic Copilot’ AI E-Commerce Tool

Alibaba International Digital Commerce Group is excited to announce...

Tractian Secures $120M to Reduce Industrial Downtime

Led by Sapphire Ventures, the round will enable Tractian...

Unisys Appoints Michael M. Thomson as CEO

The Unisys Board of Directors announced that it unanimously elected...

Veritone Expands Enterprise AI Solutions

Veritone, Inc, a leader in designing human-centered AI solutions,...
spot_imgspot_img

MinIO, the leader in high-performance storage for AI, announced compelling new optimizations and benchmarks for Arm-based chipsets powering on its object store. These optimizations underscore the relevance of low power, computationally dense chips in key AI-related tasks and demonstrates the readiness and superiority of the Arm architecture for modern AI and data processing workloads including erasure coding, bit rot protection and encryption.

MinIO’s work leveraged the latest Scalable Vector Extension Version (SVE) enhancements. SVE improves the performance and efficiency of vector operations, which are crucial for high-performance computing, artificial intelligence, machine learning and other data-intensive applications.

Specifically, MinIO was able to enhance its existing implementation of its Reed Solomon erasure coding library. The result was 2x faster throughput as compared to the previous NEON instruction set implementation, as is shown in this graph:

Just as significantly, the new code uses just a quarter of the available cores (16) to consume half the memory bandwidth. Previously it required 32 cores to achieve this same memory bandwidth consumption.

Also Read: Validation Cloud Secures $10M Lead to Scale AI for Web3

Highway Hash

MinIO was also able to achieve significant performance enhancements for the Highway Hash algorithm it uses for bit-rot detection. This is frequently run during both GET (read) as well as PUT (write) operations. SVE support for the core hashing update function delivered the following results:

The performance scales linearly as the core count goes up, and for the larger block sizes, it starts to reach the memory bandwidth limit around 50 to 52 cores.

Further, SVE offers support for lane masking via predicated execution, allowing more efficient utilization of the vector units. Extensive support for scatter and gather instructions makes it possible to efficiently access memory in a flexible manner.

“Delivering performance gains while improving power efficiency is critical to building a sustainable infrastructure for the modern, intensive data processing workloads demanded by AI,” said Eddie Ramirez, vice president of marketing and ecosystem development, Infrastructure Line of Business, Arm. “Performance efficiency is a key factor in why Arm has become pervasive across the datacenter and in the cloud, powering server SoCs and DPUs, and MinIO’s latest optimizations will further drive Arm-based AI innovation for the world’s leading data enterprises.”

The latest generation of the NVIDIA® BlueField®-3 data processing unit (DPU) offers an integrated 16-core Arm-based CPU as part of the network controller card itself. This leads to simpler server designs by interfacing the Non-Volatile Memory Express (NVMe) drives directly to the networking card and bypassing any (main server) CPUs altogether. This represents an ideal pairing for software-defined technologies like MinIO. With 400Gb/s, Ethernet BlueField-3 DPUs offload, accelerate and isolate software-defined networking, storage, security and management functions. Given the criticality of disaggregating storage and compute in modern AI architectures, the pairing is remarkably powerful.

SOURCE: PRNewswire

Subscribe

- Never miss a story with notifications


    Latest stories

    spot_img