Sunday, July 14, 2024

EdgeCortix Launches SAKURA-II Platform to Power the Next Wave of Generative AI at the Edge

Related stories

Synchron Announces Brain Computer Interface Chat Feature Powered by OpenAI

New feature includes AI-driven emotion and language predictions for...

Inspiro Wins Multiple Gold Honors from Globee® Awards

Inspiro, a leading global CX outsourcing company, is excited...

Peak Boosts Business Productivity with General Release of Agentic AI Assistant, Co:Driver

Artificial intelligence company Peak announced the general availability of Co:Driver,...

The next-generation high performance, energy efficient Edge AI accelerator addresses the latest Generative AI solutions at the edge from vision to billions of parameters large language models

EdgeCortix® Inc., a leading fabless semiconductor company specializing in energy-efficient AI processing at the edge, unveiled its next-generation SAKURA-II Edge AI accelerator.

This state-of-the-art platform, paired with EdgeCortix’s innovative second generation Dynamic Neural Accelerator (DNA) architecture, is engineered to tackle the most challenging Generative AI tasks in the industry. Designed for flexibility and power efficiency, SAKURA-II empowers users to seamlessly manage a wide range of complex tasks including Large Language Models (LLMs), Large Vision Models (LVMs), and multi-modal transformer-based applications, even within the stringent environmental constraints at the edge. Featuring low latency, best-in-class memory bandwidth, high accuracy, and compact form factors, SAKURA-II delivers unparalleled performance and cost-efficiency across the diverse spectrum of edge AI applications.

Well-suited for numerous use cases across the manufacturing, industry 4.0, security, robotics, aerospace, and telecommunications industries, SAKURA-II features EdgeCortix’s latest generation runtime reconfigurable neural processing engine, DNA-II. Leveraging this highly configurable intellectual property block, SAKURA-II delivers power efficiency and real-time processing capabilities while simultaneously executing multiple deep neural network models with low latency. SAKURA-II can deliver up to 60 trillion operations per second (TOPS) of effective 8-bit integer performance and 30 trillion 16-bit brain floating-point operations per second (TFLOPS), while also supporting built-in mixed precision for handling the rigorous demands of next-generation AI tasks.

The SAKURA-II platform, with its sophisticated MERA software suite, features a heterogeneous compiler platform, advanced quantization, and model calibration capabilities. This software suite includes native support for leading development frameworks such as PyTorch, TensorFlow Lite, and ONNX. MERA’s flexible host-to-accelerator unified runtime is adept at scaling across single, multi-chip, and multi-card systems at the edge, significantly streamlining AI inferencing and shortening deployment times for data scientists. Furthermore, the integration with the MERA Model Library, with seamless interface to Hugging Face Optimum, offers users access to an extensive range of the latest transformer models, ensuring a smooth transition from training to edge inference.

Also Read: NVIDIA Grace Hopper Ignites New Era of AI Supercomputing

“SAKURA-II’s impressive 60 TOPS performance within 8W of typical power consumption, combined with its mixed-precision and built-in memory compression capabilities, positions it as a pivotal technology for the latest Generative AI solutions at the edge,” said Sakyasingha Dasgupta, CEO and Founder of EdgeCortix. “Whether running traditional AI models or the latest Llama 2/3, Stable-diffusion, Whisper or Vision-transformer models, SAKURA-II provides deployment flexibility at superior performance per watt and cost-efficiency. We are committed to ensuring we meet our customer’s varied needs and also to securing a technological foundation that remains robust and adaptable within the swiftly evolving AI sector.”

Key Benefits of SAKURA-II include:

  • Optimized for Generative AI: Tailored specifically for processing Generative AI workloads at the edge with minimal power consumption.
  • Complex Model Handling: Capable of managing multi-billion parameter models like Llama 2, Stable Diffusion, DETR, and ViT within a typical power envelope of 8W.
  • Seamless Software Integration: Fully compatible with EdgeCortix’s MERA software suite, facilitating seamless transitions from model training to deployment.
  • Enhanced Memory Bandwidth: Offers up to four times more DRAM bandwidth than competing AI accelerators, ensuring superior performance for LLM and LVM.
  • Real-Time Data Streaming: Optimized for low-latency operations under real-time data streaming conditions.
  • Advanced Precision: Provides software-enabled mixed-precision support for near FP32 accuracy.
  • Sparse Computation: Supports sparse computation to reduce memory footprint and optimize bandwidth.
  • Versatile Functionality: Supports arbitrary activation functions with hardware approximation for enhanced adaptability.
  • Efficient Data Handling: Includes a dedicated Reshaper engine to manage complex data permutations on-chip and minimize host CPU load.
  • Power Management: Features on-chip power-gating and power management capabilities to facilitate ultra-high efficiency modes.

Source: BusinessWire


- Never miss a story with notifications

    Latest stories