-

Untether AI Dramatically Expands AI Model Support and Speeds Developer Velocity with New Generative Compiler Technology

Generative compilation supports four times more models, reduces implementation time by orders of magnitude

TORONTO--(BUSINESS WIRE)--Untether AI®, a leader in energy-centric AI inference acceleration today introduced a breakthrough in AI model support and developer velocity for users of the imAIgine® Software Development Kit (SDK). Using a breakthrough generative compiler technology, the upcoming release the imAIgine SDK will support 4 times more AI models than the previous releases. Additionally, for new neural networks users may architect, the generative compiler creates new kernels for these layers automatically, reducing development time to just minutes, increasing developer velocity by orders of magnitude.

Generative compiler technology speeds development for true push-button deployment

Kernel-based compilers require hand-coded c/c++ programs for each layer in a neural network. With potentially thousands of possible kernels required to support the exploding number of neural networks, these kernels become a bottleneck in lowering neural networks to hardware implementations. To resolve this bottleneck, Untether AI has developed a generative compiler technology, which ingests neural networks and applies a series of specialized compiler transformations that lowers high-level neural network operators into fundamental computational primitives. These primitives are then recombined into a hardware-optimized representation, maximizing throughput, minimizing latency, and boosting efficiency on Untether AI’s At-Memory Compute architecture. With the push of a button, neural networks can be lowered to Untether AI’s speedAI® devices and cards in a matter of seconds.

Out of the box, this new technology increases the number of models supported by the imAIgine SDK to over 300, spanning neural networks performing object detection, semantic segmentation, classification, error detection, and many other functions. This breadth of support builds on Untether AI’s validated performance and energy efficiency advantages shown in the last MLPerf inference submissions.

For new neural networks or modified layers of neural networks, the generative compiler can create new kernels on-the-fly, optimized for the over 1,400 RISC-V processors and At-Memory Compute processing elements on speedAI-based inference acceleration solutions.

Kernel-based optimizations available for utmost performance

The new compiler retains the ability to use hand-crafted, optimized kernels. The compiler has the option to use existing kernels, but if a kernel doesn’t exist in the library, the generative technology is used to construct new kernels on-the-fly. This ability to mix and match pre-existing, optimized kernels with generated kernels speeds the time to implement highly performant neural networks.

“This new compiler technology provides the ultimate in productivity and flexibility,” said Alex Grbic, PhD and VP of Software Engineering at Untether AI. “By introducing our generative compiler technology, customers gain an accelerated path to market, quickly achieving a deployment of their neural networks on Untether AI’s architecture.”

This new generative compiler technology will be in the imAIgine SDK version 25.04 release, scheduled for availability in early April of 2025. To obtain early access to the SDK, please visit https://www.untether.ai/imaigine-sdk-early-access-program/

About Untether AI

Untether AI® provides energy-centric AI inference acceleration from the edge to the cloud, supporting any type of neural network model. With its At-Memory compute architecture, Untether AI has solved the data movement bottleneck that costs energy and performance in traditional CPUs and GPUs, resulting in high-performance, low-latency neural network inference acceleration without sacrificing accuracy. Untether AI’s technology ships today as speedAI® devices, acceleration cards, and its imAIgine® Software Development Kit. More information can be found at www.untether.ai.

All references to Untether AI trademarks are the property of Untether AI. All other trademarks mentioned herein are the property of their respective owners.

Contacts

Media Contact for Untether AI:
Michelle Clancy Fuller, Cayenne Global
Michelle.clancy@cayennecom.com
1-503-702-4732

Company Contact:
Robert Beachler, Untether AI
beach@untether.ai
+1.650.793.8219

Untether AI


Release Versions

Contacts

Media Contact for Untether AI:
Michelle Clancy Fuller, Cayenne Global
Michelle.clancy@cayennecom.com
1-503-702-4732

Company Contact:
Robert Beachler, Untether AI
beach@untether.ai
+1.650.793.8219

More News From Untether AI

Untether AI with AI Platform Alliance Unveils AI-Powered Intelligent Video Solution at ISC West 2025

TORONTO--(BUSINESS WIRE)--Untether AI®, a leader in energy-centric AI inference acceleration, is teaming up with AI Platform Alliance members Ampere® Computing, NETINT, ZoneMinder, AVC Group, and ASA Computers to introduce the Intelligent Video Recording (IVR) solution at ISC West 2025, the premier security industry tradeshow. This groundbreaking AI-powered system sets a new benchmark in video surveillance, delivering up to 8x better AI camera efficiency while operating in an eco-friendly serve...

Untether AI and Vertical Data Join Forces to Revolutionize AI-Centric Modular Data Centers

TORONTO--(BUSINESS WIRE)--Untether AI®, a leader in energy-efficient AI inference acceleration, today announced its partnership with Vertical Data, recognized for its cutting-edge infrastructure solutions that enhance computing capabilities at the source of data generation. This collaboration aims to advance modular and portable data center solutions, enabling faster, more secure, and highly efficient AI-driven computing. Untether AI’s industry-leading AI inference accelerators, including the h...

Untether AI Ships speedAI 240 Slim: World’s Fastest, Most Energy Efficient AI Inference Accelerator for Cloud to Edge Applications

TORONTO--(BUSINESS WIRE)--Untether AI®, the leader in energy-centric AI inference acceleration, today announced broad availability of its highly anticipated speedAI 240 Slim AI inference accelerator cards. Recently receiving top marks in the MLPerf benchmark for AI inference, speedAI 240 Slim cards provide customers the performance, energy efficiency, AI model support, and scalability they need for a broad range of applications from regional clouds to the edge. J-squared and Ola-Krutrim are amo...
Back to Newsroom
  1. There was an issue with the authorization server. Please contact support if the issue persists.