Our Technology

Aspinity’s AnalogML™ is the most efficient NPU core for machine learning at the edge

AnalogML™ delivers state-of-the-art acceleration for machine learning in size and power constrained devices

  • Best compute density: 0.86 TOPS/mm2
  • Best compute efficiency: 300 TOPS/W
  • Scalable compute fabric accommodates model needs for many applications (256 to 250+k parameters)
  • Parameters embedded in compute elements to eliminate bottleneck
  • Supports CNN, RNN, and other common layer types.
Digital Accelerator Analog In-Memory-Computing (IMC) Aspinity AnalogML™
Data Movement HighContinual movement of instructions, parameters, and intermediate data to/from cache MediumParameters are embedded; still significant movement of instructions and intermediate data to/from cache LowParameters and configuration (i.e., instructions) are embedded; Signals stream through
Parameter Precision HighHigh-precision datatypes often supported Low-MediumLimited to ≤ 8 bits precision, much less in many implementations High10+ bit precision
MAC Precision HighHigh-precision datatypes often supported Low“Memory as compute” precision limited by linearity, noise, and mismatch HighParameters co-located with compute allows linear multiplication circuits to be used
MAC Efficiency Low1000’s of transistors and transfer instructions and operands from memory Medium Efficient multiplication in analog; still fetch input vector from memory HighEfficient multiplication in analog
NN Efficiency LowData movement to/from memory MediumMultiply is in memory but activation requires digital HighMAC & activation all within analog circuitry
Robustness to Temperature & Manufacturing Variation HighLimited issues with digital circuitry LowChallenged with long signal chains HighSilicon-proven and dynamic approach to on-the-fly trimming

Precision Analog Parameters

High Performance Acceleration

Aspinity patented analog memory is co-located in circuits
10+ bits precision for storing parameters (weights, etc.)
Permanent and accurate parameter storage (no memory fetch)

Weight quantization not required to save power

Scalable NPU Architecture

Aspinity scalable NPU architecture based on configurable analog blocks
  • Tiled CAB (Configurable Analog Block) architecture can be sized for model needs
  • CABs are software programmable to support a wide range of models and applications
  • All-analog MAC and activation functions in each CAB (no digital conversions required)
  • No interlayer data buffering
  • High-precision analog parameters (10+ bit) located at compute circuit eliminates time and energy spent fetching datas

AnalogML™ enables robust and accurate immunity to analog variability

  • Dynamic software-driven solution for on-the-fly variation trimming
  • Leverages high-precision analog parameters for localized fine tuning/trimming
  • Not susceptible to environmental or manufacturing variations
  • Repeatable and consistent analog computing platform can be scaled to larger neural networks and smaller process nodes to meet the needs of next generation AI computing

Modern, easy-to-use Python based training and execution

Integration with common
ML frameworks

Simple and quick compilation
to AnalogML hardware

Supports CNN, RNN, and other common layer types