Theme 2 Objectives:
- Pipeline combining computation and dimensionality reduction to dramatically reduce the data flow off the sensor
- Adaptable feature extraction, integrated with Thrust 3
- Multi-layer: analog feature extraction (AFE) followed by digital feature processing (DFP)
- Multi-modal: features are fused at the digital output
- Fundamentals: what are the theoretical limits of a given feature extraction architecture?
Theme 2 Highlights
Task 2.1: Quantization Compensation in Feature Extraction
Task 2.2: Analog CIM Chip Design for Broadband Digital Beamforming
We’ve successfully taped out BeamCIM, a cutting-edge mixed-signal Compute-In-Memory (CIM) accelerator optimized for beamforming applications. BeamCIM leverages Linear Embedding (LE) techniques to efficiently transform high-dimensional input data into compact, lower-dimensional features, enabling fast and power-efficient beamforming performance. Designed for next-generation signal processing tasks, BeamCIM pushes the boundaries of speed, scalability, and system integration.
Continued Task 2.2: Hybrid Slepian Beamformer Overview
- ADC reduction without beam squinting errors
- MAC and ADC co-design improves power and area efficiency
- Simplified multi-phase clock-generation for scalability
- Fabricated in 28-nm CMOS, 8-element, 2-beam receiver with 7.8 mW and 0.02 mm2 / beam
Task 2.3: Adaptive Multi-Modal Sensor Fusion
Objectives: Deep learning driven, adaptive lidar/radar/video sensor fusion with data and/or complexity reduction.
Current Accomplishments:
Task 2.4: A 22nm 9.51 TOPS/W HTNN Processor with 2MB MRAM
We developed the first structured-sparse Walsh-Hadamard NN processor with on-chip MRAM, enabling compact, energy-efficient radar inference. Demonstrated on a Cognisense radar workload, the chip integrates a specialized architecture and memory system optimized for sparse-orthogonal computations.
To address MRAM’s leakage and read energy, we use layer-wise power gating and a multi-clocked weight cache, achieving 9.51 TOPS/W at 33% weight density. When applied to a ChirpNet-style radar model, our transform layer delivers 2.4× better performance and 2.1× energy efficiency over recent CNN baselines—with no accuracy loss.