System-Technology Co-optimization for enablement of MRAM-based Machine Learning

Host institution

Imec is a world-leading research and innovation hub in nanoelectronics and digital technologies. The machine learning program at Imec is leading the quest for computationally- and energy-efficient machine learning accelerators. Imec‘s machine learning research is driving the co-evolution of hardware and algorithms needed to facilitate the move to this new computational paradigm.

Supervisor

Dr. Dwaipayan Biswas

Description

ML techniques such as DNNs have realized important breakthroughs in a myriad of application domains. The core operations in DNN are MVMs and in majority of use-cases, incur a dedicated training step for modelling, resulting in generating a set of parameters which are used in an inference step for classification/prediction outcomes. Training of DNNs have been traditionally carried out using software compute capabilities while a considerable research effort has been spent by the community to accelerate the inference on-chip for (near) real-time outcomes optimizing energy and accuracy. There is a need to look at optimized training procedures for reducing the energy footprint with minimal accuracy trade-off. Minimizing the data movement between the compute and memory blocks (non-Von Neuman trajectory) has had great success towards energy optimization, especially targeting the accelerated-inference landscape for ML applications. This has primarily been achieved through compute near/in memory (CnM/CiM) techniques. Devices based on standard technology as well as novel/emerging technology have been the main contributors to CiM/CnM paradigm, helping to optimize the core MVM operation. Both digital multiply-accumulate circuits and Kirchoff’s law-based analogue domain processing have been explored to avoid costly memory fetches to an external memory. 

Dense non-volatile memories (NVM), with large resistance (MOhm) and narrow parameters distribution are a promising candidate, however typical write penalties for the standard STT variant of MRAM technology could be a bottleneck for their adoption. This is mitigated by making use of MRAM emerging writing concepts: SOT and VCMA. In addition, design solutions are proposed to create multi-level bit MTJs cells and are currently being prototyped for further demonstration. This project will explore Design-Technology Co-optimization (DTCO) using in-house SOT/VGSOT-MRAM technology-based ML hardware for optimizing ML Training related system performance for a dedicated application space. This will help to close the bottom-up loop connecting device characteristics to system power/performance metrics, enabling system technology co-optimization (STCO) for CnM-centric ML applications. 

 In this PhD you will: 

  1. Understand device characteristics for binary and multi-level bit SOT-MTJ.  
  2. DTCO: architecture-level choices which help to optimize the device knobs for yielding low-energy hardware solution for ML 
  3. STCO: Exploration of novel compute near/in memory concepts using MRAM for system PPA impact estimation.   

The PhD is expected to develop a: i) device-level understanding enabling ML circuit level DTCO optimization, and ii) architecture-level choices in conjunction with MRAM technology, which help to optimize system PPA for ML training/inference. You are expected to participate in both circuit and architecture level optimization loops, enabling device benchmarking. 

 
[1] S. Cosemans et al., “Towards 10000TOPS/W DNN Inference with Analog in-Memory Computing – A Circuit Blueprint, Device Options and Requirements,” 2019 IEEE International Electron Devices Meeting (IEDM), 2019.  

[2] J. Doevenspeck et al., “SOT-MRAM Based Analog in-Memory Computing for DNN Inference,” 2020 IEEE Symposium on VLSI Technology, 2020. 

[3] G. Karunaratne et al. Robust high-dimensional memory-augmented neural networks. Nat Commun (2021). 

Requirements

  • Master’s in Computer Engineering, Electrical Engineering
  • Fundamental knowledge in neural networks
  • Computer architecture, and circuit design skills
  • Owing to the international nature of the research center, good verbal and written communication skills in English is a must.

Although not mandatory for this position, following prior experience could be beneficial:

  • Electrical device characterization
  • Knowledge within the field of magnetism, spintronics, magnetic tunnel junctions

Planned Secondments

ETH Zürich (Zurich, Switzerland), under the supervision of Pietro Gambardella.

NanOsc (Gothenburg, Sweden), under the supervision of Fredrik Magnusson.

Registering University

KU Leuven (Leuven, Belgium).


Diseño y desarrollo web Triplevdoble