Optimized Weight Programming for Analogue Memory-based Deep Neural Networks

التفاصيل البيبلوغرافية
العنوان: Optimized Weight Programming for Analogue Memory-based Deep Neural Networks
المؤلفون: Abu Sebastian, Stefano Ambrogio, Malte J. Rasch, An Chen, Nandakumar Rajaleksh, Andrea Fasoli, Jonathan Timcheck, Charles Mackin, Jose Luquin, Pritish Narayanan, Hsinyu Tsai, Robert L. Bruce, Alexander Friz, Geoffrey W. Burr, Manuel Le Gallo
بيانات النشر: Research Square Platform LLC, 2021.
سنة النشر: 2021
مصطلحات موضوعية: business.industry, Computer science, Deep neural networks, Artificial intelligence, business
الوصف: Analogue memory-based Deep Neural Networks (DNNs) provide energy-efficiency and per-area throughput gains relative to state-of-the-art digital counterparts such as graphic processing units (GPUs). Recent advances focus largely on hardware-aware algorithmic training and improvements in circuits, architectures, and memory device characteristics. Optimal translation of software-trained weights into analogue hardware weights---given the plethora of complex memory non-idealities---represents an equally important goal in realizing the full potential of this technology. We report a generalized computational framework that automates the process of crafting complex weight programming strategies for analogue memory-based DNNs, in order to minimize accuracy degradations during inference, particularly over time. This framework is agnostic to DNN structure and is shown to generalize well across Long Short-Term Memory (LSTM), Convolution Neural Networks (CNNs), and Transformer networks. Being a highly-flexible numerical heuristic, our approach can accommodate arbitrary device-level complexity, and is thus broadly applicable to a variety of analogue memories and their continually evolving device characteristics. Interestingly, this computational technique is capable of optimizing inference accuracy without the need to run inference simulations or evaluate large training, validation, or test datasets. Lastly, by quantifying the limit of achievable inference accuracy given imperfections in analogue memory, weight programming optimization represents a unique and foundational tool for enabling analogue memory-based DNN accelerators to reach their full inference potential.
الوصول الحر: https://explore.openaire.eu/search/publication?articleId=doi_________::3cc478e8fcea6d6dfc345071fdfad6b8Test
https://doi.org/10.21203/rs.3.rs-1028668/v1Test
حقوق: OPEN
رقم الانضمام: edsair.doi...........3cc478e8fcea6d6dfc345071fdfad6b8
قاعدة البيانات: OpenAIRE