scholarly journals Toward Software-Equivalent Accuracy on Transformer-Based Deep Neural Networks With Analog Memory Devices

2021 ◽  
Vol 15 ◽  
Author(s):  
Katie Spoon ◽  
Hsinyu Tsai ◽  
An Chen ◽  
Malte J. Rasch ◽  
Stefano Ambrogio ◽  
...  

Recent advances in deep learning have been driven by ever-increasing model sizes, with networks growing to millions or even billions of parameters. Such enormous models call for fast and energy-efficient hardware accelerators. We study the potential of Analog AI accelerators based on Non-Volatile Memory, in particular Phase Change Memory (PCM), for software-equivalent accurate inference of natural language processing applications. We demonstrate a path to software-equivalent accuracy for the GLUE benchmark on BERT (Bidirectional Encoder Representations from Transformers), by combining noise-aware training to combat inherent PCM drift and noise sources, together with reduced-precision digital attention-block computation down to INT6.

2021 ◽  
Vol 11 (7) ◽  
pp. 3184
Author(s):  
Ismael Garrido-Muñoz  ◽  
Arturo Montejo-Ráez  ◽  
Fernando Martínez-Santiago  ◽  
L. Alfonso Ureña-López 

Deep neural networks are hegemonic approaches to many machine learning areas, including natural language processing (NLP). Thanks to the availability of large corpora collections and the capability of deep architectures to shape internal language mechanisms in self-supervised learning processes (also known as “pre-training”), versatile and performing models are released continuously for every new network design. These networks, somehow, learn a probability distribution of words and relations across the training collection used, inheriting the potential flaws, inconsistencies and biases contained in such a collection. As pre-trained models have been found to be very useful approaches to transfer learning, dealing with bias has become a relevant issue in this new scenario. We introduce bias in a formal way and explore how it has been treated in several networks, in terms of detection and correction. In addition, available resources are identified and a strategy to deal with bias in deep NLP is proposed.


Author(s):  
Shubham Jain ◽  
Swagath Venkataramani ◽  
Vijayalakshmi Srinivasan ◽  
Jungwook Choi ◽  
Pierce Chuang ◽  
...  

2020 ◽  
Vol 49 (4) ◽  
pp. 482-494
Author(s):  
Jurgita Kapočiūtė-Dzikienė ◽  
Senait Gebremichael Tesfagergish

Deep Neural Networks (DNNs) have proven to be especially successful in the area of Natural Language Processing (NLP) and Part-Of-Speech (POS) tagging—which is the process of mapping words to their corresponding POS labels depending on the context. Despite recent development of language technologies, low-resourced languages (such as an East African Tigrinya language), have received too little attention. We investigate the effectiveness of Deep Learning (DL) solutions for the low-resourced Tigrinya language of the Northern-Ethiopic branch. We have selected Tigrinya as the testbed example and have tested state-of-the-art DL approaches seeking to build the most accurate POS tagger. We have evaluated DNN classifiers (Feed Forward Neural Network – FFNN, Long Short-Term Memory method – LSTM, Bidirectional LSTM, and Convolutional Neural Network – CNN) on a top of neural word2vec word embeddings with a small training corpus known as Nagaoka Tigrinya Corpus. To determine the best DNN classifier type, its architecture and hyper-parameter set both manual and automatic hyper-parameter tuning has been performed. BiLSTM method was proved to be the most suitable for our solving task: it achieved the highest accuracy equal to 92% that is 65% above the random baseline.


2022 ◽  
pp. 25-52
Author(s):  
Abhinav Goel ◽  
Caleb Tung ◽  
Xiao Hu ◽  
Haobo Wang ◽  
Yung-Hsiang Lu ◽  
...  

Author(s):  
Shihui Yin ◽  
Zhewei Jiang ◽  
Minkyu Kim ◽  
Tushar Gupta ◽  
Mingoo Seok ◽  
...  

Author(s):  
Stefano Ambrogio ◽  
Pritish Narayanan ◽  
Hsinyu Tsai ◽  
Charles Mackin ◽  
Katherine Spoon ◽  
...  

2018 ◽  
Vol 14 (4) ◽  
pp. 520-534 ◽  
Author(s):  
Muhammad Abdullah Hanif ◽  
Alberto Marchisio ◽  
Tabasher Arif ◽  
Rehan Hafiz ◽  
Semeen Rehman ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document