Search In this Thesis
   Search In this Thesis  
العنوان
Design and implementation of hardware accelerator of deep neural networks /
المؤلف
Mohamed,Khaled Salah Eldin Hassan Gamal Eldin
هيئة الاعداد
باحث / د صلاح الدين حسن جمال الدين محمد
مشرف / أشرف الفرغلي سالم
مناقش / أحمد حسن كامل مدين
مناقش / محمد واثق علي كامل الخراشي
تاريخ النشر
2023
عدد الصفحات
72P.:
اللغة
الإنجليزية
الدرجة
ماجستير
التخصص
هندسة النظم والتحكم
تاريخ الإجازة
1/1/2023
مكان الإجازة
جامعة عين شمس - كلية الهندسة - كهرباء حاسبات
الفهرس
Only 14 pages are availabe for public view

from 115

from 115

Abstract

Deep neural networks have been one of the hottest research topics due to their wide applications and importance in industry. They are used in many fields and in different scopes such as speech recognition, object detection and autonomous vehicles. However, they call for huge computational burden to process the input data and do the inevitable heavy calculations in the form of matrix multiplications and accumulations.
Recurrent neural networks (RNNs) are considered to be among the most important types of neural networks especially for the applications where processing of a sequence of data comes to place. RNNs are in general computationally expensive and need a lot of processing time and power. Therefore, there is a strong need to reduce the processing time in order to use them in an embedded environment with limited resources.
In this thesis we present an accelerated field-programmable gate array (FPGA) design for RNNs with an emphasis on long short term memory neural networks (LSTMs). A new configurable block capable of calculating Tanh and Sigmoid activation functions is proposed and analysed. The solution is based on a look-up table and additional simple math operations, which leads to a speedup of the proposed model of the neural network.
.