dc.contributor.author |
Iqbal, Muhammad Jawad |
|
dc.date.accessioned |
2023-07-31T07:33:09Z |
|
dc.date.available |
2023-07-31T07:33:09Z |
|
dc.date.issued |
2021 |
|
dc.identifier.other |
00000204949 |
|
dc.identifier.uri |
http://10.250.8.41:8080/xmlui/handle/123456789/35294 |
|
dc.description |
Supervisor: Dr. Farhan Hussain |
en_US |
dc.description.abstract |
Deep neural networks have become the most advanced method for solving computer vision
problems. Although they are much Powerful, big amount of weights requires significant
storage space. So, these deep Neural networks have high computational cost and require more
memory resources which makes them hard to run it on systems with less hardware resources.
As of Today, the basic memory optimization technique after quantization is Huffman encoding.
However, the lossless encoding schemes based on Huffman encoding are significantly slower
than others. So in our proposed design in order to reduce memory requirement and make
decoding phase fast we use two stage process Quantization and a Proposed Encoding Scheme
which works together to decrease the memory requirement of Neural Network. Our method
first quantizes the layer weights and then encoding scheme is applied which further reduces the
storage requirement. In our work for testing purpose we use CIFAR10 and MNIST datasets.
However, in terms of compression ratio our encoding scheme is not as efficient as Huffman
encoding but in terms of decoding time it is way more effective than Huffman encoding. |
en_US |
dc.language.iso |
en |
en_US |
dc.publisher |
College of Electrical & Mechanical Engineering (CEME), NUST |
en_US |
dc.subject |
Key Words: Neural Networks, Huffman Encoding, Quantization, Embedded Systems |
en_US |
dc.title |
Memory Optimization of CNN for Reconfigurable Architectures |
en_US |
dc.type |
Thesis |
en_US |