hh.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
DNNARA: A Deep Neural Network Accelerator using Residue Arithmetic and Integrated Photonics
George Washington University, Washington, DC, United States.
Halmstad University, School of Information Technology, Halmstad Embedded and Intelligent Systems Research (EIS), Centre for Research on Embedded Systems (CERES).ORCID iD: 0000-0001-8806-8146
George Washington University, Washington, DC, United States.
George Washington University, Washington, DC, United States.
Show others and affiliations
2020 (English)In: Proceedings of the 49th International Conference on Parallel Processing, New York: Association for Computing Machinery (ACM), 2020, p. 1-11, article id 3404467Conference paper, Published paper (Refereed)
Abstract [en]

Deep Neural Networks (DNNs) are currently used in many fields, including critical real-time applications. Due to its compute-intensive nature, speeding up DNNs has become an important topic in current research. We propose a hybrid opto-electronic computing architecture targeting the acceleration of DNNs based on the residue number system (RNS). In this novel architecture, we combine the use of Wavelength Division Multiplexing (WDM) and RNS for efficient execution. WDM is used to enable a high level of parallelism while reducing the number of optical components needed to decrease the area of the accelerator. Moreover, RNS is used to generate optical components with short optical critical paths. In addition to speed, this has the advantage of lowering the optical losses and reducing the need for high laser power. Our RNS compute modules use one-hot encoding and thus enable fast switching between the electrical and optical domains. 

In this work, we demonstrate how to implement the different DNN computational kernels using WDM-enabled RNS based integrated photonics. We provide an accelerator architecture that uses our designed components and perform design space exploration to select efficient architecture parameters. Compared to memristor crossbars, our residue matrix-vector multiplication unit has two orders of magnitude higher peak performance. Our experimental evaluation using DNN benchmarks illustrates that our architecture can perform more than 19 times faster than the state of the art GPUs under the same power budget. © 2020 ACM.

Place, publisher, year, edition, pages
New York: Association for Computing Machinery (ACM), 2020. p. 1-11, article id 3404467
Series
ACM International Conference Proceeding Series
Keywords [en]
deep learning, neural network accelerators, optical computing, residue number system
National Category
Telecommunications
Identifiers
URN: urn:nbn:se:hh:diva-43447DOI: 10.1145/3404397.3404467Scopus ID: 2-s2.0-85090555072ISBN: 978-1-4503-8816-0 (electronic)OAI: oai:DiVA.org:hh-43447DiVA, id: diva2:1501563
Conference
49th International Conference on Parallel Processing (ICPP 2020), Virtual/Online, Canada, 17-20 August, 2020
Note

Funding text: This project is supported by Air Force Office of Scientific Research (AFOSR) award number FA9550-19-1-0277

Available from: 2020-11-17 Created: 2020-11-17 Last updated: 2020-11-24Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Alkabani, Yousra

Search in DiVA

By author/editor
Alkabani, Yousra
By organisation
Centre for Research on Embedded Systems (CERES)
Telecommunications

Search outside of DiVA

GoogleGoogle Scholar

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 170 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf