The patent badge is an abbreviated version of the USPTO patent document. The patent badge does contain a link to the full patent document.
The patent badge is an abbreviated version of the USPTO patent document. The patent badge covers the following: Patent number, Date patent was issued, Date patent was filed, Title of the patent, Applicant, Inventor, Assignee, Attorney firm, Primary examiner, Assistant examiner, CPCs, and Abstract. The patent badge does contain a link to the full patent document (in Adobe Acrobat format, aka pdf). To download or print any patent click here.
Patent No.:
Date of Patent:
Aug. 16, 2022
Filed:
Oct. 15, 2018
Intel Corporation, Santa Clara, CA (US);
Amrita Mathuriya, Portland, OR (US);
Sasikanth Manipatruni, Portland, OR (US);
Victor Lee, Santa Clara, CA (US);
Huseyin Sumbul, Portland, OR (US);
Gregory Chen, Portland, CA (US);
Raghavan Kumar, Hillsboro, OR (US);
Phil Knag, Hillsboro, OR (US);
Ram Krishnamurthy, Portland, OR (US);
Ian Young, Portland, OR (US);
Abhishek Sharma, Hillsboro, OR (US);
Intel Corporation, Santa Clara, CA (US);
Abstract
The present disclosure is directed to systems and methods of implementing a neural network using in-memory, bit-serial, mathematical operations performed by a pipelined SRAM architecture (bit-serial PISA) circuitry disposed in on-chip processor memory circuitry. The on-chip processor memory circuitry may include processor last level cache (LLC) circuitry. The bit-serial PISA circuitry is coupled to PISA memory circuitry via a relatively high-bandwidth connection to beneficially facilitate the storage and retrieval of layer weights by the bit-serial PISA circuitry during execution. Direct memory access (DMA) circuitry transfers the neural network model and input data from system memory to the bit-serial PISA memory and also transfers output data from the PISA memory circuitry to system memory circuitry. Thus, the systems and methods described herein beneficially leverage the on-chip processor memory circuitry to perform a relatively large number of vector/tensor calculations without burdening the processor circuitry.