The patent badge is an abbreviated version of the USPTO patent document. The patent badge does contain a link to the full patent document.

The patent badge is an abbreviated version of the USPTO patent document. The patent badge covers the following: Patent number, Date patent was issued, Date patent was filed, Title of the patent, Applicant, Inventor, Assignee, Attorney firm, Primary examiner, Assistant examiner, CPCs, and Abstract. The patent badge does contain a link to the full patent document (in Adobe Acrobat format, aka pdf). To download or print any patent click here.

Date of Patent:
Oct. 17, 2023

Filed:

Mar. 18, 2019
Applicant:

Microsoft Technology Licensing, Llc, Redmond, WA (US);

Inventors:

Kalin Ovtcharov, Snoqualmie, WA (US);

Eric S. Chung, Redmond, WA (US);

Vahideh Akhlaghi, Redmond, WA (US);

Ritchie Zhao, Ithaca, NY (US);

Assignee:
Attorneys:
Primary Examiner:
Int. Cl.
CPC ...
G06N 3/00 (2023.01); G06F 7/483 (2006.01); G06N 3/045 (2023.01); G06N 3/063 (2023.01);
U.S. Cl.
CPC ...
G06N 3/045 (2023.01); G06F 7/483 (2013.01); G06N 3/063 (2013.01);
Abstract

Quantization-aware neural architecture search ('QNAS') can be utilized to learn optimal hyperparameters for configuring an artificial neural network ('ANN') that quantizes activation values and/or weights. The hyperparameters can include model topology parameters, quantization parameters, and hardware architecture parameters. Model topology parameters specify the structure and connectivity of an ANN. Quantization parameters can define a quantization configuration for an ANN such as, for example, a bit width for a mantissa for storing activation values or weights generated by the layers of an ANN. The activation values and weights can be represented using a quantized-precision floating-point format, such as a block floating-point format (“BFP”) having a mantissa that has fewer bits than a mantissa in a normal-precision floating-point representation and a shared exponent.


Find Patent Forward Citations

Loading…