The patent badge is an abbreviated version of the USPTO patent document. The patent badge does contain a link to the full patent document.
The patent badge is an abbreviated version of the USPTO patent document. The patent badge covers the following: Patent number, Date patent was issued, Date patent was filed, Title of the patent, Applicant, Inventor, Assignee, Attorney firm, Primary examiner, Assistant examiner, CPCs, and Abstract. The patent badge does contain a link to the full patent document (in Adobe Acrobat format, aka pdf). To download or print any patent click here.
Patent No.:
Date of Patent:
Feb. 05, 2008
Filed:
Oct. 02, 2003
Xinmin Tian, Union City, CA (US);
Shih-wei Liao, San Jose, CA (US);
Hong Wang, Fremont, CA (US);
Milind Girkar, Sunnyvale, CA (US);
John Shen, San Jose, CA (US);
Perry Wang, San Jose, CA (US);
Grant Haab, Mahomet, IL (US);
Gerolf Hoflehner, Santa Clara, CA (US);
Daniel Lavery, Santa Clara, CA (US);
Hideki Saito, Sunnyvale, CA (US);
Sanjiv Shah, Champaign, IL (US);
Dongkeun Kim, San Jose, CA (US);
Xinmin Tian, Union City, CA (US);
Shih-wei Liao, San Jose, CA (US);
Hong Wang, Fremont, CA (US);
Milind Girkar, Sunnyvale, CA (US);
John Shen, San Jose, CA (US);
Perry Wang, San Jose, CA (US);
Grant Haab, Mahomet, IL (US);
Gerolf Hoflehner, Santa Clara, CA (US);
Daniel Lavery, Santa Clara, CA (US);
Hideki Saito, Sunnyvale, CA (US);
Sanjiv Shah, Champaign, IL (US);
Dongkeun Kim, San Jose, CA (US);
Intel Corporation, Santa Clara, CA (US);
Abstract
Methods and apparatus for reducing memory latency in a software application are disclosed. A disclosed system uses one or more helper threads to prefetch variables for a main thread to reduce performance bottlenecks due to memory latency and/or a cache miss. A performance analysis tool is used to profile the software application's resource usage and identifies areas in the software application experiencing performance bottlenecks. Compiler-runtime instructions are generated into the software application to create and manage the helper thread. The helper thread prefetches data in the identified areas of the software application experiencing performance bottlenecks. A counting mechanism is inserted into the helper thread and a counting mechanism is inserted into the main thread to coordinate the execution of the helper thread with the main thread and to help ensure the prefetched data is not removed from the cache before the main thread is able to take advantage of the prefetched data.