Patents by Inventor Zehra Sura
Zehra Sura has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 11568235Abstract: Embodiments for implementing mixed precision learning for neural networks by a processor. A neural network may be replicated into a plurality of replicated instances and each of the plurality of replicated instances differ in precision used for representing and determining parameters of the neural network. Data instances may be routed to one or more of the plurality of replicated instances for processing according to a data pre-processing operation.Type: GrantFiled: November 19, 2018Date of Patent: January 31, 2023Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Zehra Sura, Parijat Dube, Bishwaranjan Bhattacharjee, Tong Chen
-
Patent number: 11403213Abstract: A method for transparently moving a block of memory with respect to an application using the block of memory, includes inserting, by a compiler, in an application that includes a memory allocation call, instructions for transparently moving a block of memory with respect to an application using the block of memory. The instructions include obtaining a first pointer returned by a memory allocator, where the first pointer points to an internal data structure, the internal data structure includes a read-write lock and a second pointer, and the second pointer points to an actual memory block. The instructions further include acquiring a read lock on a read-write lock in the internal data structure, before the first pointer is used by the application, obtaining the second pointer to the actual memory block, and dereferencing the second pointer to access the actual memory block for the application data.Type: GrantFiled: June 28, 2019Date of Patent: August 2, 2022Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Wenqi Cao, Arun Iyengar, Gong Su, Zehra Sura, Qi Zhang
-
Publication number: 20200409833Abstract: A method for transparently moving a block of memory with respect to an application using the block of memory, includes inserting, by a compiler, in an application that includes a memory allocation call, instructions for transparently moving a block of memory with respect to an application using the block of memory. The instructions include obtaining a first pointer returned by a memory allocator, where the first pointer points to an internal data structure, the internal data structure includes a read-write lock and a second pointer, and the second pointer points to an actual memory block. The instructions further include acquiring a read lock on a read-write lock in the internal data structure, before the first pointer is used by the application, obtaining the second pointer to the actual memory block, and dereferencing the second pointer to access the actual memory block for the application data.Type: ApplicationFiled: June 28, 2019Publication date: December 31, 2020Inventors: WENQI CAO, ARUN IYENGAR, GONG SU, ZEHRA SURA, QI ZHANG
-
Partial synchronization between compute tasks based on threshold specification in a computing system
Patent number: 10824481Abstract: Embodiments for implementing partial synchronization between compute processes based on threshold specification in a computing environment. One or more compute processes may be synchronized in one of a plurality of types of computing platforms using a barrier having a barrier release condition based on a threshold of one or more measures. The barrier is defined according to one or more parameters. The one or more compute processes may be released via the barrier upon exceeding the threshold of the one or more measures.Type: GrantFiled: November 13, 2018Date of Patent: November 3, 2020Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Zehra Sura, Li Zhang, Ashish Kundu, Ravi Nair -
Publication number: 20200160169Abstract: Embodiments for implementing mixed precision learning for neural networks by a processor. A neural network may be replicated into a plurality of replicated instances and each of the plurality of replicated instances differ in precision used for representing and determining parameters of the neural network. Data instances may be routed to one or more of the plurality of replicated instances for processing according to a data pre-processing operation.Type: ApplicationFiled: November 19, 2018Publication date: May 21, 2020Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Zehra SURA, Parijat DUBE, Bishwaranjan BHATTACHARJEE, Tong CHEN
-
PARTIAL SYNCHRONIZATION BETWEEN COMPUTE TASKS BASED ON THRESHOLD SPECIFICATION IN A COMPUTING SYSTEM
Publication number: 20200151028Abstract: Embodiments for implementing partial synchronization between compute processes based on threshold specification in a computing environment. One or more compute processes may be synchronized in one of a plurality of types of computing platforms using a barrier having a barrier release condition based on a threshold of one or more measures. The barrier is defined according to one or more parameters. The one or more compute processes may be released via the barrier upon exceeding the threshold of the one or more measures.Type: ApplicationFiled: November 13, 2018Publication date: May 14, 2020Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Zehra SURA, Li ZHANG, Ashish KUNDU, Ravi NAIR -
Patent number: 8930921Abstract: According to one embodiment of the present invention, a computer system is provided where the computer system includes a main processor, first and second active memory device. The computer system is configured to perform a method including receiving an executable module generated by a compiler, wherein the executable module includes a code section identified as executable by a first processing element in the first active memory device and a second processing element in the second active memory device. The method includes copying the code section to memory in the first device based on the code section being executable on the first device, copying the code section from the first active memory device to an instruction buffer of the first processing element and copying the code section from the first device to the second device based on the code section being executable on the second device.Type: GrantFiled: November 20, 2012Date of Patent: January 6, 2015Assignee: International Business Machines CorporationInventors: Tong Chen, John K. O'Brien, Zehra Sura
-
Patent number: 8914778Abstract: According to one embodiment, a method for a compiler to produce an executable module to be executed by a computer system including a main processor and active memory devices includes dividing source code into code sections, identifying a first code section to be executed by the active memory devices, wherein the first code section is one of the code sections and identifying data structures that are used by the first code section. The method also includes classifying the data structures based on pre-defined attributes, formulating, by the compiler, a storage mapping plan for the data structures based on the classifying and generating, by the compiler, mapping code that implements the storage mapping plan, wherein the mapping code is part of the executable module and wherein the mapping code maps storing of the data structures to storage locations in the active memory devices.Type: GrantFiled: November 5, 2012Date of Patent: December 16, 2014Assignee: International Business Machines CorporationInventors: Tong Chen, John K. O'Brien, Zehra Sura
-
Patent number: 8914779Abstract: According to one embodiment, a system including a compiler to produce an executable module to be executed by a computer system including a main processor and active memory devices is provided. The system configured to perform a method including dividing source code into code sections, identifying a first code section to be executed by the active memory devices and identifying data structures that are used by the first code section. The method also includes classifying the data structures based on pre-defined attributes, formulating, by the compiler, a storage mapping plan for the data structures based on the classifying and generating, by the compiler, mapping code that implements the storage mapping plan, wherein the mapping code is part of the executable module and wherein the mapping code maps storing of the data structures to storage locations in the active memory devices.Type: GrantFiled: November 26, 2012Date of Patent: December 16, 2014Assignee: International Business Machines CorporationInventors: Tong Chen, John K. O'Brien, Zehra Sura
-
Patent number: 8863099Abstract: According to one embodiment of the present invention, a method for operation of a computer system including a main processor, a first and a second active memory device includes receiving an executable module generated by a compiler, wherein the executable module includes a code section identified as executable by a first processing element in the first active memory device and a second processing element in the second active memory device. The method further includes copying the code section to memory in the first device based on the code section being executable on the first device, copying the code section from the memory in the first active memory device to an instruction buffer of the first processing element and copying the code section from the memory in the first device to the second device based on the code section being executable on the second device.Type: GrantFiled: November 5, 2012Date of Patent: October 14, 2014Assignee: International Business Machines CorporationInventors: Tong Chen, John K. O'Brien, Zehra Sura
-
Publication number: 20140129787Abstract: According to one embodiment, a method for a compiler to produce an executable module to be executed by a computer system including a main processor and active memory devices includes dividing source code into code sections, identifying a first code section to be executed by the active memory devices, wherein the first code section is one of the code sections and identifying data structures that are used by the first code section. The method also includes classifying the data structures based on pre-defined attributes, formulating, by the compiler, a storage mapping plan for the data structures based on the classifying and generating, by the compiler, mapping code that implements the storage mapping plan, wherein the mapping code is part of the executable module and wherein the mapping code maps storing of the data structures to storage locations in the active memory devices.Type: ApplicationFiled: November 5, 2012Publication date: May 8, 2014Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Tong Chen, John K. O'Brien, Zehra Sura
-
Publication number: 20140130023Abstract: According to one embodiment of the present invention, a computer system is provided where the computer system includes a main processor, first and second active memory device. The computer system is configured to perform a method including receiving an executable module generated by a compiler, wherein the executable module includes a code section identified as executable by a first processing element in the first active memory device and a second processing element in the second active memory device. The method includes copying the code section to memory in the first device based on the code section being executable on the first device, copying the code section from the first active memory device to an instruction buffer of the first processing element and copying the code section from the first device to the second device based on the code section being executable on the second device.Type: ApplicationFiled: November 20, 2012Publication date: May 8, 2014Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Tong Chen, John K. O'Brien, Zehra Sura
-
Publication number: 20140130022Abstract: According to one embodiment of the present invention, a method for operation of a computer system including a main processor, a first and a second active memory device includes receiving an executable module generated by a compiler, wherein the executable module includes a code section identified as executable by a first processing element in the first active memory device and a second processing element in the second active memory device. The method further includes copying the code section to memory in the first device based on the code section being executable on the first device, copying the code section from the memory in the first active memory device to an instruction buffer of the first processing element and copying the code section from the memory in the first device to the second device based on the code section being executable on the second device.Type: ApplicationFiled: November 5, 2012Publication date: May 8, 2014Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Tong Chen, John K. O'Brien, Zehra Sura
-
Publication number: 20110283067Abstract: Target memory hierarchy specification in a multi-core computer processing system is provided including a system for implementing prefetch instructions. The system includes a first core processor, a dedicated cache corresponding to the first core processor, and a second core processor. The second core processor includes instructions for executing a prefetch instruction that specifies a memory location and the dedicated local cache corresponding to the first core processor. Executing the prefetch instruction includes retrieving data from the memory location and storing the retrieved data on the dedicated local cache corresponding to the first core processor.Type: ApplicationFiled: May 11, 2010Publication date: November 17, 2011Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Tong Chen, Yaoqing Gao, Kevin K. O'Brien, Zehra Sura, Lixin Zhang
-
Patent number: 7836256Abstract: One embodiment of the present method and apparatus for application-specific dynamic cache placement includes grouping sets of data in a cache memory system into two or more virtual partitions and processing a load/store instruction in accordance with the virtual partitions, where the load/store instruction specifies at least one of the virtual partitions to which the load/store instruction is assigned.Type: GrantFiled: June 30, 2008Date of Patent: November 16, 2010Assignee: International Business Machines CorporationInventors: Krishnan Kunjunny Kailas, Rajiv Alazhath Ravindran, Zehra Sura
-
Patent number: 7502890Abstract: One embodiment of the present method and apparatus for dynamic priority-based cache replacement includes selectively assigning relative priority values to at least a subset of data items in the cache memory system, fetching a new data item to load into the cache memory system, the data item being associated with a priority value, and selecting an existing data item from the cache memory system to replace with the new data item, in accordance with the relative priority values and the priority value of the new data item.Type: GrantFiled: July 7, 2006Date of Patent: March 10, 2009Assignee: International Business Machines CorporationInventors: Krishnan Kunjunny Kailas, Rajiv Alazhath Ravindran, Zehra Sura
-
Publication number: 20080270705Abstract: One embodiment of the present method and apparatus for application-specific dynamic cache placement includes grouping sets of data in a cache memory system into two or more virtual partitions and processing a load/store instruction in accordance with the virtual partitions, where the load/store instruction specifies at least one of the virtual partitions to which the load/store instruction is assigned.Type: ApplicationFiled: June 30, 2008Publication date: October 30, 2008Inventors: KRISHNAN KUNJUNNY KAILAS, Rajiv Alazhath Ravindran, Zehra Sura
-
Publication number: 20080010413Abstract: One embodiment of the present method and apparatus for application-specific dynamic cache placement includes grouping sets of data in a cache memory system into two or more virtual partitions and processing a load/store instruction in accordance with the virtual partitions, where the load/store instruction specifies at least one of the virtual partitions to which the load/store instruction is assigned.Type: ApplicationFiled: July 7, 2006Publication date: January 10, 2008Inventors: Krishnan Kunjunny Kailas, Rajiv Alazhath Ravindran, Zehra Sura
-
Publication number: 20080010414Abstract: One embodiment of the present method and apparatus for dynamic priority-based cache replacement includes selectively assigning relative priority values to at least a subset of data items in the cache memory system, fetching a new data item to load into the cache memory system, the data item being associated with a priority value, and selecting an existing data item from the cache memory system to replace with the new data item, in accordance with the relative priority values and the-priority value of the new data item.Type: ApplicationFiled: July 7, 2006Publication date: January 10, 2008Inventors: Krishnan Kunjunny Kailas, Rajiv Alazhath Ravindran, Zehra Sura
-
Publication number: 20070261042Abstract: A compiler implemented software cache apparatus and method in which non-aliased explicitly fetched data are excluded are provided. With the mechanisms of the illustrative embodiments, a compiler uses a forward data flow analysis to prove that there is no alias between the cached data and explicitly fetched data. Explicitly fetched data that has no alias in the cached data are excluded from the software cache. Explicitly fetched data that has aliases in the cached data are allowed to be stored in the software cache. In this way, there is no runtime overhead to maintain the correctness of the two copies of data. Moreover, the number of lines of the software cache that must be protected from eviction is decreased. This leads to a decrease in the amount of computation cycles required by the cache miss handler when evicting cache lines during cache miss handling.Type: ApplicationFiled: April 14, 2006Publication date: November 8, 2007Inventors: Tong Chen, John O'Brien, Kathryn O'Brien, Byoungro So, Zehra Sura, Tao Zhang