Patents by Inventor Arun Radhakrishnan

Arun Radhakrishnan has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11899588
    Abstract: A graphics processing unit (GPU) includes a table located in a memory of the GPU and a cache hierarchy. The table contains an address of inactive data in a cache hierarchy of the GPU in which the inactive data is associated with an intermediate render target. The cache hierarchy is responsive to an eviction event by discarding the inactive data from the cache hierarchy without performing a writeback to a system memory associated with the GPU based on the address of the inactive data being contained in the table. The cache hierarchy may obtain the address of the inactive data from the table, and the inactive data may be located in a last-level cache of the cache hierarchy. In one embodiment, the address of inactive data in a cache hierarchy of the GPU includes a range of addresses for the inactive data.
    Type: Grant
    Filed: February 12, 2021
    Date of Patent: February 13, 2024
    Inventors: Anshujit Sharma, Sushant Kondguli, Zhenhong Liu, Wilson Wai Lun Fung, Arun Radhakrishnan, Wayne Yamamoto
  • Patent number: 11675782
    Abstract: A disparate data source aggregation system and methods are provided which may pull or retrieve talent data or features from disparate data sources, automatically correlate the data across the different data sources, build a self-adjusting system database that captures the talent data from the disparate data sources, and lets users search, query and build model insights on the aggregated data of the system database without human intervention. A method for disparate data source aggregation may include: extracting a first feature set having a first extracted feature and a second feature set having a second extracted feature; determining, if the first extracted feature of the first feature set matches the second extracted feature of the second feature set; and aggregating the first feature set with the second feature set if the first extracted feature of the first feature set matches the second extracted feature of the second feature set.
    Type: Grant
    Filed: December 20, 2021
    Date of Patent: June 13, 2023
    Assignee: SAP SE
    Inventors: Stacy Chapman, Satish Sallakonda, Arun Radhakrishnan
  • Patent number: 11610281
    Abstract: A method of processing a workload in a graphics processing unit (GPU) may include detecting a work item of the workload in the GPU, determining a cache policy for the work item, and operating at least a portion of a cache memory hierarchy in the GPU for at least a portion of the work item based on the cache policy. The work item may be detected based on information received from an application and/or monitoring one or more performance counters by a driver and/or hardware detection logic. The method may further include monitoring one or more performance counters, wherein the cache policy for the work item may be determined and/or changed based on the one or more performance counters. The cache policy for the work item may be selected based on a runtime learning model.
    Type: Grant
    Filed: January 11, 2021
    Date of Patent: March 21, 2023
    Inventors: Sushant Kondguli, Arun Radhakrishnan, Zachary D. Neyland, David C. Tannenbaum
  • Publication number: 20220114167
    Abstract: A disparate data source aggregation system and methods are provided which may pull or retrieve talent data or features from disparate data sources, automatically correlate the data across the different data sources, build a self-adjusting system database that captures the talent data from the disparate data sources, and lets users search, query and build model insights on the aggregated data of the system database without human intervention. A method for disparate data source aggregation may include: extracting a first feature set having a first extracted feature and a second feature set having a second extracted feature; determining, if the first extracted feature of the first feature set matches the second extracted feature of the second feature set; and aggregating the first feature set with the second feature set if the first extracted feature of the first feature set matches the second extracted feature of the second feature set.
    Type: Application
    Filed: December 20, 2021
    Publication date: April 14, 2022
    Inventors: Stacy Chapman, Satish Sallakonda, Arun Radhakrishnan
  • Publication number: 20220083473
    Abstract: A graphics processing unit (GPU) includes a table located in a memory of the GPU and a cache hierarchy. The table contains an address of inactive data in a cache hierarchy of the GPU in which the inactive data is associated with an intermediate render target. The cache hierarchy is responsive to an eviction event by discarding the inactive data from the cache hierarchy without performing a writeback to a system memory associated with the GPU based on the address of the inactive data being contained in the table. The cache hierarchy may obtain the address of the inactive data from the table, and the inactive data may be located in a last-level cache of the cache hierarchy. In one embodiment, the address of inactive data in a cache hierarchy of the GPU includes a range of addresses for the inactive data.
    Type: Application
    Filed: February 12, 2021
    Publication date: March 17, 2022
    Inventors: Anshujit SHARMA, Sushant KONDGULI, Zhenhong LIU, Wilson Wai Lun FUNG, Arun RADHAKRISHNAN, Wayne YAMAMOTO
  • Publication number: 20220067876
    Abstract: A method of processing a workload in a graphics processing unit (GPU) may include detecting a work item of the workload in the GPU, determining a cache policy for the work item, and operating at least a portion of a cache memory hierarchy in the GPU for at least a portion of the work item based on the cache policy. The work item may be detected based on information received from an application and/or monitoring one or more performance counters by a driver and/or hardware detection logic. The method may further include monitoring one or more performance counters, wherein the cache policy for the work item may be determined and/or changed based on the one or more performance counters. The cache policy for the work item may be selected based on a runtime learning model.
    Type: Application
    Filed: January 11, 2021
    Publication date: March 3, 2022
    Inventors: Sushant KONDGULI, Arun RADHAKRISHNAN, Zachary D. NEYLAND, David C. TANNENBAUM
  • Patent number: 11216452
    Abstract: A disparate data source aggregation system and methods are provided which may pull or retrieve talent data or features from disparate data sources, automatically correlate the data across the different data sources, build a self-adjusting system database that captures the talent data from the disparate data sources, and lets users search, query and build model insights on the aggregated data of the system database without human intervention. A method for disparate data source aggregation may include: extracting a first feature set having a first extracted feature and a second feature set having a second extracted feature; determining, if the first extracted feature of the first feature set matches the second extracted feature of the second feature set; and aggregating the first feature set with the second feature set if the first extracted feature of the first feature set matches the second extracted feature of the second feature set.
    Type: Grant
    Filed: October 27, 2018
    Date of Patent: January 4, 2022
    Assignee: SAP SE
    Inventors: Stacy Chapman, Satish Sallakonda, Arun Radhakrishnan
  • Patent number: 10540287
    Abstract: Apparatuses and methods of manufacturing same, systems, and methods for a spatial memory streaming (SMS) prefetch engine are described. In one aspect, the SMS prefetch engine includes a pattern history table (PHT), which has a table in which each entry has an offset list field comprising sub-fields for offset values from a base offset value within a region and a per-offset confidence field comprising sub-fields for per-offset confidence levels corresponding to each offset value. When a PHT entry is activated, the per-offset confidence values corresponding to each offset value in the offset list field of the PHT entry are updated by matching current accesses to the stored offset values in the offset list field of the activated PHT entry. Continuous learning may be provided to the SMS engine at least by the per-offset confidence levels.
    Type: Grant
    Filed: August 30, 2017
    Date of Patent: January 21, 2020
    Assignee: Samsung Electronics Co., Ltd
    Inventors: Edward A Brekelbaum, Arun Radhakrishnan
  • Patent number: 10417130
    Abstract: Apparatuses, systems, methods for a spatial memory streaming (SMS) prefetch engine are described. In one aspect, an SMS prefetch engine uses trigger-to-trigger stride detection to promote training table entries to pattern history table (PHT) entries and to drive spatially related prefetches in more distant regions. In another aspect, an SMS prefetch engine maintains a blacklist of program counter (PC) values to not use as trigger values. In yet another aspect, an SMS prefetch engine uses hashed values of certain fields, such as the trigger PC, in entries of, e.g., filter tables, training tables, and PHTs, as index values for the table.
    Type: Grant
    Filed: October 10, 2017
    Date of Patent: September 17, 2019
    Assignee: Samsung Electronics Co., Ltd
    Inventors: Edward A Brekelbaum, Arun Radhakrishnan
  • Publication number: 20190129996
    Abstract: A disparate data source aggregation system and methods are provided which may pull or retrieve talent data or features from disparate data sources, automatically correlate the data across the different data sources, build a self-adjusting system database that captures the talent data from the disparate data sources, and lets users search, query and build model insights on the aggregated data of the system database without human intervention. A method for disparate data source aggregation may include: extracting a first feature set having a first extracted feature and a second feature set having a second extracted feature; determining, if the first extracted feature of the first feature set matches the second extracted feature of the second feature set; and aggregating the first feature set with the second feature set if the first extracted feature of the first feature set matches the second extracted feature of the second feature set.
    Type: Application
    Filed: October 27, 2018
    Publication date: May 2, 2019
    Applicant: Tahlent Inc DBA SwoopTalent
    Inventors: Stacy Chapman, Satish Sallakonda, Arun Radhakrishnan
  • Publication number: 20180329823
    Abstract: Apparatuses, systems, methods for a spatial memory streaming (SMS) prefetch engine are described. In one aspect, an SMS prefetch engine uses trigger-to-trigger stride detection to promote training table entries to pattern history table (PHT) entries and to drive spatially related prefetches in more distant regions. In another aspect, an SMS prefetch engine maintains a blacklist of program counter (PC) values to not use as trigger values. In yet another aspect, an SMS prefetch engine uses hashed values of certain fields, such as the trigger PC, in entries of, e.g., filter tables, training tables, and PHTs, as index values for the table.
    Type: Application
    Filed: October 10, 2017
    Publication date: November 15, 2018
    Inventors: Edward A. BREKELBAUM, Arun RADHAKRISHNAN
  • Publication number: 20180329822
    Abstract: Apparatuses and methods of manufacturing same, systems, and methods for a spatial memory streaming (SMS) prefetch engine are described. In one aspect, the SMS prefetch engine includes a pattern history table (PHT), which has a table in which each entry has an offset list field comprising sub-fields for offset values from a base offset value within a region and a per-offset confidence field comprising sub-fields for per-offset confidence levels corresponding to each offset value. When a PHT entry is activated, the per-offset confidence values corresponding to each offset value in the offset list field of the PHT entry are updated by matching current accesses to the stored offset values in the offset list field of the activated PHT entry. Continuous learning may be provided to the SMS engine at least by the per-offset confidence levels.
    Type: Application
    Filed: August 30, 2017
    Publication date: November 15, 2018
    Inventors: Edward A. BREKELBAUM, Arun Radhakrishnan
  • Patent number: 10031851
    Abstract: A computing system includes: an instruction dispatch module module configured to receive a program instruction; and an address reordering module, coupled to the instruction dispatch module, configured to filter the program instruction when the program instruction is a hit in a cache-line in a prefetch filter. The computer system further includes: an instruction dispatch module configured to receive a program instruction; an address reordering module, coupled to the instruction dispatch module, configured to: allocate a tag in a tag module for the program instruction in a program order, allocate a virtual address in a virtual address module for the program instruction and out-of-order relative to the program order, and insert a pointer associated with the tag to link the tag to the virtual address.
    Type: Grant
    Filed: January 9, 2017
    Date of Patent: July 24, 2018
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Karthik Sundaram, Arun Radhakrishnan
  • Patent number: 9665491
    Abstract: According to one general aspect, a method may include monitoring a plurality of pre-fetch cache requests associated with a data stream. The method may also include evaluating an accuracy of the pre-fetch cache requests. The method may further include, based at least in part upon the accuracy of the pre-fetch cache requests, adjusting a maximum amount of data that is allowably pre-fetched in excess of a data stream's current actual demand for data.
    Type: Grant
    Filed: June 3, 2015
    Date of Patent: May 30, 2017
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Arun Radhakrishnan, Karthik Sundaram
  • Publication number: 20170116128
    Abstract: A computing system includes: an instruction dispatch module module configured to receive a program instruction; and an address reordering module, coupled to the instruction dispatch module, configured to filter the program instruction when the program instruction is a hit in a cache-line in a prefetch filter. The computer system further includes: an instruction dispatch module configured to receive a program instruction; an address reordering module, coupled to the instruction dispatch module, configured to: allocate a tag in a tag module for the program instruction in a program order, allocate a virtual address in a virtual address module for the program instruction and out-of-order relative to the program order, and insert a pointer associated with the tag to link the tag to the virtual address.
    Type: Application
    Filed: January 9, 2017
    Publication date: April 27, 2017
    Inventors: Karthik Sundaram, Arun Radhakrishnan
  • Patent number: 9569361
    Abstract: According to one general aspect, an apparatus may include a cache pre-fetcher, and a pre-fetch scheduler. The cache pre-fetcher may be configured to predict, based at least in part upon a virtual address, data to be retrieved from a memory system. The pre-fetch scheduler may be configured to convert the virtual address of the data to a physical address of the data, and request the data from one of a plurality of levels of the memory system. The memory system may include a plurality of levels, each level of the memory system configured to store data.
    Type: Grant
    Filed: July 7, 2014
    Date of Patent: February 14, 2017
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Arun Radhakrishnan, Kevin Lepak, Rama Gopal, Murali Chinnakonda, Karthik Sundaram, Brian Grayson
  • Patent number: 9542323
    Abstract: A computing system includes: an instruction dispatch module configured to receive a program instruction; an address reordering module, coupled to the instruction dispatch module, configured to filter the program instruction when the program instruction is a hit in a cache-line in a prefetch filter. The computer system further includes: an instruction dispatch module configured to receive a program instruction; an address reordering module, coupled to the instruction dispatch module, configured to: allocate a tag in a tag module for the program instruction in a program order, allocate a virtual address in a virtual address module for the program instruction in an out-of-order relative to the program order, and insert a pointer associated with the tag to link the tag to the virtual address.
    Type: Grant
    Filed: September 26, 2014
    Date of Patent: January 10, 2017
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Karthik Sundaram, Arun Radhakrishnan
  • Publication number: 20160054997
    Abstract: A computing system includes: an instruction dispatch module configured to receive an address stream; a prefetch module, coupled to the instruction dispatch module, configured to: train to concurrently detect a single-stride pattern or a multi-stride pattern from the address stream, speculatively fetch a program data based on the single-stride pattern or the multi-stride pattern, and continue to train for the single-stride pattern with a larger value for a stride count or for the multi-stride pattern.
    Type: Application
    Filed: August 21, 2015
    Publication date: February 25, 2016
    Inventors: Arun Radhakrishnan, Karthik Sundaram, Brian Grayson
  • Publication number: 20160019155
    Abstract: According to one general aspect, a method may include monitoring a plurality of pre-fetch cache requests associated with a data stream. The method may also include evaluating an accuracy of the pre-fetch cache requests. The method may further include, based at least in part upon the accuracy of the pre-fetch cache requests, adjusting a maximum amount of data that is allowably pre-fetched in excess of a data stream's current actual demand for data.
    Type: Application
    Filed: June 3, 2015
    Publication date: January 21, 2016
    Inventors: Arun RADHAKRISHNAN, Karthik SUNDARAM
  • Publication number: 20150278100
    Abstract: A computing system includes: an instruction dispatch module configured to receive a program instruction; an address reordering module, coupled to the instruction dispatch module, configured to filter the program instruction when the program instruction is a hit in a cache-line in a prefetch filter. The computer system further includes: an instruction dispatch module configured to receive a program instruction; an address reordering module, coupled to the instruction dispatch module, configured to: allocate a tag in a tag module for the program instruction in a program order, allocate a virtual address in a virtual address module for the program instruction in an out-of-order relative to the program order, and insert a pointer associated with the tag to link the tag to the virtual address.
    Type: Application
    Filed: September 26, 2014
    Publication date: October 1, 2015
    Inventors: Karthik Sundaram, Arun Radhakrishnan