Patents by Inventor Roy Kim

Roy Kim has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 10275285
    Abstract: Data transformation caching in an artificial intelligence infrastructure that includes one or more storage systems and one or more graphical processing unit (‘GPU’) servers, including: identifying, in dependence upon one or more machine learning models to be executed on the GPU servers, one or more transformations to apply to a dataset; generating, in dependence upon the one or more transformations, a transformed dataset; storing, within one or more of the storage systems, the transformed dataset; receiving a plurality of requests to transmit the transformed dataset to one or more of the GPU servers; and responsive to each request, transmitting, from the one or more storage systems to the one or more GPU servers without re-performing the one or more transformations on the dataset, the transformed dataset.
    Type: Grant
    Filed: July 26, 2018
    Date of Patent: April 30, 2019
    Assignee: Pure Storage, Inc.
    Inventors: Brian Gold, Emily Watkins, Ivan Jibaja, Igor Ostrovsky, Roy Kim
  • Patent number: 10275176
    Abstract: Data transformation offloading in an artificial intelligence infrastructure that includes one or more storage systems and one or more graphical processing unit (‘GPU’) servers, including: storing, within the storage system, a dataset; identifying, in dependence upon one or more machine learning models to be executed on the GPU servers, one or more transformations to apply to the dataset; and generating, by the storage system in dependence upon the one or more transformations, a transformed dataset.
    Type: Grant
    Filed: July 26, 2018
    Date of Patent: April 30, 2019
    Assignee: Pure Storage, Inc.
    Inventors: Brian Gold, Emily Watkins, Ivan Jibaja, Igor Ostrovsky, Roy Kim
  • Publication number: 20190121673
    Abstract: Data transformation caching in an artificial intelligence infrastructure that includes one or more storage systems and one or more graphical processing unit (‘GPU’) servers, including: identifying, in dependence upon one or more machine learning models to be executed on the GPU servers, one or more transformations to apply to a dataset; generating, in dependence upon the one or more transformations, a transformed dataset; storing, within one or more of the storage systems, the transformed dataset; receiving a plurality of requests to transmit the transformed dataset to one or more of the GPU servers; and responsive to each request, transmitting, from the one or more storage systems to the one or more GPU servers without re-performing the one or more transformations on the dataset, the transformed dataset.
    Type: Application
    Filed: July 26, 2018
    Publication date: April 25, 2019
    Inventors: BRIAN GOLD, EMILY WATKINS, IVAN JIBAJA, IGOR OSTROVSKY, ROY KIM
  • Publication number: 20190121566
    Abstract: Data transformation offloading in an artificial intelligence infrastructure that includes one or more storage systems and one or more graphical processing unit (‘GPU’) servers, including: storing, within the storage system, a dataset; identifying, in dependence upon one or more machine learning models to be executed on the GPU servers, one or more transformations to apply to the dataset; and generating, by the storage system in dependence upon the one or more transformations, a transformed dataset.
    Type: Application
    Filed: July 26, 2018
    Publication date: April 25, 2019
    Inventors: BRIAN GOLD, EMILY WATKINS, IVAN JIBAJA, IGOR OSTROVSKY, ROY KIM
  • Publication number: 20190121889
    Abstract: Ensuring reproducibility in an artificial intelligence infrastructure that includes one or more storage systems and one or more graphical processing unit (‘GPU’) servers, including: identifying, by a unified management plane, one or more transformations applied to a dataset by the artificial intelligence infrastructure, wherein applying the one or more transformations to the dataset causes the artificial intelligence infrastructure to generate a transformed dataset; storing, within the one or more storage systems, information describing the dataset, the one or more transformations applied to the dataset, and the transformed dataset; identifying, by the unified management plane, one or more machine learning models executed by the artificial intelligence infrastructure using the transformed dataset as input; and storing, within the one or more storage systems, information describing one or more machine learning models executed using the transformed dataset as input.
    Type: Application
    Filed: July 26, 2018
    Publication date: April 25, 2019
    Inventors: BRIAN GOLD, EMILY WATKINS, IVAN JIBAJA, IGOR OSTROVSKY, ROY KIM
  • Patent number: 10200508
    Abstract: A special-purpose processing system, a method of carrying out sharing special-purpose processing resources and a graphics processing system. In one embodiment, the special-purpose processing system includes: (1) a special-purpose processing resource and (2) a Representational State Transfer (ReST) application programming interface operable to process data using the special-purpose processing resource in response to stateless commands based on a standard protocol selected from the group consisting of: (2a) a standard network protocol and (2b) a standard database query protocol.
    Type: Grant
    Filed: January 7, 2014
    Date of Patent: February 5, 2019
    Assignee: Nvidia Corporation
    Inventors: Jonathan Cohen, Michael Houston, Frank Jargstorff, Eric Young, Roy Kim
  • Publication number: 20150081866
    Abstract: A special-purpose processing system, a method of carrying out sharing special-purpose processing resources and a graphics processing system. In one embodiment, the special-purpose processing system includes: (1) a special-purpose processing resource and (2) a Representational State Transfer (ReST) application programming interface operable to process data using the special-purpose processing resource in response to stateless commands based on a standard protocol selected from the group consisting of: (2a) a standard network protocol and (2b) a standard database query protocol.
    Type: Application
    Filed: January 7, 2014
    Publication date: March 19, 2015
    Applicant: Nvidia Corporation
    Inventors: Jonathan Cohen, Michael Houston, Frank Jargstorff, Eric Young, Roy Kim
  • Publication number: 20060015689
    Abstract: The present invention provides parallel processing of write-back and reload operations in a cache system and optimum circuit utilisation by implementing moveable buffers in a cache storage. However, the data and associated pointers are not permanently assigned to a particular buffer—hence, the buffers can move logically around in the facility. Reload pointer is pointing to an empty entry so that retrieved data from the main memory or equal hierarchy cache on cache miss can be always be accommodated. Victim pointer is always pointing to a modified entry for the next candidate of write-back operation. Write-back operation is necessary with reload operation in order to make a free entry for further cache miss handling unless free entry exists. Because of these moveable pointers for reload buffer and victim buffer and integrated write-back buffer in the cache, intra cache data movement is not necessary which improves cache miss handling performance.
    Type: Application
    Filed: July 15, 2004
    Publication date: January 19, 2006
    Applicants: International Business Machines Corporation, Sony Computer Entertainment Inc.
    Inventors: Yasukichi Okawa, Roy Kim, Peichun Liu, Thuong Truong
  • Publication number: 20050289300
    Abstract: The present invention provides for managing an atomic facility cache write back state machine. A first write back selection is made. A reservation pointer pointing to the reserved line in the atomic facility data array is established. A next write back selection is made. An entry for the reservation point for the next write back selection is removed, whereby the valid reservation line is precluded form being selected for the write back. This prevents a modified command from being invalidated.
    Type: Application
    Filed: June 24, 2004
    Publication date: December 29, 2005
    Applicants: International Business Machines Corporation, Sony Computer Entertainment Inc.
    Inventors: Roy Kim, Yasukichi Okawa, Thuong Truong
  • Publication number: 20050273563
    Abstract: A cache write back operation, write back modified data to memory from cache data array to fix inconsistency between them can be cancelled by the results of a comparison of the progress between a write back and snoop push or snoop kill operation. Write back is intended to make an empty slot to accommodate a reload data due to a cache miss and since a snoop push or snoop kill operation creates an invalid entry in the cache, write back is not needed. If simultaneous push or kill with write back operation exist, then write back machine is late cancelled. System performance improves due to preserving more cache lines in cache data array for possible future reuse.
    Type: Application
    Filed: June 3, 2004
    Publication date: December 8, 2005
    Applicants: International Business Machines Corporation, Sony Computer Entertainment Inc.
    Inventors: Roy Kim, Yasukichi Okawa, Thuong Truong
  • Publication number: 20050259811
    Abstract: The present invention relates to headsets for communication devices, including, but not limited to, cell phones, smart phones, PDAs, desk top computers, and lap top computers. Specifically, this invention teaches a head set which may be fully retracted and then integrated with a communication device. For a wired head set, the microphone is located in proximity to the earpiece. A flexible sound tube extends from the microphone to the user's mouth. This flexible sound tube is pliable such that the user can bend the opening of the sound tube towards the user's mouth. In addition, the sound tube is coupled to the same wire that transmits signals between the microphone/earpiece and the communication device. The coupling of the flexible sound tube to the transmission wire allows the mouthpiece to be retracted along with the wire, up and until the microphone/earpiece. For wireless headsets, a flexible sound tube connects the user's mouth with the microphone/earpiece.
    Type: Application
    Filed: May 24, 2004
    Publication date: November 24, 2005
    Inventors: Daniel Kimm, Roy Kim, Jino Kim