Patents by Inventor Joo Hwan Lee
Joo Hwan Lee has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 11940837Abstract: A display that includes a display panel and a window laminated with the display panel is presented. The display panel may include: a main panel region including a first side extending in a first direction and a second side extending in a second direction crossing the first direction; a first sub-panel region that is in contact with the first side and is bent; and a second sub-panel region that is in contact with the second side and is bent. A panel corner part of the main panel region adjacent to the first sub-panel region and the second sub-panel region is rounded.Type: GrantFiled: May 29, 2019Date of Patent: March 26, 2024Assignee: Samsung Display Co., Ltd.Inventors: Seung-Hwan Cho, Jong Hyun Choi, Ju Chan Park, Joo Sun Yoon, Jong Hyuk Lee
-
Patent number: 11907814Abstract: A system and method for machine learning. The system includes a GPU with a GPU memory, and a key value storage device connected to the GPU memory. The method includes, writing, by the GPU, a key value request to a key value request queue in a input-output region of the GPU memory, the key value request including a key. The method further includes reading, by the key value storage device, the key value request from the key value request queue, and writing, by the key value storage device, in response to the key value request, a value to the input-output region of the GPU memory, the value corresponding to the key of the key value request.Type: GrantFiled: November 22, 2021Date of Patent: February 20, 2024Assignee: Samsung Electronics Co., Ltd.Inventors: Joo Hwan Lee, Yang Seok Ki
-
Publication number: 20230367675Abstract: According to one general aspect, an apparatus may include a host interface circuit configured to receive offloading instructions from a host processing device, wherein the offloading instructions instruct the apparatus to compute an error correction code associated with a plurality of data elements. The apparatus may include a memory interface circuit configured to receive the plurality of data elements. The apparatus may include a plurality of memory buffer circuits configured to temporarily store the plurality of data elements. The apparatus may include a plurality of error code computation circuits configured to, at least in part, compute the error correction code without additional processing by the host processing device.Type: ApplicationFiled: July 17, 2023Publication date: November 16, 2023Inventors: Mian QIN, Joo Hwan LEE, Rekha PITCHUMANI, Yang Seok KI
-
Patent number: 11791838Abstract: An accelerator is disclosed. The accelerator may include a memory that may store a dictionary table. An address generator may be configured to generate an address in the dictionary table based on an encoded value, which may have an encoded width. An output filter may be configured to filter a decoded value from the dictionary table based on the encoded value, the encoded width, and a decoded width of the decoded data. The accelerator may be configured to support at least two different encoded widths.Type: GrantFiled: June 24, 2021Date of Patent: October 17, 2023Inventors: Sahand Salamat, Joo Hwan Lee, Armin Haj Aboutalebi, Praveen Krishnamoorthy, Xiaodong Zhao, Hui Zhang, Yang Seok Ki
-
Patent number: 11726876Abstract: According to one general aspect, an apparatus may include a host interface circuit configured to receive offloading instructions from a host processing device, wherein the offloading instructions instruct the apparatus to compute an error correction code associated with a plurality of data elements. The apparatus may include a memory interface circuit configured to receive the plurality of data elements. The apparatus may include a plurality of memory buffer circuits configured to temporarily store the plurality of data elements. The apparatus may include a plurality of error code computation circuits configured to, at least in part, compute the error correction code without additional processing by the host processing device.Type: GrantFiled: July 2, 2021Date of Patent: August 15, 2023Inventors: Mian Qin, Joo Hwan Lee, Rekha Pitchumani, Yang Seok Ki
-
Patent number: 11687771Abstract: Computing resources are optimally allocated for a multipath neural network using a multipath neural network analyzer that includes an interface and a processing device. The interface receives a multipath neural network that includes two or more paths. A first path includes one or more layers. A first layer of the first path corresponds to a first kernel that runs on a compute unit that includes two or more cores. The processing device allocates to the first kernel a minimum number of cores of the compute unit and a maximum number of cores of the compute unit. The minimum number of cores of the compute unit is allocated based on the first kernel being run concurrently with at least one other kernel on the compute unit and the maximum number of cores of the compute unit is allocated based on the first kernel being run alone on the compute unit.Type: GrantFiled: June 14, 2019Date of Patent: June 27, 2023Inventors: Joo Hwan Lee, Yang Seok Ki, Behnam Pourghassemi Najafabadi
-
Patent number: 11620510Abstract: Computing resources may be optimally allocated for a multipath neural network using a multipath neural network analyzer that includes an interface and a processing device. The interface receives a multipath neural network. The processing device generates the multipath neural network to include one or more layers of a critical path through the multipath neural network that are allocated a first allocation of computing resources that are available to execute the multipath neural network. The critical path limits throughput of the multipath neural network. The first allocation of computing resources reduces an execution time of the multipath neural network to be less than a baseline execution time of a second allocation of computing resources for the multipath neural network. The first allocation of computing resources for a first layer of the critical path is different than the second allocation of computing resources for the first layer of the critical path.Type: GrantFiled: June 14, 2019Date of Patent: April 4, 2023Inventors: Behnam Pourghassemi Najafabadi, Joo Hwan Lee, Yang Seok Ki
-
Patent number: 11567971Abstract: A method of processing data in a system having a host and a storage node may include performing a shuffle operation on data stored at the storage node, wherein the shuffle operation may include performing a shuffle write operation, and performing a shuffle read operation, wherein at least a portion of the shuffle operation is performed by an accelerator at the storage node. A method for partitioning data may include sampling, at a device, data from one or more partitions based on a number of samples, transferring the sampled data from the device to a host, determining, at the host, one or more splitters based on the sampled data, communicating the one or more splitters from the host to the device, and partitioning, at the device, data for the one or more partitions based on the one or more splitters.Type: GrantFiled: December 4, 2020Date of Patent: January 31, 2023Inventors: Hui Zhang, Joo Hwan Lee, Yiqun Zhang, Armin Haj Aboutalebi, Xiaodong Zhao, Praveen Krishnamoorthy, Andrew Chang, Yang Seok Ki
-
Patent number: 11468312Abstract: A system and method for memory management. In one embodiment, the method includes generating a dependency structure comprising one or more task identifiers and one or more data object identifiers. The dependency structure includes a list of one or more dependencies for a first data object identifier of the one or more data object identifiers, a first dependency of the list identifying a first task for which a data object identified by the first data object identifier is an input. The method further includes counting the number of dependencies for the first data object identifier, decrementing the count by one when the first task completes execution, and, when the count reaches zero, deallocating the first data object.Type: GrantFiled: April 9, 2018Date of Patent: October 11, 2022Assignee: Samsung Electronics Co., Ltd.Inventors: Joo Hwan Lee, Yang Seok Ki
-
Patent number: 11461869Abstract: A system and method for machine learning, with a processing circuit executing an operating system, the processing circuit being connected to a first memory and to a second memory. In some embodiments, the method includes: requesting, by a user level process, from the operating system, a first memory allocation from the first memory, the first memory allocation including a plurality of first segments, a first segment of the plurality of first segments having a size sufficient to store a data object of a first kind; determining, by the user level process, that a result of a first calculation will be a data object of the first kind; and in response to determining that the result of the first calculation will be a data object of the first kind: determining, by the user level process, that none of the first segments are unused.Type: GrantFiled: May 10, 2018Date of Patent: October 4, 2022Assignee: Samsung Electronics Co., Ltd.Inventors: Joo Hwan Lee, Yang Seok Ki
-
Publication number: 20220231698Abstract: An accelerator is disclosed. The accelerator may include a memory that may store a dictionary table. An address generator may be configured to generate an address in the dictionary table based on an encoded value, which may have an encoded width. An output filter may be configured to filter a decoded value from the dictionary table based on the encoded value, the encoded width, and a decoded width of the decoded data. The accelerator may be configured to support at least two different encoded widths.Type: ApplicationFiled: June 24, 2021Publication date: July 21, 2022Inventors: Sahand SALAMAT, JOO HWAN LEE, ARMIN HAJ ABOUTALEBI, PRAVEEN KRISHNAMOORTHY, XIAODONG ZHAO, HUI ZHANG, YANG SEOK KI
-
Publication number: 20220207040Abstract: A method of processing data may include receiving a stream of first keys associated with first data, receiving a stream of second keys associated with second data, comparing, in parallel, a batch of the first keys and a batch of the second keys, collecting one or more results from the comparing, and gathering one or more results from the collecting. The collecting may include reducing an index matrix and a mask matrix. Gathering one or more results may include storing, in a leftover vector, at least a portion of the one or more results from the collecting. Gathering one or more results further may include combining at least a portion of the leftover vector from a first cycle with at least a portion of the one or more results from the collecting from a second cycle.Type: ApplicationFiled: February 11, 2021Publication date: June 30, 2022Inventors: Shiyu LI, Yiqun ZHANG, Joo Hwan LEE, Yang Seok KI, Andrew CHANG
-
Publication number: 20220164122Abstract: A method of shuffling data may include shuffling a first batch of data using a first memory on a first level of a memory hierarchy to generate a first batch of shuffled data, shuffling a second batch of data using the first memory to generate a second batch of shuffled data, and storing the first batch of shuffled data and the second batch of shuffled data in a second memory on a second level of the memory hierarchy. The method may further include merging the first batch of shuffled data and the second batch of shuffled data. A data shuffling device may include a buffer memory configured to stream one or more records to a partitioning circuit and transfer, by random access, one or more records to a grouping circuit.Type: ApplicationFiled: April 7, 2021Publication date: May 26, 2022Inventors: Chen ZOU, Hui ZHANG, Joo Hwan LEE, Yang Seok KI
-
Publication number: 20220164708Abstract: A system and method for machine learning. The system includes a GPU with a GPU memory, and a key value storage device connected to the GPU memory. The method includes, writing, by the GPU, a key value request to a key value request queue in a input-output region of the GPU memory, the key value request including a key. The method further includes reading, by the key value storage device, the key value request from the key value request queue, and writing, by the key value storage device, in response to the key value request, a value to the input-output region of the GPU memory, the value corresponding to the key of the key value request.Type: ApplicationFiled: November 22, 2021Publication date: May 26, 2022Inventors: Joo Hwan Lee, Yang Seok Ki
-
Publication number: 20220156287Abstract: A method of processing data in a system having a host and a storage node may include performing a shuffle operation on data stored at the storage node, wherein the shuffle operation may include performing a shuffle write operation, and performing a shuffle read operation, wherein at least a portion of the shuffle operation is performed by an accelerator at the storage node. A method for partitioning data may include sampling, at a device, data from one or more partitions based on a number of samples, transferring the sampled data from the device to a host, determining, at the host, one or more splitters based on the sampled data, communicating the one or more splitters from the host to the device, and partitioning, at the device, data for the one or more partitions based on the one or more splitters.Type: ApplicationFiled: December 4, 2020Publication date: May 19, 2022Inventors: HUI ZHANG, JOO HWAN LEE, YIQUN ZHANG, ARMIN HAJ ABOUTALEBI, XIAODONG ZHAO, PRAVEEN KRISHNAMOORTHY, ANDREW CHANG, YANG SEOK KI
-
Patent number: 11249651Abstract: A storage system includes: a storage device to store an array of data elements associated with a sort operation; a storage interface to facilitate communications between the storage device and a host computer; and a reconfigurable processing device communicably connected to the storage device, the reconfigurable processing device including: memory to store input data read from the storage device, the input data corresponding to the array of data elements stored in the storage device; and a kernel including one or more compute components to execute the sort operation on the input data stored in the memory according to a SORT command received from the host computer. The reconfigurable processing device is to dynamically instantiate the one or more compute components to accelerate the sort operation.Type: GrantFiled: March 17, 2020Date of Patent: February 15, 2022Assignee: Samsung Electronics Co., Ltd.Inventors: Sahand Salamat, Hui Zhang, Joo Hwan Lee, Yang Seok Ki
-
Patent number: 11182694Abstract: A system and method for machine learning. The system includes a GPU with a GPU memory, and a key value storage device connected to the GPU memory. The method includes, writing, by the GPU, a key value request to a key value request queue in a input-output region of the GPU memory, the key value request including a key. The method further includes reading, by the key value storage device, the key value request from the key value request queue, and writing, by the key value storage device, in response to the key value request, a value to the input-output region of the GPU memory, the value corresponding to the key of the key value request.Type: GrantFiled: March 30, 2018Date of Patent: November 23, 2021Assignee: Samsung Electronics Co., Ltd.Inventors: Joo Hwan Lee, Yang Seok Ki
-
Publication number: 20210334162Abstract: According to one general aspect, an apparatus may include a host interface circuit configured to receive offloading instructions from a host processing device, wherein the offloading instructions instruct the apparatus to compute an error correction code associated with a plurality of data elements. The apparatus may include a memory interface circuit configured to receive the plurality of data elements. The apparatus may include a plurality of memory buffer circuits configured to temporarily store the plurality of data elements. The apparatus may include a plurality of error code computation circuits configured to, at least in part, compute the error correction code without additional processing by the host processing device.Type: ApplicationFiled: July 2, 2021Publication date: October 28, 2021Inventors: Mian QIN, Joo Hwan LEE, Rekha PITCHUMANI, Yang Seok KI
-
Patent number: 11061772Abstract: According to one general aspect, an apparatus may include a host interface circuit configured to receive offloading instructions from a host processing device, wherein the offloading instructions instruct the apparatus to compute an error correction code associated with a plurality of data elements. The apparatus may include a memory interface circuit configured to receive the plurality of data elements. The apparatus may include a plurality of memory buffer circuits configured to temporarily store the plurality of data elements. The apparatus may include a plurality of error code computation circuits configured to, at least in part, compute the error correction code without additional processing by the host processing device.Type: GrantFiled: February 8, 2019Date of Patent: July 13, 2021Inventors: Mian Qin, Joo Hwan Lee, Rekha Pitchumani, Yang Seok Ki
-
Publication number: 20210124500Abstract: A storage system includes: a storage device to store an array of data elements associated with a sort operation; a storage interface to facilitate communications between the storage device and a host computer; and a reconfigurable processing device communicably connected to the storage device, the reconfigurable processing device including: memory to store input data read from the storage device, the input data corresponding to the array of data elements stored in the storage device; and a kernel including one or more compute components to execute the sort operation on the input data stored in the memory according to a SORT command received from the host computer. The reconfigurable processing device is to dynamically instantiate the one or more compute components to accelerate the sort operation.Type: ApplicationFiled: March 17, 2020Publication date: April 29, 2021Inventors: Sahand Salamat, Hui Zhang, Joo Hwan Lee, Yang Seok Ki