Patents by Inventor Seung Hyub Jeon

Seung Hyub Jeon has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 10296379
    Abstract: Scheduling threads in a system with many cores includes generating a thread map where a connection relationship between a plurality of threads is represented by a frequency of inter-process communication (IPC) between threads, generating a core map where a connection relationship between a plurality of cores is represented by a hop between cores, and respectively allocating the plurality of threads to the plurality of cores defined by the core map, based on a thread allocation policy defining a mapping rule between the thread map and the core map.
    Type: Grant
    Filed: March 17, 2017
    Date of Patent: May 21, 2019
    Assignee: ELECTRONICS AND TELECOMMUNICATIONS RESEARCH INSTITUTE
    Inventors: Kang Ho Kim, Kwang Won Koh, Jin Mee Kim, Jeong Hwan Lee, Seung Hyub Jeon, Sung In Jung, Yeon Jeong Jeong, Seung Jun Cha
  • Publication number: 20190114193
    Abstract: Disclosed is an apparatus and method of processing input and output in a multi-kernel system. A method of processing input and output in a multi-kernel system according to the present disclosure includes: setting a shared memory between a first kernel on a main processor and a lightweight kernel on a parallel processor; setting a data transmission and reception channel between the first kernel on the main processor and the lightweight kernel on the parallel processor using the shared memory; providing, on the basis of the data transmission and reception channel, an input/output task that occurs in the lightweight kernel to the first kernel on the main processor; processing, by the first kernel on the main processor, an operation corresponding to the input/output task; and providing a result of the processing to the lightweight kernel.
    Type: Application
    Filed: October 12, 2018
    Publication date: April 18, 2019
    Inventors: Seung Jun CHA, Jin Mee KIM, Seung Hyub JEON, Sung In JUNG, Yeon Jeong JEONG
  • Publication number: 20170329642
    Abstract: Provided is a many-core system including a resource unit including a resource needed for execution of an operating system and a resource needed for execution of a lightweight kernel, a programing constructing unit configured to convert an input program into an application program and to load the application program into the resource unit, a run-time management unit configured to manage a running environment for executing the application program, and a self-organization management unit configured to monitor the application program and the resources in the resource unit, to dynamically adjust the running environment to prevent a risk factor from occurring during the execution of the application program, and to cure a risk factor which occurred.
    Type: Application
    Filed: August 11, 2016
    Publication date: November 16, 2017
    Inventors: Jin Mee KIM, Kwang Won KOH, Kang Ho KIM, Jeong Hwan LEE, Seung Hyub JEON, Sung In JUNG, Yeon Jeong JEONG, Seung Jun CHA
  • Patent number: 9804903
    Abstract: Disclosed herein are a data processing apparatus for pipeline execution acceleration and a method thereof. According to an exemplary embodiment of the present invention, the data processing apparatus for pipeline execution acceleration includes: a processor configured to sequentially execute a first application program and a second application program reading or writing a specific file; and a file system configured to complete a write for a file data for the specific file to a data block previously allocated from the first application program and provide the file data for the specific file to the second application program prior to executing a close call for the specific file from the first application program, when executing a read call for the specific file from the second application program.
    Type: Grant
    Filed: February 9, 2015
    Date of Patent: October 31, 2017
    Assignee: ELECTRONICS AND TELECOMMUNICATIONS RESEARCH INSTITUTE
    Inventors: Kang-Ho Kim, Kwang-Won Koh, Seung-Hyub Jeon
  • Publication number: 20170269966
    Abstract: Provided is a method of scheduling threads in a many-cores system. The method includes generating a thread map where a connection relationship between a plurality of threads is represented by a frequency of inter-process communication (IPC) between threads, generating a core map where a connection relationship between a plurality of cores is represented by a hop between cores, and respectively allocating the plurality of threads to the plurality of cores defined by the core map, based on a thread allocation policy defining a mapping rule between the thread map and the core map.
    Type: Application
    Filed: March 17, 2017
    Publication date: September 21, 2017
    Inventors: Kang Ho KIM, Kwang Won KOH, Jin Mee KIM, Jeong Hwan LEE, Seung Hyub JEON, Sung In JUNG, Yeon Jeong JEONG, Seung Jun CHA
  • Patent number: 9501394
    Abstract: Apparatus, method and systems for managing reference data, which can prevent duplicated data loading of reference data and eliminate redundancy of I/O operations for loading of the same reference data required by different virtual machines present in the same physical node to reduce use memory and I/O through sharing virtual machine leveled memories, are provided.
    Type: Grant
    Filed: June 9, 2015
    Date of Patent: November 22, 2016
    Assignee: ELECTRONICS AND TELECOMMUNICATIONS RESEARCH INSTITUTE
    Inventors: Kwang-Won Koh, Kang-Ho Kim, Seung-Hyub Jeon, Seung-Jo Bae
  • Publication number: 20160275162
    Abstract: A method and an apparatus for partitioning or combining massive data, which can efficiently partition and combine data when an operation is executed by being distributed to a plurality of nodes in an environment such as genome analysis, in which massive data can be partitioned and executed. The method includes storing meta information on partition or combination of at least one data, if a request for data is sensed, acquiring meta information corresponding to the data, partitioning or combining the data, based on the meta information, and transmitting the partitioned or combined data in response to the request.
    Type: Application
    Filed: March 15, 2016
    Publication date: September 22, 2016
    Inventor: Seung Hyub JEON
  • Publication number: 20150356003
    Abstract: Apparatus, method and systems for managing reference data, which can prevent duplicated data loading of reference data and eliminate redundancy of I/O operations for loading of the same reference data required by different virtual machines present in the same physical node to reduce use memory and I/O through sharing virtual machine leveled memories, are provided.
    Type: Application
    Filed: June 9, 2015
    Publication date: December 10, 2015
    Inventors: Kwang-Won KOH, Kang-Ho KIM, Seung-Hyub JEON, Seung-Jo BAE
  • Patent number: 9189165
    Abstract: A method for memory management, include allocating an empty page of a physical memory for reference data according to execution of an application program, and mapping the empty page to a virtual memory; checking a physical address of the physical memory to which the reference data has been loaded; mapping the checked physical address to the virtual memory to which the empty page has been mapped, and mapping the reference data; and releasing allocation of the allocated physical memory when the reference data is mapped to the virtual memory.
    Type: Grant
    Filed: July 15, 2013
    Date of Patent: November 17, 2015
    Assignee: ELECTRONICS AND TELECOMMUNICATIONS RESEARCH INSTITUTE
    Inventors: Kwang-Won Koh, Kang Ho Kim, Seung Hyub Jeon, Seungjo Bae
  • Patent number: 9158562
    Abstract: Disclosed herein is a method and apparatus for supporting virtualization. In the method, conversion of source code of a loadable module is initiated. A virtualization-sensitive instruction is searched for during the conversion of the source code. If the virtualization-sensitive instruction has been found, a virtualization-sensitive instruction table is generated based on the found virtualization-sensitive instruction. The virtualization-sensitive instruction is substituted with an instruction recognizable in a privileged mode, based on the generated virtualization-sensitive instruction table. The loadable module is loaded and executed in a kernel. Accordingly, the present invention supports virtualization, thus minimizing overhead occurring in full virtualization, and guaranteeing the high performance provided by para-virtualization without modifying a source.
    Type: Grant
    Filed: March 15, 2013
    Date of Patent: October 13, 2015
    Assignee: ELECTRONICS AND TELECOMMUNICATIONS RESEARCH INSTITUTE
    Inventors: Seung-Hyub Jeon, Kwang-Won Koh, Kang-Ho Kim, Chei-Yol Kim, Chang-Won Ahn
  • Publication number: 20150254116
    Abstract: Disclosed herein are a data processing apparatus for pipeline execution acceleration and a method thereof. According to an exemplary embodiment of the present invention, the data processing apparatus for pipeline execution acceleration includes: a processor configured to sequentially execute a first application program and a second application program reading or writing a specific file; and a file system configured to complete a write for a file data for the specific file to a data block previously allocated from the first application program and provide the file data for the specific file to the second application program prior to executing a close call for the specific file from the first application program, when executing a read call for the specific file from the second application program.
    Type: Application
    Filed: February 9, 2015
    Publication date: September 10, 2015
    Applicant: Electronics and Telecommunications Research Institute
    Inventors: Kang-Ho KIM, Kwang-Won KOH, Seung-Hyub JEON
  • Publication number: 20140310497
    Abstract: A method for memory management, include allocating an empty page of a physical memory for reference data according to execution of an application program, and mapping the empty page to a virtual memory; checking a physical address of the physical memory to which the reference data has been loaded; mapping the checked physical address to the virtual memory to which the empty page has been mapped, and mapping the reference data; and releasing allocation of the allocated physical memory when the reference data is mapped to the virtual memory.
    Type: Application
    Filed: July 15, 2013
    Publication date: October 16, 2014
    Inventors: Kwang-Won KOH, Kang Ho Kim, Seung Hyub Jeon, Seungjo Bae
  • Patent number: 7873609
    Abstract: Provided is a contents distribution management system and method for supporting a plurality of global servers that provide the contents and managing the contents by applying different policies based on the global server and service type of the contents. The inventive system comprises a plurality of global servers for supplying contents to a plurality of local servers, each global server having a large capacity contents library, and the local servers for managing the contents provided from the global servers based on global servers and service types using contents tables, and providing a contents service in response to a contents streaming service demand from a last terminal using a local contents cache.
    Type: Grant
    Filed: December 20, 2005
    Date of Patent: January 18, 2011
    Assignee: Electronics and Telecommunications Research Institute
    Inventors: Chang-Soo Kim, Seung-Jo Bae, Jin-Mee Kim, Yu-Hyeon Bak, Sang-Min Woo, Seung-Hyub Jeon, Won-Jae Lee, Hag-Young Kim
  • Patent number: 7584292
    Abstract: A hierarchical system configuration method and an integrated scheduling method for offering a multimedia streaming service on a two-level double cluster system are provided.
    Type: Grant
    Filed: November 30, 2005
    Date of Patent: September 1, 2009
    Assignee: Electronics and Telecommunications Research Institute
    Inventors: Seung Jo Bae, Jin Mee Kim, Seung Hyub Jeon, Hag Young Kim, Myung Joon Kim
  • Publication number: 20060155779
    Abstract: Provided is a contents distribution management system and method for supporting a plurality of global servers that provide the contents and managing the contents by applying different policies based on the global server and service type of the contents. The inventive system comprises a plurality of global servers for supplying contents to a plurality of local servers, each global server having a large capacity contents library, and the local servers for managing the contents provided from the global servers based on global servers and service types using contents tables, and providing a contents service in response to a contents streaming service demand from a last terminal using a local contents cache.
    Type: Application
    Filed: December 20, 2005
    Publication date: July 13, 2006
    Inventors: Chang-Soo Kim, Seung-Jo Bae, Jin-Mee Kim, Yu-Hyeon Bak, Sang-Min Woo, Seung-Hyub Jeon, Won-Jae Lee, Hag-Young Kim