Patents by Inventor Yi Ge

Yi Ge has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20160335094
    Abstract: A method for controlling a pipeline-based processor. The method includes determining a change in a workload. The method also includes activating or shutting down, by at least one controller circuit, one or more of the plurality of enhanced pipeline stages based on at least one corresponding enhanced stage priority level of the one or more of the plurality of enhanced pipeline stages and requirements for performance of the workload. The method additionally includes activating or shutting down, by the at least one controller circuit, one or more of the plurality of enhanced modules of the particular pipeline stage based on at least one corresponding priority level of the one or more of the plurality of enhanced modules of the particular pipeline stage and the requirements for the performance of the workload.
    Type: Application
    Filed: July 29, 2016
    Publication date: November 17, 2016
    Inventors: Wen Bo Shen, Peng Shao, Yu Li, Xiao Tao Chang, Yi Ge, Hua Yong Wang, Huan Hao Zou
  • Publication number: 20160266918
    Abstract: A data assignment method for a physical machine having multiple virtual machines (VMs) running thereon includes receiving data to be processed from a scheduler for assigning job data; storing the received data to be processed into a single queue; and in response to an idle VM being among the multiple VMs, assigning data in the queue to the idle VM to be processed by the idle VM.
    Type: Application
    Filed: May 23, 2016
    Publication date: September 15, 2016
    Inventors: Yi Ge, Li Li, Liang Liu, Ju Wei Shi
  • Patent number: 9430217
    Abstract: A method for patching an application running in a computing system , the method comprising: in response to that there is a need to patch a first content and the first content has been in the memory, distinguishing between a new content and an old content, the new content being the patched first content, the old content being the first content that has been in the memory; and in response to that the new content is loaded to the memory, mapping to the new content a new process that needs to apply the first content, wherein the new process comprises a process that is started after loading the new content to the memory. An apparatus for patching an application is further disclosed. With the technical solution provided by the present invention, it is possible to perform dynamic patching to a virtual machine or a physical machine without stopping a running process.
    Type: Grant
    Filed: May 23, 2013
    Date of Patent: August 30, 2016
    Assignee: International Business Machines Corporation
    Inventors: Chun Hai Chen, Yi Ge, Li Li, Liang Liu, Jun Mei Qu
  • Patent number: 9411625
    Abstract: An apparatus for a hypervisor to obtain a faulting instruction, wherein the hypervisor runs between a physical machine including a central processing unit (CPU) and a virtual machine includes a content addressable memory (CAM); a special-purpose register (SPR) which is accessible by the hypervisor; and a control logic circuit with an input terminal connected to the CPU and an output terminal connected to the CAM, the input terminal receiving data from an instruction fetching (IF) stage and a write-back (WB) stage of a CPU instruction pipeline respectively, the output terminal causing instructions from the IF stage of the CPU instruction pipeline to be stored into the CAM and triggering the CAM to output a faulting instruction among the instructions stored therein to the SPR.
    Type: Grant
    Filed: December 13, 2013
    Date of Patent: August 9, 2016
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Xiao Tao Chang, Yi Ge, Hao Li, Tao Liu, Kun Wang
  • Patent number: 9396014
    Abstract: The present invention provides a method and apparatus for data swap in a virtual machine environment. The present invention provides a method for data swap in a virtual machine environment, including: in response to a swap request from a virtual machine, looking up storage space associated with the swap request; and allocating to the virtual machine free physical storage space, in a host, which matches the storage space, so that the free physical storage space logically becomes available storage space to the virtual machine; the virtual machine is running on the host, and the storage space is physical storage space in the host.
    Type: Grant
    Filed: February 27, 2014
    Date of Patent: July 19, 2016
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Chun Hai Chen, Yi Ge, Li Li, Liang Liu, Junmei Qu
  • Publication number: 20160205210
    Abstract: The present invention provides a method and apparatus of controlling subscription requests, which can be used in a publish/subscribe engine. The method includes parsing a received subscription request into a predicate expression; determining an occurrence probability of the subscription request based on occurrence probabilities of predicates in the predicate expression; estimating a message increment to be caused by the subscription request based on a message publishing rate associated with the subscription request and on the occurrence probability of the subscription request; and controlling a quality of service of messages based on the estimated message increment.
    Type: Application
    Filed: August 21, 2014
    Publication date: July 14, 2016
    Inventors: Yi Ge, Li Li, Ju Wei Shi, Qi Yu, Guotao Zhao
  • Patent number: 9372886
    Abstract: This invention relates to the Internet of Things (IOT), and discloses a method and apparatus of data filtering in the IOT, where the IOT includes a plurality of sensor devices. The method includes: inputting an application deployed rule; converting the rule into at least one sub-predicate expression having static predicates and dynamic predicates; inputting data collected by the sensor devices; matching the collected data with the static predicates and the dynamic predicates of the sub-predicate expression in sequence; and distributing matched data to the application. In this invention, a rule is divided into static and dynamic predicates, and a match is performed on static predicates before dynamic predicates, so that the speed of predicate match can be improved and fast and efficient data filtering in the IOT can be achieved.
    Type: Grant
    Filed: April 24, 2013
    Date of Patent: June 21, 2016
    Assignee: International Business Machines Corporation
    Inventors: Zhen Tan Feng, Yi Ge, Chi Liu, Wei Lu, Bo Yang, Qi Yu
  • Patent number: 9355042
    Abstract: Method and apparatus for managing a translation lookaside buffer (TLB) at hardware in a virtualization enabled system. According to embodiments of the present invention, a series of operations caused by TLB miss would not need intervening from the hypervisor. On the contrary, when a TLB miss occurs, the hardware directly issues an interrupt to a virtual machine. In this way, the TLB can be efficiently managed by means of a hardware-level auxiliary translation table. Therefore, system overheads can be greatly reduced and system performance can be improved. Methods and apparatuses associated with hardware, hypervisor, and virtual machine in a virtualization enabled system are disclosed, respectively.
    Type: Grant
    Filed: June 17, 2014
    Date of Patent: May 31, 2016
    Assignee: GLOBALFOUNDRIES INC.
    Inventors: Xiao Tao Chang, Hubertus Franke, Yi Ge, Kun Wang
  • Publication number: 20160147288
    Abstract: A pipeline-based processor and method. The method includes partitioning a particular pipeline into one or more base pipeline stages and a plurality of enhanced pipeline stages, each enhanced pipeline stage configured to be either a shutdown enhanced pipeline stage or an activated enhanced pipeline stage. Each enhanced pipeline stage has an enhanced stage priority level. The method also includes configuring each enhanced pipeline stage to be activated or shut down based at least on the enhanced stage priority level. The method additionally includes partitioning a particular pipeline stage into at least one base module and a plurality of enhanced modules, each enhanced pipeline stage configured to be either a shutdown enhanced pipeline stage or an activated enhanced pipeline stage. Each enhanced module has a particular priority level. The method further includes configuring each enhanced module to be activated or shut down based at least on the particular priority level.
    Type: Application
    Filed: November 20, 2015
    Publication date: May 26, 2016
    Inventors: Wen Bo Shen, Peng Shao, Yu Li, Xiao Tao Chang, Yi Ge, Hua Yong Wang, Huan Hao Zou
  • Patent number: 9348406
    Abstract: The present invention discloses a method of processing instructions in a pipeline-based central processing unit, wherein the pipeline is partitioned into base pipeline stages and enhanced pipeline stages according to functions, the base pipeline stages being activated all the while, and the enhanced pipeline stages being activated or shutdown according to requirements for performance of a workload. The present invention further discloses a method of processing instructions in a pipeline-based central processing unit, wherein the pipeline is partitioned into base pipeline stages and enhanced pipeline stages according to functions, each pipeline stage being partitioned into a base module and at least one enhanced module, the base module being activated all the while, and the enhanced module being activated or shutdown according to requirements for performance of a workload.
    Type: Grant
    Filed: April 26, 2012
    Date of Patent: May 24, 2016
    Assignee: International Business Machines Corporation
    Inventors: Wen Bo Shen, Peng Shao, Yu Li, Xiao Tao Chang, Yi Ge, Hua Yong Wang, Huan Hao Zou
  • Patent number: 9348632
    Abstract: A data assignment method for a physical machine having multiple virtual machines (VMs) running thereon includes receiving data to be processed from a scheduler for assigning job data; storing the received data to be processed into a single queue; and in response to an idle VM being among the multiple VMs, assigning data in the queue to the idle VM to be processed by the idle VM.
    Type: Grant
    Filed: September 19, 2014
    Date of Patent: May 24, 2016
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Yi Ge, Li Li, Liang Liu, Ju Wei Shi
  • Publication number: 20160065657
    Abstract: A method and system for processing a message and subscription information is provided. Subscription information is obtained and parsed into a first set of predicates, whereby these predicates are not repetitive. One or more tree structures for the predicates and the subscription information are constructed based on possibilities of the predicates in hitting a message. The possibility of the predicate for a lower internal node in the tree structure hitting the message is higher than that of an upper internal node. Each of the subscription information corresponds to one leaf node in one tree structure, and each of the subscription information corresponds to all information of the predicates in a branch of the leaf node in the tree structure thereof from the root node. The one or more tree structures for matching with the message are stored.
    Type: Application
    Filed: August 26, 2015
    Publication date: March 3, 2016
    Inventors: Yi Ge, Qi Yu, Guotao Zhao, Yu Chen Zhou
  • Publication number: 20150379024
    Abstract: A method and system for processing file storage is provided. The method includes receiving a file storage request and determining a file size of a file which is requested to be stored. In response to determining that the file is a small file, a request is sent to an HBase to store the file into the HBase. In response to determining that the file is a large file, a request is sent to a HDFS to store the file into the HDFS. Information associated with the file is stored in a record.
    Type: Application
    Filed: June 2, 2015
    Publication date: December 31, 2015
    Inventors: Yi Ge, Liang Liu, Junmei Qu, Wen Jun Yin, Wei Zhuang
  • Patent number: 9152444
    Abstract: A method used by a virtual machine in communication with an external machine includes providing a single sharing page that is shared between a plurality of virtual machines and a particular virtual machine, wherein the particular virtual machine and the plurality of virtual machines run on a same physical machine; writing into the single sharing page a data packet to be sent by the virtual machine to the external machine; scheduling a page swap between the single sharing page and a blank memory page of the particular virtual machine; and sending, to the external machine, the data packet in the memory page of the particular virtual machine subsequent to the page swap.
    Type: Grant
    Filed: January 7, 2013
    Date of Patent: October 6, 2015
    Assignee: International Business Machines Corporation
    Inventors: Yi Ge, Li Li, Liang Liu, Jun Mei Qu, Yu Yuan
  • Publication number: 20150242231
    Abstract: The present invention provides a method and apparatus for data swap in a virtual machine environment. The present invention provides a method for data swap in a virtual machine environment, including: in response to a swap request from a virtual machine, looking up storage space associated with the swap request; and allocating to the virtual machine free physical storage space, in a host, which matches the storage space, so that the free physical storage space logically becomes available storage space to the virtual machine; the virtual machine is running on the host, and the storage space is physical storage space in the host.
    Type: Application
    Filed: February 27, 2014
    Publication date: August 27, 2015
    Applicant: International Business Machines Corporation
    Inventors: Chun Hai Chen, Yi Ge, Li Li, Liang Liu, Junmei Qu
  • Publication number: 20150220320
    Abstract: The present invention belongs to the computer field and discloses a method for patching, the method comprising: in response to that there is a need to patch a first content and the first content has been in the memory, distinguishing between a new content and an old content, the new content being the patched first content, the old content being the first content that has been in the memory; and in response to that the new content is loaded to the memory, mapping to the new content a new process that needs to apply the first content, wherein the new process comprises a process that is started after loading the new content to the memory. The present invention further discloses an apparatus for patching. With the technical solution provided by the present invention, it is possible to perform dynamic patching to a virtual machine or a physical machine without stopping a running process.
    Type: Application
    Filed: May 23, 2013
    Publication date: August 6, 2015
    Applicant: International Business Machines Corporation
    Inventors: Chun Hai Chen, Yi Ge, Li Li, Liang Liu, Jun Mei Qu
  • Patent number: 9086980
    Abstract: The present disclosure provides a method, device, and system for processing a request in a multi-core system. The method comprises steps of: receiving a request for data by a filter from a requesting unit; comparing an indicator indicative of a logical partition in the request with an indicator indicative of the logical partition in a record of the filter; searching in a unit where the filter is located based on the request and returning a search result to the requesting unit if a comparison result matches; and returning a NONE response to the requesting unit from the filter if the comparison result does not match.
    Type: Grant
    Filed: August 1, 2012
    Date of Patent: July 21, 2015
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Yi Ge, Rui Hou, Kun Wang, Hong Bo Zeng, Yu Zhang
  • Patent number: 9069716
    Abstract: A matrix calculation unit may include a matrix operation unit and a converting unit. The matrix operation unit may include functions to perform a matrix operation of a first size with respect to data stored in a memory, and to perform a matrix operation of a second size with respect to the data stored in the memory, where the second size is enlarged from the first size. The converting unit may convert in at least one direction in the memory between a data array suited for the matrix operation of the first size and a data array suited for the matrix operation of the second size.
    Type: Grant
    Filed: February 27, 2013
    Date of Patent: June 30, 2015
    Assignees: FUJITSU LIMITED, SOCIONEXT INC.
    Inventors: Yi Ge, Hiroshi Hatano, Kazuo Horio
  • Publication number: 20150169348
    Abstract: An apparatus for a hypervisor to obtain a faulting instruction, wherein the hypervisor runs between a physical machine including a central processing unit (CPU) and a virtual machine includes a content addressable memory (CAM); a special-purpose register (SPR) which is accessible by the hypervisor; and a control logic circuit with an input terminal connected to the CPU and an output terminal connected to the CAM, the input terminal receiving data from an instruction fetching (IF) stage and a write-back (WB) stage of a CPU instruction pipeline respectively, the output terminal causing instructions from the IF stage of the CPU instruction pipeline to be stored into the CAM and triggering the CAM to output a faulting instruction among the instructions stored therein to the SPR.
    Type: Application
    Filed: December 13, 2013
    Publication date: June 18, 2015
    Applicant: International Business Machines Corporation
    Inventors: Xiao Tao Chang, Yi Ge, Hao Li, Tao Liu, Kun Wang
  • Publication number: 20150095910
    Abstract: A data assignment method for a physical machine having multiple virtual machines (VMs) running thereon includes receiving data to be processed from a scheduler for assigning job data; storing the received data to be processed into a single queue; and in response to an idle VM being among the multiple VMs, assigning data in the queue to the idle VM to be processed by the idle VM.
    Type: Application
    Filed: September 19, 2014
    Publication date: April 2, 2015
    Inventors: Yi Ge, Li Li, Liang Liu, Ju Wei Shi