Patents by Inventor Jen Cheng Huang

Jen Cheng Huang has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20240136401
    Abstract: The present disclosure relates to an integrated chip. The integrated chip includes a substrate having a first semiconductor material. A second semiconductor material is disposed on the first semiconductor material and a passivation layer is disposed on the second semiconductor material. A first doped region and a second doped region extend through the passivation layer and into the second semiconductor material. A silicide is arranged within the passivation layer and along tops of the first doped region and the second doped region.
    Type: Application
    Filed: January 5, 2024
    Publication date: April 25, 2024
    Inventors: Yin-Kai Liao, Sin-Yi Jiang, Hsiang-Lin Chen, Yi-Shin Chu, Po-Chun Liu, Kuan-Chieh Huang, Jyh-Ming Hung, Jen-Cheng Liu
  • Patent number: 11915977
    Abstract: A stacked integrated circuit (IC) device and a method are disclosed. The stacked IC device includes a first semiconductor element. The first substrate includes a dielectric block in the first substrate; and a plurality of first conductive features formed in first inter-metal dielectric layers over the first substrate. The stacked IC device also includes a second semiconductor element bonded on the first semiconductor element. The second semiconductor element includes a second substrate and a plurality of second conductive features formed in second inter-metal dielectric layers over the second substrate. The stacked IC device also includes a conductive deep-interconnection-plug coupled between the first conductive features and the second conductive features. The conductive deep-interconnection-plug is isolated by dielectric block, the first inter-metal-dielectric layers and the second inter-metal-dielectric layers.
    Type: Grant
    Filed: April 12, 2021
    Date of Patent: February 27, 2024
    Assignee: TAIWAN SEMICONDUCTOR MANUFACTURING COMPANY, LTD.
    Inventors: Shu-Ting Tsai, Jeng-Shyan Lin, Dun-Nian Yaung, Jen-Cheng Liu, Feng-Chi Hung, Chih-Hui Huang, Sheng-Chau Chen, Shih Pei Chou, Chia-Chieh Lin
  • Patent number: 11386058
    Abstract: Provided herein are data cloud administration techniques that achieve autonomy by using a rules engine that reacts to a database system event by autonomously submitting an asynchronous job to reconfigure a database. In an embodiment, a rules engine receives an event from a DBMS. Based on the event, the rules engine executes a rule to generate a request that indicates configuration details for a database. The rules engine sends the request to a request broker. The request broker dispatches an asynchronous job based on the request. The asynchronous job configures the database based on the configuration details. Thus, databases in a cloud, data grid, or data center may be administered autonomously (without human intervention) base on dynamic conditions that are foreseen and unforeseen.
    Type: Grant
    Filed: September 19, 2018
    Date of Patent: July 12, 2022
    Assignee: Oracle International Corporation
    Inventors: Hochak Hung, Kumar Rajamani, Jaebock Lee, Sanket Jain, Giridhar Ravipati, John McHugh, Churn Don Lee, Alexander Chen, Jen-Cheng Huang
  • Patent number: 11327932
    Abstract: Provided herein are workload management techniques that asynchronously configure pluggable databases within a compute cloud. In an embodiment, the compute cloud receives an administrative request that indicates configuration details for a pluggable database. The compute cloud generates a configuration descriptor that specifies an asynchronous job based on the configuration details of the request. The compute cloud accesses hosting metadata to detect at least one of: a) a current container database that already hosts the pluggable database, b) a target container database that will host the pluggable database, or c) a particular computer that hosts at least one of: the current container database, or the target container database. The compute cloud executes the asynchronous job to configure the pluggable database based on at least one of: the hosting metadata, or the configuration descriptor.
    Type: Grant
    Filed: August 6, 2018
    Date of Patent: May 10, 2022
    Assignee: Oracle International Corporation
    Inventors: Hochak Hung, Kumar Rajamani, Sanket Jain, Giridhar Ravipati, Jaebock Lee, Alexander Chen, Jen-Cheng Huang
  • Publication number: 20190102384
    Abstract: Provided herein are workload management techniques that asynchronously configure pluggable databases within a compute cloud. In an embodiment, the compute cloud receives an administrative request that indicates configuration details for a pluggable database. The compute cloud generates a configuration descriptor that specifies an asynchronous job based on the configuration details of the request. The compute cloud accesses hosting metadata to detect at least one of: a) a current container database that already hosts the pluggable database, b) a target container database that will host the pluggable database, or c) a particular computer that hosts at least one of: the current container database, or the target container database. The compute cloud executes the asynchronous job to configure the pluggable database based on at least one of: the hosting metadata, or the configuration descriptor.
    Type: Application
    Filed: August 6, 2018
    Publication date: April 4, 2019
    Inventors: Hochak Hung, Kumar Rajamani, Sanket Jain, Giridhar Ravipati, Jaebock Lee, Alexander Chen, Jen-Cheng Huang
  • Publication number: 20190102411
    Abstract: Provided herein are data cloud administration techniques that achieve autonomy by using a rules engine that reacts to a database system event by autonomously submitting an asynchronous job to reconfigure a database. In an embodiment, a rules engine receives an event from a DBMS. Based on the event, the rules engine executes a rule to generate a request that indicates configuration details for a database. The rules engine sends the request to a request broker. The request broker dispatches an asynchronous job based on the request. The asynchronous job configures the database based on the configuration details. Thus, databases in a cloud, data grid, or data center may be administered autonomously (without human intervention) base on dynamic conditions that are foreseen and unforeseen.
    Type: Application
    Filed: September 19, 2018
    Publication date: April 4, 2019
    Inventors: Hochak Hung, Kumar Rajamani, Jaebock Lee, Sanket Jain, Giridhar Ravipati, John McHugh, Churn Don Lee, Alexander Chen, Jen-Cheng Huang
  • Patent number: 8396953
    Abstract: In a method for processing packets among at least a first computing device and a second computing device, in which the first computing device is configured to transmit and receive packets through a Network Interface Card (NIC), in the second computing device, descriptors of packets to be one of transmitted and received by the first computing device through a device descriptor queue are received and placed in a virtualized descriptor queue accessible by the second computing device. In addition, the packets associated with the descriptors placed in the virtualized descriptor queue are processed prior to one of transmission and receipt of the packets by the first computing device.
    Type: Grant
    Filed: March 12, 2010
    Date of Patent: March 12, 2013
    Assignee: Hewlett-Packard Development Company, L.P.
    Inventors: Matteo Monchiero, Jen Cheng Huang, Yoshio Turner
  • Publication number: 20110225278
    Abstract: In a method for processing packets among at least a first computing device and a second computing device, in which the first computing device is configured to transmit and receive packets through a Network Interface Card (NIC), in the second computing device, descriptors of packets to be one of transmitted and received by the first computing device through a device descriptor queue are received and placed in a virtualized descriptor queue accessible by the second computing device. In addition, the packets associated with the descriptors placed in the virtualized descriptor queue are processed prior to one of transmission and receipt of the packets by the first computing device.
    Type: Application
    Filed: March 12, 2010
    Publication date: September 15, 2011
    Inventors: Matteo Monchiero, Jen Cheng Huang, Yoshio Turner