Patents by Inventor Xiaobo Sharon Hu

Xiaobo Sharon Hu has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20230379140
    Abstract: A secure system configured to perform operations on data can include a plurality of N look-up-table fabric modules each operatively coupled to a respective column of M rows of data included in an input state matrix having N columns and M rows and a processor circuit operatively coupled to the plurality of N look-up-table fabric modules, the processor circuit configured to provide unencrypted data to the plurality of N look-up-table fabric modules for encryption and configured to provide encrypted data to the plurality of N look-up-table fabric modules for decryption.
    Type: Application
    Filed: May 15, 2023
    Publication date: November 23, 2023
    Inventors: Xiaobo Sharon Hu, Dayane Alfenas Reis, Michael Niemier, Haoran Geng
  • Patent number: 11599798
    Abstract: A method operating a Graphics Processing Unit (GPU) memory can be provided by accessing specified training parameters used to train a Deep Neural Network (DNN) using a GPU with a local GPU memory, the specified training parameters including at least a specified batch size of samples configured to train the DNN. A sub-batch size of the samples can be defined that is less than or equal to the specified batch size of samples in response to determining that an available size of the local GPU memory is insufficient to store all data associated with training the DNN using one batch of the samples. Instructions configured to train the DNN using the sub-batch size can be defined so that an accuracy of the DNN trained using the sub-batch size is about equal to an accuracy of the DNN trained using the specified batch size of the samples.
    Type: Grant
    Filed: March 16, 2020
    Date of Patent: March 7, 2023
    Assignee: UNIVERSITY OF NOTRE DAME DU LAC
    Inventors: Xiaobo Sharon Hu, Danny Ziyi Chen, Xiaoming Chen
  • Publication number: 20200302304
    Abstract: A method operating a Graphics Processing Unit (GPU) memory can be provided by accessing specified training parameters used to train a Deep Neural Network (DNN) using a GPU with a local GPU memory, the specified training parameters including at least a specified batch size of samples configured to train the DNN. A sub-batch size of the samples can be defined that is less than or equal to the specified batch size of samples in response to determining that an available size of the local GPU memory is insufficient to store all data associated with training the DNN using one batch of the samples. Instructions configured to train the DNN using the sub-batch size can be defined so that an accuracy of the DNN trained using the sub-batch size is about equal to an accuracy of the DNN trained using the specified batch size of the samples.
    Type: Application
    Filed: March 16, 2020
    Publication date: September 24, 2020
    Inventors: Xiaobo Sharon Hu, Danny Ziyi Chen, Xiaoming Chen
  • Patent number: 9825132
    Abstract: An electrical circuit is disclosed that comprises plurality of tunneling field-effect transistors (TFETs) arranged in a diffusion network matrix having a plurality of nodes wherein, for each of the TFETs that is not on an end of the matrix, a drain of the TFET is electrically coupled with the source of at least one of the other TFETs at a node of the matrix and a source of the TFET is electrically coupled with the drain of at least one of the other TFETs at another node of the matrix. The electrical circuit further comprises a plurality of capacitors, wherein a respective one of the plurality of capacitors is electrically coupled with each node that includes the source of at least one TFET and the drain of at least one TFET. The TFETs may be symmetrical graphene-insulator-graphene field-effect transistors (SymFETs), for example.
    Type: Grant
    Filed: October 13, 2015
    Date of Patent: November 21, 2017
    Assignee: University of Notre Dame du Lac
    Inventors: Behnam Sedighi, Xiaobo Sharon Hu, Michael Niemier, Joseph Nahas
  • Patent number: 9712146
    Abstract: Various processor architectures for mixed signal computation exploit the unique characteristics of advanced CMOS technologies, such as fin-based, multi-gate field effect transistors, and/or emerging technologies such as tunnel field effect transistors (TFETs). The example processors disclosed herein are cellular neural network (CNN)-inspired and eliminate the need for voltage controlled current sources (VCCSs), which have previously been utilized to realize feedback and feed-forward templates in CNNs and are the dominant source of power consumption in a CNN array. The example processors replace VCCSs with comparators, which can be efficiently realized with TFETs given their high intrinsic gain. Power efficiencies are in the order of 10,000 giga-operations per second per Watt (GOPS/W), which represents an improvement of more than ten times over state-of-the-art architectures seeking to accomplish similar information processing tasks.
    Type: Grant
    Filed: September 18, 2015
    Date of Patent: July 18, 2017
    Assignee: University of Notre Dame du Lac
    Inventors: Behnam Sedighi, Michael Niemier, Xiaobo Sharon Hu, Indranil Palit
  • Publication number: 20170103979
    Abstract: An electrical circuit is disclosed that comprises plurality of tunneling field-effect transistors (TFETs) arranged in a diffusion network matrix having a plurality of nodes wherein, for each of the TFETs that is not on an end of the matrix, a drain of the TFET is electrically coupled with the source of at least one of the other TFETs at a node of the matrix and a source of the TFET is electrically coupled with the drain of at least one of the other TFETs at another node of the matrix. The electrical circuit further comprises a plurality of capacitors, wherein a respective one of the plurality of capacitors is electrically coupled with each node that includes the source of at least one TFET and the drain of at least one TFET. The TFETs may be symmetrical graphene-insulator-graphene field-effect transistors (SymFETs), for example.
    Type: Application
    Filed: October 13, 2015
    Publication date: April 13, 2017
    Inventors: Behnam Sedighi, Xiaobo Sharon Hu, Michael Niemier, Joseph Nahas
  • Patent number: 8494115
    Abstract: Disclosed is an example method to calculate radiation dose. The method includes receiving a tissue matrix in which the tissue matrix includes a plurality of voxels. The example method also includes producing a first plurality of transport lines with a direction controller in which each transport line is indicative of a cone of irradiated energy, and calculating at least one radiation dose with at least one deposit engine substantially in parallel with producing a second plurality of transport lines with the direction controller.
    Type: Grant
    Filed: March 13, 2007
    Date of Patent: July 23, 2013
    Assignees: The University of Notre Dame du Lac, The University of Maryland, Baltimore
    Inventors: Xiaobo Sharon Hu, Cedric Xinsheng Yu, Bo Zhou, Danny Ziyi Chen, Kevin Whitton
  • Publication number: 20130158879
    Abstract: Disclosed is an example method to calculate radiation dose. The method includes receiving a tissue matrix in which the tissue matrix includes a plurality of voxels. The example method also includes producing a first plurality of transport lines with a direction controller in which each transport line is indicative of a cone of irradiated energy, and calculating at least one radiation dose with at least one deposit engine substantially in parallel with producing a second plurality of transport lines with the direction controller.
    Type: Application
    Filed: March 13, 2007
    Publication date: June 20, 2013
    Inventors: Xiaobo Sharon Hu, Cedric Xinsheng Yu, Bo Zhou, Danny Ziyi Chen, Kevin Whitton
  • Patent number: 8058906
    Abstract: A non-majority magnetic logic gate device for use in constructing compact and power efficient logical magnetic arrays is presented. The non-majority magnetic logic gate device includes a substrate, symmetrically aligned magnetic islands (SAMIs), at least one misaligned magnetic island (MAMI), magnetic field inputs (MFIs), and at least one magnetic field output (MFO). The SAMIs and MAMI are electrically isolated from each other but are magnetically coupled to one another through their respective magnetic fringe fields. The MAMI is geometrically and/or angularly configured to exhibit a magnetization ground state bias which is dependent upon which direction the applied magnetic clock field is swept. Non-majority logic gates can be made from layouts containing the SAMIs and the MAMI which contain a smaller number of components as comparable majority logic gate layouts.
    Type: Grant
    Filed: June 16, 2010
    Date of Patent: November 15, 2011
    Assignee: The University of Notre Dame Du Lac
    Inventors: Michael T. Niemier, Mohammad T. Alam, Gary H. Bernstein, Xiaobo Sharon Hu, Wolfgang Porod, Edit Varga
  • Publication number: 20100315123
    Abstract: A non-majority magnetic logic gate device for use in constructing compact and power efficient logical magnetic arrays is presented. The non-majority magnetic logic gate device includes a substrate, symmetrically aligned magnetic islands (SAMIs), at least one misaligned magnetic island (MAMI), magnetic field inputs (MFIs), and at least one magnetic field output (MFO). The SAMIs and MAMI are electrically isolated from each other but are magnetically coupled to one another through their respective magnetic fringe fields. The MAMI is geometrically and/or angularly configured to exhibit a magnetization ground state bias which is dependent upon which direction the applied magnetic clock field is swept. Non-majority logic gates can be made from layouts containing the SAMIs and the MAMI which contain a smaller number of components as comparable majority logic gate layouts.
    Type: Application
    Filed: June 16, 2010
    Publication date: December 16, 2010
    Inventors: Michael T. Niemier, Mohammad T. Alam, Gary H. Bernstein, Xiaobo Sharon Hu, Wolfgang Porod, Edit Varga