Patents Examined by Johnathan R Germick
  • Patent number: 12354024
    Abstract: An intelligent table-based question answering system receives an input user query requesting information from an input table to generate a response to the input query. Two types of user queries including textual user queries and scalar user queries that require at least one mathematical operation to be executed can be processed for answer generation. The input table is processed to generate a paragraph of table text which includes row and column information extracted from the input table. The input query is provided along with the paragraph of table text to a deep Quans model which outputs a candidate answer that forms a textual portion of the response. Also, the candidate answer is reverse mapped to the input table and a portion of the input table including the candidate answer is provided as a tabular portion of the response.
    Type: Grant
    Filed: August 10, 2020
    Date of Patent: July 8, 2025
    Assignee: ACCENTURE GLOBAL SOLUTIONS LIMITED
    Inventors: Eldhose Joy, Sundar Prasad Jayaraman, Harsha Jawagal, Vishal Radhesham Supekar
  • Patent number: 12277490
    Abstract: Persistent storage contains a representation of a neural network including an input layer, and output layer, and a hidden layer, wherein nodes of the hidden layer incorporate serialized activation functions, wherein the serialized activation functions for each of the nodes include a sigmoid function and a Beta function, wherein the sigmoid function is applied to weighted outputs from nodes of a previous layer of the neural network, wherein the Beta function is applied to a conductance hyper-parameter and respective outputs of the sigmoid function, and wherein outputs of the Beta function are provided to a subsequent layer of the neural network. One or more processors are configured to train the neural network until the outputs of the sigmoid function for the nodes of the hidden layer are substantially binary.
    Type: Grant
    Filed: April 14, 2020
    Date of Patent: April 15, 2025
    Assignee: Google LLC
    Inventors: Ryan McDonald, Luis Sarmento
  • Patent number: 12229675
    Abstract: Generally, the present disclosure is directed to systems and methods that perform adaptive optimization with improved convergence properties. The adaptive optimization techniques described herein are useful in various optimization scenarios, including, for example, training a machine-learned model such as, for example, a neural network. In particular, according to one aspect of the present disclosure, a system implementing the adaptive optimization technique can, over a plurality of iterations, employ an adaptive learning rate while also ensuring that the learning rate is non-increasing.
    Type: Grant
    Filed: December 14, 2022
    Date of Patent: February 18, 2025
    Assignee: GOOGLE LLC
    Inventors: Sashank Jakkam Reddi, Sanjiv Kumar, Satyen Chandrakant Kale
  • Patent number: 12182719
    Abstract: A method of operating a neural network. The input layer of the network may have n input nodes connected to output nodes via a hidden layer. The hidden layer may include m hidden nodes. The n input nodes may connect to a subset of k nodes of the m hidden nodes via respective synaptic connections, to which training weights are associated, which form an n×k input matrix Win, whereas a subset of m?k nodes of the hidden layer are not connected by any node of the input layer. Running the network may include performing a first matrix vector multiplication between the input matrix Win and a vector of values obtained in output of the input nodes and a second matrix vector multiplication between a fixed matrix Wrec of fixed weights and a vector of values obtained in output of the m nodes of the hidden layer.
    Type: Grant
    Filed: July 28, 2020
    Date of Patent: December 31, 2024
    Assignee: International Business Machines Corporation
    Inventors: Lorenz K. Muller, Pascal Stark, Stefan Abel
  • Patent number: 12131260
    Abstract: During training of deep neural networks, a Copernican loss (LC) is designed to augment a primary loss function, for example, a standard Softmax loss, to explicitly minimize intra-class variation and simultaneously maximize inter-class variation. Copernican loss operates using the cosine distance and thereby affects angles leading to a cosine embedding, which removes the disconnect between training and testing.
    Type: Grant
    Filed: April 10, 2023
    Date of Patent: October 29, 2024
    Assignee: Carnegie Mellon University
    Inventors: Marios Savvides, Dipan Kumar Pal
  • Patent number: 12106215
    Abstract: Knowledge transfer between recurrent neural networks is performed by obtaining a first output sequence from a bidirectional Recurrent Neural Network (RNN) model for an input sequence, obtaining a second output sequence from a unidirectional RNN model for the input sequence, selecting at least one first output from the first output sequence based on a similarity between the at least one first output and a second output from the second output sequence; and training the unidirectional RNN model to increase the similarity between the at least one first output and the second output.
    Type: Grant
    Filed: February 14, 2023
    Date of Patent: October 1, 2024
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Gakuto Kurata, Kartik Audhkhasi
  • Patent number: 12073321
    Abstract: Embodiments of this application disclose a method for training an image caption model, the image caption model including an encoding convolutional neural network (CNN) and a decoding recurrent neural network (RNN). The method includes: obtaining an image eigenvector of an image sample by using the encoding CNN; decoding the image eigenvector by using the decoding RNN, to obtain a sentence used for describing the image sample; determining a matching degree between the sentence obtained through decoding and the image sample and a smoothness degree of the sentence obtained through decoding, respectively; and adjusting the decoding RNN according to the matching degree and the smoothness degree.
    Type: Grant
    Filed: October 20, 2020
    Date of Patent: August 27, 2024
    Assignee: TENCENT TECHNOLOGY (SHENZHEN) COMPANY LIMITED
    Inventors: Yang Feng, Lin Ma, Wei Liu, Jiebo Luo
  • Patent number: 12050998
    Abstract: Systems and methods are provided for data shuffling for distributed machine learning training, including each training node in the network receiving a shard of training data, wherein the training data set is divided into shards having data items. Each data item is assigned to a working set such that each of the working set includes data items from multiple shards. The training nodes perform training using the data items of a first working set that are in each node's shard. Upon completion of the training using the data items of the first working set, the training nodes performing training using the data items of a second working set that are in their shards; and while the training nodes are performing training on their respective subsets of shards of the second working set, the nodes randomly shuffling data items in the first working set to create a shuffled first working set.
    Type: Grant
    Filed: September 2, 2020
    Date of Patent: July 30, 2024
    Assignee: Hewlett Packard Enterprise Development LP
    Inventors: Sergey Serebryakov, Cong Xu
  • Patent number: 11972327
    Abstract: A method for action automation includes determining, using an electronic device, an action based on domain information. Activity patterns associated with the action are retrieved. For each activity pattern, a candidate action rule is determined. Each candidate action rule specifies one or more pre-conditions when the action occurs. One or more preferred candidate action rules are determined from multiple candidate action rules for automation of the action.
    Type: Grant
    Filed: April 30, 2018
    Date of Patent: April 30, 2024
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Vijay Srinivasan, Christian Koehler, Hongxia Jin
  • Patent number: 11972408
    Abstract: A method may include embedding, in a hidden layer and/or an output layer of a first machine learning model, a first digital watermark. The first digital watermark may correspond to input samples altering the low probabilistic regions of an activation map associated with the hidden layer of the first machine learning model. Alternatively, the first digital watermark may correspond to input samples rarely encountered by the first machine learning model. The first digital watermark may be embedded in the first machine learning model by at least training, based on training data including the input samples, the first machine learning model. A second machine learning model may be determined to be a duplicate of the first machine learning model based on a comparison of the first digital watermark embedded in the first machine learning model and a second digital watermark extracted from the second machine learning model.
    Type: Grant
    Filed: March 21, 2019
    Date of Patent: April 30, 2024
    Assignee: The Regents of the University of California
    Inventors: Bita Darvish Rouhani, Huili Chen, Farinaz Koushanfar
  • Patent number: 11842264
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for a neural network system comprising one or more gated linear networks. A system includes: one or more gated linear networks, wherein each gated linear network corresponds to a respective data value in an output data sample and is configured to generate a network probability output that defines a probability distribution over possible values for the corresponding data value, wherein each gated linear network comprises a plurality of layers, wherein the plurality of layers comprises a plurality of gated linear layers, wherein each gated linear layer has one or more nodes, and wherein each node is configured to: receive a plurality of inputs, receive side information for the node; combine the plurality of inputs according to a set of weights defined by the side information, and generate and output a node probability output for the corresponding data value.
    Type: Grant
    Filed: November 30, 2018
    Date of Patent: December 12, 2023
    Assignee: DeepMind Technologies Limited
    Inventors: Agnieszka Grabska-Barwinska, Peter Toth, Christopher Mattern, Avishkar Bhoopchand, Tor Lattimore, Joel William Veness
  • Patent number: 11803756
    Abstract: A method of operating a neural network system includes parsing, by a processor, at least one item of information related to a neural network operation from an input neural network model; determining, by the processor, information of at least one dedicated hardware device; and generating, by the processor, a reshaped neural network model by changing information of the input neural network model according to a result of determining the information of the at least one dedicated hardware device such that the reshaped neural network model is tailored for execution by the dedicated hardware device.
    Type: Grant
    Filed: June 12, 2018
    Date of Patent: October 31, 2023
    Assignee: Samsung Electronics Co., Ltd.
    Inventor: Seung-soo Yang
  • Patent number: 11663483
    Abstract: According to embodiments, an encoder neural network receives a one-hot representation of a real text. The encoder neural network outputs a latent representation of the real text. A decoder neural network receives random noise data or artificial code generated by a generator neural network from random noise data. The decoder neural network outputs softmax representation of artificial text. The decoder neural network receives the latent representation of the real text. The decoder neural network outputs a reconstructed softmax representation of the real text. A hybrid discriminator neural network receives a first combination of the soft-text and the latent representation of the real text and a second combination of the softmax representation of artificial text and the artificial code. The hybrid discriminator neural network outputs a probability indicating whether the second combination is similar to the first combination. Additional embodiments for utilizing latent representation are also disclosed.
    Type: Grant
    Filed: October 30, 2018
    Date of Patent: May 30, 2023
    Assignee: Huawei Technologies Co., Ltd.
    Inventors: Md Akmal Haidar, Mehdi Rezagholizadeh
  • Patent number: 11636344
    Abstract: During training of deep neural networks, a Copernican loss (LC) is designed to augment the standard Softmax loss to explicitly minimize intra-class variation and simultaneously maximize inter-class variation. Copernican loss operates using the cosine distance and thereby affects angles leading to a cosine embedding, which removes the disconnect between training and testing.
    Type: Grant
    Filed: March 12, 2019
    Date of Patent: April 25, 2023
    Assignee: Carnegie Mellon University
    Inventors: Marios Savvides, Dipan Kumar Pal
  • Patent number: 11625595
    Abstract: Knowledge transfer between recurrent neural networks is performed by obtaining a first output sequence from a bidirectional Recurrent Neural Network (RNN) model for an input sequence, obtaining a second output sequence from a unidirectional RNN model for the input sequence, selecting at least one first output from the first output sequence based on a similarity between the at least one first output and a second output from the second output sequence; and training the unidirectional RNN model to increase the similarity between the at least one first output and the second output.
    Type: Grant
    Filed: August 29, 2018
    Date of Patent: April 11, 2023
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Gakuto Kurata, Kartik Audhkhasi
  • Patent number: 11604960
    Abstract: Machine learning is utilized to learn an optimized quantization configuration for an artificial neural network (ANN). For example, an ANN can be utilized to learn an optimal bit width for quantizing weights for layers of the ANN. The ANN can also be utilized to learn an optimal bit width for quantizing activation values for the layers of the ANN. Once the bit widths have been learned, they can be utilized at inference time to improve the performance of the ANN by quantizing the weights and activation values of the layers of the ANN.
    Type: Grant
    Filed: March 18, 2019
    Date of Patent: March 14, 2023
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Kalin Ovtcharov, Eric S. Chung, Vahideh Akhlaghi, Ritchie Zhao
  • Patent number: 11586904
    Abstract: Generally, the present disclosure is directed to systems and methods that perform adaptive optimization with improved convergence properties. The adaptive optimization techniques described herein are useful in various optimization scenarios, including, for example, training a machine-learned model such as, for example, a neural network. In particular, according to one aspect of the present disclosure, a system implementing the adaptive optimization technique can, over a plurality of iterations, employ an adaptive learning rate while also ensuring that the learning rate is non-increasing.
    Type: Grant
    Filed: September 13, 2018
    Date of Patent: February 21, 2023
    Assignee: GOOGLE LLC
    Inventors: Sashank Jakkam Reddi, Sanjiv Kumar, Satyen Chandrakant Kale
  • Patent number: 11568303
    Abstract: An electronic apparatus is provided. The electronic apparatus includes a first memory configured to store a first artificial intelligence (AI) model including a plurality of first elements and a processor configured to include a second memory. The second memory is configured to store a second AI model including a plurality of second elements. The processor is configured to acquire output data from input data based on the second AI model. The first AI model is trained through an AI algorithm. Each of the plurality of second elements includes at least one higher bit of a plurality of bits included in a respective one of the plurality of first elements.
    Type: Grant
    Filed: October 5, 2018
    Date of Patent: January 31, 2023
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Kyoung-hoon Kim, Young-hwan Park, Dong-soo Lee, Dae-hyun Kim, Han-su Cho, Hyun-jung Kim
  • Patent number: 11568266
    Abstract: Described herein are embodiments for systems and methods for mutual machine learning with global topic discovery and local word embedding. Both topic modeling and word embedding map documents onto a low-dimensional space, with the former clustering words into a global topic space and the latter mapping word into a local continuous embedding space. Embodiments of Topic Modeling and Sparse Autoencoder (TMSA) framework unify these two complementary patterns by constructing a mutual learning mechanism between word co-occurrence based topic modeling and autoencoder. In embodiments, word topics generated with topic modeling are passed into auto-encoder to impose topic sparsity for the autoencoder to learn topic-relevant word representations. In return, word embedding learned by autoencoder is sent back to topic modeling to improve the quality of topic generations. Performance evaluation on various datasets demonstrates the effectiveness of the disclosed TMSA framework in discovering topics and embedding words.
    Type: Grant
    Filed: March 15, 2019
    Date of Patent: January 31, 2023
    Assignee: Baidu USA LLC
    Inventors: Dingcheng Li, Jingyuan Zhang, Ping Li
  • Patent number: 11562213
    Abstract: Logic may reduce the size of runtime memory for deep neural network inference computations. Logic may determine, for two or more stages of a neural network, a count of shared block allocations, or shared memory block allocations, that concurrently exist during execution of the two or more stages. Logic may compare counts of the shared block allocations to determine a maximum count of the counts. Logic may reduce inference computation time for deep neural network inference computations. Logic may determine a size for each of the shared block allocations of the count of shared memory block allocations, to accommodate data to store in a shared memory during execution of the two or more stages of the cascaded neural network. Logic may determine a batch size per stage of the two or more stages of a cascaded neural network based on a lack interdependencies between input data.
    Type: Grant
    Filed: April 17, 2018
    Date of Patent: January 24, 2023
    Assignee: INTEL CORPORATION
    Inventors: Byoungwon Choe, Kwangwoong Park, Seokyong Byun