Patents Examined by Michael J Huntley
  • Patent number: 12112259
    Abstract: Features related to systems and methods for reinforcement learning are described. The environment includes one or more agents for automating the training of reinforcement learning (RL) models. The environment may include a simulator or real-world observations. The features described identify key training parameters, resource configurations, virtual network configurations, and simulators based on historical training data.
    Type: Grant
    Filed: November 21, 2018
    Date of Patent: October 8, 2024
    Assignee: Amazon Technologies, Inc.
    Inventor: Leo Parker Dirac
  • Patent number: 12106203
    Abstract: Systems and methods for analyzing the usage of a set of workloads in a hyper-converged infrastructure are disclosed. A neural network model is trained based upon historical usage data of the set of workloads. The neural network model can make usage predictions of future demands on the set of workloads to minimize over-allocation or under-allocation of resources to the workloads.
    Type: Grant
    Filed: July 12, 2018
    Date of Patent: October 1, 2024
    Assignee: VMware LLC
    Inventors: Alaa Shaabana, Gregory Jean-Baptiste, Anant Agarwal, Rahul Chandrasekaran, Pawan Saxena
  • Patent number: 12106212
    Abstract: The present disclosure is directed to methods and apparatus for validating and authenticating use of machine learning models. For example, various techniques are described herein to limit the vulnerability of machine learning models to attack and/or exploitation of the model for malicious use, and for detecting when such attack/exploitation has occurred. Additionally, various embodiments described herein promote the protection of sensitive and/or valuable data, for example by ensuring only licensed use is permissible. Moreover, techniques are described for version tracking, usage tracking, permission tracking, and evolution of machine learning models.
    Type: Grant
    Filed: January 14, 2020
    Date of Patent: October 1, 2024
    Assignee: Koninklijke Philips N.V.
    Inventors: Shawn Arie Peter Stapleton, Amir Mohammad Tahmasebi Maraghoosh
  • Patent number: 12099922
    Abstract: A computer-implemented method for detecting an operation tendency is disclosed. The method includes preparing a general model for generating a general anomaly score. The method also includes preparing a specific model, for generating a specific anomaly score, trained with a set of a plurality of operation data related to operation by a target operator. The method further includes receiving input operation data. The method includes also calculating a detection score related to the operation tendency by using a general anomaly score and a specific anomaly score generated for the input operation data. Further the method includes outputting a result based on the detection score.
    Type: Grant
    Filed: May 30, 2019
    Date of Patent: September 24, 2024
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Kun Zhao, Takayuki Katsuki, Takayuki Yoshizumi
  • Patent number: 12086695
    Abstract: A system for training a multi-task model includes a processor and a memory in communication with the processor. The memory has a multi-task training module having instructions that, when executed by the processor, causes the processor to provide simulation training data having a plurality of samples to a multi-task model capable of performing at least a first task and a second task using at least one shared. The training module further causes the processor to determine a first value (gradience or loss) for the first task and a second value (gradience or loss) for a second task using the simulation training data and the at least one shared parameter, determine a task induced variance between the first value and the second value, and iteratively adjust the at least one shared parameter to reduce the task induced variance.
    Type: Grant
    Filed: March 18, 2021
    Date of Patent: September 10, 2024
    Assignee: Toyota Research Institute, Inc.
    Inventors: Dennis Park, Adrien David Gaidon
  • Patent number: 12086714
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for training a neural network used to select actions performed by a reinforcement learning agent interacting with an environment. In one aspect, a method includes maintaining a replay memory, where the replay memory stores pieces of experience data generated as a result of the reinforcement learning agent interacting with the environment. Each piece of experience data is associated with a respective expected learning progress measure that is a measure of an expected amount of progress made in the training of the neural network if the neural network is trained on the piece of experience data. The method further includes selecting a piece of experience data from the replay memory by prioritizing for selection pieces of experience data having relatively higher expected learning progress measures and training the neural network on the selected piece of experience data.
    Type: Grant
    Filed: January 30, 2023
    Date of Patent: September 10, 2024
    Assignee: DeepMind Technologies Limited
    Inventors: Tom Schaul, John Quan, David Silver
  • Patent number: 12086713
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for evaluating candidate output sequences using language model neural networks. In particular, an auto-regressive language model neural network is used to generate a candidate output sequence. The same auto-regressive language model neural network is used to evaluate the candidate output sequence to determine rating scores for each of one or more criteria. The rating score(s) are then used to determine whether to provide the candidate output sequence.
    Type: Grant
    Filed: July 28, 2022
    Date of Patent: September 10, 2024
    Assignee: Google LLC
    Inventors: Daniel De Freitas Adiwardana, Noam M. Shazeer
  • Patent number: 12086716
    Abstract: A method for constructing a multimodality-based medical large model, and a related device thereof are provided. The medical large model includes a multimodal transformer T, a prompt manager M, a dialogue engine L, a task controller H, and a multimodal foundation (MMF) that includes at least one medical foundation model (MFM). Five stages, namely modal analysis, model allocation, downstream task result feedback, modal transformation normalization, and response generation are designed.
    Type: Grant
    Filed: November 13, 2023
    Date of Patent: September 10, 2024
    Assignee: AthenaEyes CO., LTD.
    Inventors: Weihua Liu, Jianhua Qiu, Jinmin Ma
  • Patent number: 12067483
    Abstract: Embodiments of the present invention provide a machine learning model training method, including: obtaining target task training data and N categories of support task training data; inputting the target task training data and the N categories of support task training data into a memory model to obtain target task training feature data and N categories of support task training feature data; training the target task model based on the target task training feature data and obtaining a first loss of the target task model, and separately training respectively corresponding support task models based on the N categories of support task training feature data and obtaining respective second losses of the N support task models; and updating the memory model, the target task model, and the N support task models based on the first loss and the respective second losses of the N support task models.
    Type: Grant
    Filed: June 4, 2019
    Date of Patent: August 20, 2024
    Assignee: Huawei Technologies Co., Ltd.
    Inventors: Bin Wu, Fengwei Zhou, Zhenguo Li
  • Patent number: 12061960
    Abstract: A learning device is configured to perform learning of a decision tree. The learning device includes a gradient output unit and a branch score calculator. The gradient output unit is configured to output a cumulative sum of gradient information corresponding to each value of a feature amount of learning data. The branch score calculator is configured to calculate a branch score used for determining a branch condition for a node of the decision tree, from the cumulative sum without using a dividing circuit.
    Type: Grant
    Filed: July 31, 2019
    Date of Patent: August 13, 2024
    Assignee: RICOH COMPANY, LTD.
    Inventors: Takuya Tanaka, Ryosuke Kasahara
  • Patent number: 12045705
    Abstract: A system receives information associated with an interaction with an individual in a context. Then, the system analyzes the information to extract features associated with one or more attributes of the individual. Moreover, the system generates, based at least in part on the extracted features, a group of behavioral agents in a multi-layer hierarchy that automatically mimics the one or more attributes. Next, the system calculates one or more performance metrics associated with the group of behavioral agents and the one or more attributes. Furthermore, the system determines, based at least in part on the one or more performance metrics, one or more deficiencies in the extracted features. Additionally, the system selectively acquires second information associated with additional interaction with the individual in the context based at least in part on the one or more deficiencies to at least in part correct for the one or more deficiencies.
    Type: Grant
    Filed: May 20, 2018
    Date of Patent: July 23, 2024
    Assignee: Artificial Intelligence Foundation, Inc.
    Inventors: Robert Marc Meadows, Lars Ulrich Buttler, Alan Peter Swearengen
  • Patent number: 12045716
    Abstract: A method of updating a first neural network is disclosed. The method includes providing a computer system with a computer-readable memory that stores specific computer-executable instructions for the first neural network and a second neural network separate from the first neural network. The method also includes providing one or more processors in communication with the computer-readable memory. The one or more processors are programmed by the computer-executable instructions to at least process a first data with the first neural network, process a second data with the second neural network, update a weight in a node of the second neural network by a delta amount as a function of the processing of the second data with the second neural network, and update a weight in a node of the first neural network as a function of the delta amount. A computer system for updating a first neural network is also disclosed. Other features of the preferred embodiments are also disclosed.
    Type: Grant
    Filed: September 14, 2020
    Date of Patent: July 23, 2024
    Assignee: Lucinity ehf
    Inventors: Justin Bercich, Theresa Bercich, Gudmundur Runar Kristjansson, Anush Vasudevan
  • Patent number: 12039429
    Abstract: An equilibrium computation acceleration method and system for a sparse recurrent neural network determine scheduling information based on an arbitration result of sparsity of a weight matrix, select a computation submodule having operating voltage and operating frequency that match the scheduling information or a computation submodule having operating voltage and operating frequency that are adjusted to match the scheduling information, and use the selected computation submodule to perform a zero-hop operation and a multiply-add operation in sequence, to accelerate equilibrium computation. The equilibrium computation acceleration system includes a data transmission module, an equilibrium computation scheduling module with a plurality of independent built-in computation submodules, and a voltage-adjustable equilibrium computation module. An error monitor is configured to achieve dynamic voltage adjustment.
    Type: Grant
    Filed: October 29, 2020
    Date of Patent: July 16, 2024
    Assignee: NANJING PROCHIP ELECTRONIC TECHNOLOGY CO., LTD
    Inventor: Zhen Wang
  • Patent number: 12026609
    Abstract: A resistive processing unit (RPU) that includes a pair of transistors connected in series providing an update function for a weight of a training methodology to the RPU, and a read transistor for reading the weight of the training methodology. In some embodiments, the resistive processing unit (RPU) further includes a capacitor connecting a gate of the read transistor to the air of transistors providing the update function for the resistive processing unit (RPU). The capacitor stores said weight of training methodology for the RPU.
    Type: Grant
    Filed: May 18, 2023
    Date of Patent: July 2, 2024
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Tayfun Gokmen, Seyoung Kim, Dennis M. Newns, Yurii A. Vlasov
  • Patent number: 12020154
    Abstract: Disclosed are a data processing method, a device and a medium. The method includes: acquiring first feature data and a source identification of data to be processed; determining a first unshared hidden unit, corresponding to the source identification, in an autoencoder, wherein the autoencoder includes a plurality of first unshared hidden units that do not share a parameter with each other; inputting the first feature data into the determined first unshared hidden unit, to perform noise cancellation, and outputting second feature data meeting a set standard; inputting the second feature data into a first shared hidden unit of the autoencoder to map the second feature data to a set feature space through the first shared hidden unit, and outputting mapping data; and inputting the mapping data into a shared feature layer of the autoencoder, and outputting common feature data in the first feature data, extracted by the shared feature layer.
    Type: Grant
    Filed: April 23, 2020
    Date of Patent: June 25, 2024
    Assignee: BEIJING XINTANG SICHUANG EDUCATION TECHNOLOGY CO., LTD
    Inventors: Song Yang, Jian Huang, Fei Yang, Zitao Liu, Yan Huang
  • Patent number: 12020147
    Abstract: Effectively training machine learning systems with incomplete/partial labels is a practical, technical problem that solutions described herein attempt to overcome. In particular, an approach to modify loss functions on a proportionality basis is noted in some embodiments. In other embodiments, a graph neural network is provided to help identify correlations/causations as between categories. In another set of embodiments, a prediction approach is described to, based on originally provided labels, predict labels for unlabelled training samples such that the proportion of labelled labels relative to all labels is increased.
    Type: Grant
    Filed: November 15, 2019
    Date of Patent: June 25, 2024
    Assignee: ROYAL BANK OF CANADA
    Inventors: Thibaut Durand, Nazanin Mehrasa, Gregory Mori
  • Patent number: 12008469
    Abstract: A single neural network model can be used by each computing engine (CE) in a neural network processor to perform convolution operations in parallel for one or more stacks of convolutional layers. An input feature map can be divided into N chunks to be processed by N CEs, respectively. Each CE can process a last portion of a respective chunk to generate respective shared states to be used by a subsequent CE. A first CE uses pre-computed states to generate a first portion of an output feature map, while other CEs use shared states computed by a preceding CE to generate respective portions of the output feature map.
    Type: Grant
    Filed: September 1, 2020
    Date of Patent: June 11, 2024
    Assignee: Amazon Technologies, Inc.
    Inventors: Thiam Khean Hah, Randy Renfu Huang, Richard John Heaton, Ron Diamant, Vignesh Vivekraja
  • Patent number: 12008473
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for augmenting machine learning language models using search engine results. One of the methods includes obtaining question data representing a question; generating, from the question data, a search engine query for a search engine; obtaining a plurality of documents identified by the search engine in response to processing the search engine query; generating, from the plurality of documents, a plurality of conditioning inputs each representing at least a portion of one or more of the obtained documents; for each of a plurality of the generated conditioning inputs, processing a network input generated from (i) the question data and (ii) the conditioning input using a neural network to generate a network output representing a candidate answer to the question; and generating, from the network outputs representing respective candidate answers, answer data representing a final answer to the question.
    Type: Grant
    Filed: January 31, 2023
    Date of Patent: June 11, 2024
    Assignee: DeepMind Technologies Limited
    Inventors: Angeliki Lazaridou, Elena Gribovskaya, Nikolai Grigorev, Wojciech Jan Stokowiec
  • Patent number: 11995540
    Abstract: A computer-implemented method, a computer program product, and a computer processing system are provided for online learning for a Dynamic Boltzmann Machine (DyBM) with hidden units. The method includes imposing, by a processor device, limited connections in the DyBM where (i) a current observation x[t] depends only on latest hidden units h[t-1/2] and all previous observations x[<t] and (ii) the latest hidden units h[t-1/2] depend on all the previous observations x[<t] while being independent of older hidden units h[t-1/2]. The method further includes computing, by the processor device, gradients of an objective function. The method also includes optimizing, by the processor device, the objective function in polynomial time using a stochastic Gradient Descent algorithm applied to the gradients.
    Type: Grant
    Filed: October 11, 2018
    Date of Patent: May 28, 2024
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Hiroshi Kajino, Takayuki Osogami
  • Patent number: 11972344
    Abstract: A method, system, and computer program product, including generating, using a linear probe, confidence scores through flattened intermediate representations and theoretically-justified weighting of samples during a training of the simple model using the confidence scores of the intermediate representations.
    Type: Grant
    Filed: November 28, 2018
    Date of Patent: April 30, 2024
    Assignee: International Business Machines Corporation
    Inventors: Amit Dhurandhar, Karthikeyan Shanmugam, Ronny Luss, Peder Andreas Olsen