Patents Examined by Michael J Huntley
-
Patent number: 12112259Abstract: Features related to systems and methods for reinforcement learning are described. The environment includes one or more agents for automating the training of reinforcement learning (RL) models. The environment may include a simulator or real-world observations. The features described identify key training parameters, resource configurations, virtual network configurations, and simulators based on historical training data.Type: GrantFiled: November 21, 2018Date of Patent: October 8, 2024Assignee: Amazon Technologies, Inc.Inventor: Leo Parker Dirac
-
Patent number: 12106203Abstract: Systems and methods for analyzing the usage of a set of workloads in a hyper-converged infrastructure are disclosed. A neural network model is trained based upon historical usage data of the set of workloads. The neural network model can make usage predictions of future demands on the set of workloads to minimize over-allocation or under-allocation of resources to the workloads.Type: GrantFiled: July 12, 2018Date of Patent: October 1, 2024Assignee: VMware LLCInventors: Alaa Shaabana, Gregory Jean-Baptiste, Anant Agarwal, Rahul Chandrasekaran, Pawan Saxena
-
Patent number: 12106212Abstract: The present disclosure is directed to methods and apparatus for validating and authenticating use of machine learning models. For example, various techniques are described herein to limit the vulnerability of machine learning models to attack and/or exploitation of the model for malicious use, and for detecting when such attack/exploitation has occurred. Additionally, various embodiments described herein promote the protection of sensitive and/or valuable data, for example by ensuring only licensed use is permissible. Moreover, techniques are described for version tracking, usage tracking, permission tracking, and evolution of machine learning models.Type: GrantFiled: January 14, 2020Date of Patent: October 1, 2024Assignee: Koninklijke Philips N.V.Inventors: Shawn Arie Peter Stapleton, Amir Mohammad Tahmasebi Maraghoosh
-
Patent number: 12099922Abstract: A computer-implemented method for detecting an operation tendency is disclosed. The method includes preparing a general model for generating a general anomaly score. The method also includes preparing a specific model, for generating a specific anomaly score, trained with a set of a plurality of operation data related to operation by a target operator. The method further includes receiving input operation data. The method includes also calculating a detection score related to the operation tendency by using a general anomaly score and a specific anomaly score generated for the input operation data. Further the method includes outputting a result based on the detection score.Type: GrantFiled: May 30, 2019Date of Patent: September 24, 2024Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Kun Zhao, Takayuki Katsuki, Takayuki Yoshizumi
-
Patent number: 12086695Abstract: A system for training a multi-task model includes a processor and a memory in communication with the processor. The memory has a multi-task training module having instructions that, when executed by the processor, causes the processor to provide simulation training data having a plurality of samples to a multi-task model capable of performing at least a first task and a second task using at least one shared. The training module further causes the processor to determine a first value (gradience or loss) for the first task and a second value (gradience or loss) for a second task using the simulation training data and the at least one shared parameter, determine a task induced variance between the first value and the second value, and iteratively adjust the at least one shared parameter to reduce the task induced variance.Type: GrantFiled: March 18, 2021Date of Patent: September 10, 2024Assignee: Toyota Research Institute, Inc.Inventors: Dennis Park, Adrien David Gaidon
-
Patent number: 12086714Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for training a neural network used to select actions performed by a reinforcement learning agent interacting with an environment. In one aspect, a method includes maintaining a replay memory, where the replay memory stores pieces of experience data generated as a result of the reinforcement learning agent interacting with the environment. Each piece of experience data is associated with a respective expected learning progress measure that is a measure of an expected amount of progress made in the training of the neural network if the neural network is trained on the piece of experience data. The method further includes selecting a piece of experience data from the replay memory by prioritizing for selection pieces of experience data having relatively higher expected learning progress measures and training the neural network on the selected piece of experience data.Type: GrantFiled: January 30, 2023Date of Patent: September 10, 2024Assignee: DeepMind Technologies LimitedInventors: Tom Schaul, John Quan, David Silver
-
Patent number: 12086713Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for evaluating candidate output sequences using language model neural networks. In particular, an auto-regressive language model neural network is used to generate a candidate output sequence. The same auto-regressive language model neural network is used to evaluate the candidate output sequence to determine rating scores for each of one or more criteria. The rating score(s) are then used to determine whether to provide the candidate output sequence.Type: GrantFiled: July 28, 2022Date of Patent: September 10, 2024Assignee: Google LLCInventors: Daniel De Freitas Adiwardana, Noam M. Shazeer
-
Patent number: 12086716Abstract: A method for constructing a multimodality-based medical large model, and a related device thereof are provided. The medical large model includes a multimodal transformer T, a prompt manager M, a dialogue engine L, a task controller H, and a multimodal foundation (MMF) that includes at least one medical foundation model (MFM). Five stages, namely modal analysis, model allocation, downstream task result feedback, modal transformation normalization, and response generation are designed.Type: GrantFiled: November 13, 2023Date of Patent: September 10, 2024Assignee: AthenaEyes CO., LTD.Inventors: Weihua Liu, Jianhua Qiu, Jinmin Ma
-
Patent number: 12067483Abstract: Embodiments of the present invention provide a machine learning model training method, including: obtaining target task training data and N categories of support task training data; inputting the target task training data and the N categories of support task training data into a memory model to obtain target task training feature data and N categories of support task training feature data; training the target task model based on the target task training feature data and obtaining a first loss of the target task model, and separately training respectively corresponding support task models based on the N categories of support task training feature data and obtaining respective second losses of the N support task models; and updating the memory model, the target task model, and the N support task models based on the first loss and the respective second losses of the N support task models.Type: GrantFiled: June 4, 2019Date of Patent: August 20, 2024Assignee: Huawei Technologies Co., Ltd.Inventors: Bin Wu, Fengwei Zhou, Zhenguo Li
-
Patent number: 12061960Abstract: A learning device is configured to perform learning of a decision tree. The learning device includes a gradient output unit and a branch score calculator. The gradient output unit is configured to output a cumulative sum of gradient information corresponding to each value of a feature amount of learning data. The branch score calculator is configured to calculate a branch score used for determining a branch condition for a node of the decision tree, from the cumulative sum without using a dividing circuit.Type: GrantFiled: July 31, 2019Date of Patent: August 13, 2024Assignee: RICOH COMPANY, LTD.Inventors: Takuya Tanaka, Ryosuke Kasahara
-
Patent number: 12045705Abstract: A system receives information associated with an interaction with an individual in a context. Then, the system analyzes the information to extract features associated with one or more attributes of the individual. Moreover, the system generates, based at least in part on the extracted features, a group of behavioral agents in a multi-layer hierarchy that automatically mimics the one or more attributes. Next, the system calculates one or more performance metrics associated with the group of behavioral agents and the one or more attributes. Furthermore, the system determines, based at least in part on the one or more performance metrics, one or more deficiencies in the extracted features. Additionally, the system selectively acquires second information associated with additional interaction with the individual in the context based at least in part on the one or more deficiencies to at least in part correct for the one or more deficiencies.Type: GrantFiled: May 20, 2018Date of Patent: July 23, 2024Assignee: Artificial Intelligence Foundation, Inc.Inventors: Robert Marc Meadows, Lars Ulrich Buttler, Alan Peter Swearengen
-
Patent number: 12045716Abstract: A method of updating a first neural network is disclosed. The method includes providing a computer system with a computer-readable memory that stores specific computer-executable instructions for the first neural network and a second neural network separate from the first neural network. The method also includes providing one or more processors in communication with the computer-readable memory. The one or more processors are programmed by the computer-executable instructions to at least process a first data with the first neural network, process a second data with the second neural network, update a weight in a node of the second neural network by a delta amount as a function of the processing of the second data with the second neural network, and update a weight in a node of the first neural network as a function of the delta amount. A computer system for updating a first neural network is also disclosed. Other features of the preferred embodiments are also disclosed.Type: GrantFiled: September 14, 2020Date of Patent: July 23, 2024Assignee: Lucinity ehfInventors: Justin Bercich, Theresa Bercich, Gudmundur Runar Kristjansson, Anush Vasudevan
-
Patent number: 12039429Abstract: An equilibrium computation acceleration method and system for a sparse recurrent neural network determine scheduling information based on an arbitration result of sparsity of a weight matrix, select a computation submodule having operating voltage and operating frequency that match the scheduling information or a computation submodule having operating voltage and operating frequency that are adjusted to match the scheduling information, and use the selected computation submodule to perform a zero-hop operation and a multiply-add operation in sequence, to accelerate equilibrium computation. The equilibrium computation acceleration system includes a data transmission module, an equilibrium computation scheduling module with a plurality of independent built-in computation submodules, and a voltage-adjustable equilibrium computation module. An error monitor is configured to achieve dynamic voltage adjustment.Type: GrantFiled: October 29, 2020Date of Patent: July 16, 2024Assignee: NANJING PROCHIP ELECTRONIC TECHNOLOGY CO., LTDInventor: Zhen Wang
-
Patent number: 12026609Abstract: A resistive processing unit (RPU) that includes a pair of transistors connected in series providing an update function for a weight of a training methodology to the RPU, and a read transistor for reading the weight of the training methodology. In some embodiments, the resistive processing unit (RPU) further includes a capacitor connecting a gate of the read transistor to the air of transistors providing the update function for the resistive processing unit (RPU). The capacitor stores said weight of training methodology for the RPU.Type: GrantFiled: May 18, 2023Date of Patent: July 2, 2024Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Tayfun Gokmen, Seyoung Kim, Dennis M. Newns, Yurii A. Vlasov
-
Patent number: 12020154Abstract: Disclosed are a data processing method, a device and a medium. The method includes: acquiring first feature data and a source identification of data to be processed; determining a first unshared hidden unit, corresponding to the source identification, in an autoencoder, wherein the autoencoder includes a plurality of first unshared hidden units that do not share a parameter with each other; inputting the first feature data into the determined first unshared hidden unit, to perform noise cancellation, and outputting second feature data meeting a set standard; inputting the second feature data into a first shared hidden unit of the autoencoder to map the second feature data to a set feature space through the first shared hidden unit, and outputting mapping data; and inputting the mapping data into a shared feature layer of the autoencoder, and outputting common feature data in the first feature data, extracted by the shared feature layer.Type: GrantFiled: April 23, 2020Date of Patent: June 25, 2024Assignee: BEIJING XINTANG SICHUANG EDUCATION TECHNOLOGY CO., LTDInventors: Song Yang, Jian Huang, Fei Yang, Zitao Liu, Yan Huang
-
Patent number: 12020147Abstract: Effectively training machine learning systems with incomplete/partial labels is a practical, technical problem that solutions described herein attempt to overcome. In particular, an approach to modify loss functions on a proportionality basis is noted in some embodiments. In other embodiments, a graph neural network is provided to help identify correlations/causations as between categories. In another set of embodiments, a prediction approach is described to, based on originally provided labels, predict labels for unlabelled training samples such that the proportion of labelled labels relative to all labels is increased.Type: GrantFiled: November 15, 2019Date of Patent: June 25, 2024Assignee: ROYAL BANK OF CANADAInventors: Thibaut Durand, Nazanin Mehrasa, Gregory Mori
-
Patent number: 12008469Abstract: A single neural network model can be used by each computing engine (CE) in a neural network processor to perform convolution operations in parallel for one or more stacks of convolutional layers. An input feature map can be divided into N chunks to be processed by N CEs, respectively. Each CE can process a last portion of a respective chunk to generate respective shared states to be used by a subsequent CE. A first CE uses pre-computed states to generate a first portion of an output feature map, while other CEs use shared states computed by a preceding CE to generate respective portions of the output feature map.Type: GrantFiled: September 1, 2020Date of Patent: June 11, 2024Assignee: Amazon Technologies, Inc.Inventors: Thiam Khean Hah, Randy Renfu Huang, Richard John Heaton, Ron Diamant, Vignesh Vivekraja
-
Patent number: 12008473Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for augmenting machine learning language models using search engine results. One of the methods includes obtaining question data representing a question; generating, from the question data, a search engine query for a search engine; obtaining a plurality of documents identified by the search engine in response to processing the search engine query; generating, from the plurality of documents, a plurality of conditioning inputs each representing at least a portion of one or more of the obtained documents; for each of a plurality of the generated conditioning inputs, processing a network input generated from (i) the question data and (ii) the conditioning input using a neural network to generate a network output representing a candidate answer to the question; and generating, from the network outputs representing respective candidate answers, answer data representing a final answer to the question.Type: GrantFiled: January 31, 2023Date of Patent: June 11, 2024Assignee: DeepMind Technologies LimitedInventors: Angeliki Lazaridou, Elena Gribovskaya, Nikolai Grigorev, Wojciech Jan Stokowiec
-
Patent number: 11995540Abstract: A computer-implemented method, a computer program product, and a computer processing system are provided for online learning for a Dynamic Boltzmann Machine (DyBM) with hidden units. The method includes imposing, by a processor device, limited connections in the DyBM where (i) a current observation x[t] depends only on latest hidden units h[t-1/2] and all previous observations x[<t] and (ii) the latest hidden units h[t-1/2] depend on all the previous observations x[<t] while being independent of older hidden units h[t-1/2]. The method further includes computing, by the processor device, gradients of an objective function. The method also includes optimizing, by the processor device, the objective function in polynomial time using a stochastic Gradient Descent algorithm applied to the gradients.Type: GrantFiled: October 11, 2018Date of Patent: May 28, 2024Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Hiroshi Kajino, Takayuki Osogami
-
Patent number: 11972344Abstract: A method, system, and computer program product, including generating, using a linear probe, confidence scores through flattened intermediate representations and theoretically-justified weighting of samples during a training of the simple model using the confidence scores of the intermediate representations.Type: GrantFiled: November 28, 2018Date of Patent: April 30, 2024Assignee: International Business Machines CorporationInventors: Amit Dhurandhar, Karthikeyan Shanmugam, Ronny Luss, Peder Andreas Olsen