Patents by Inventor MILOS VASIC

MILOS VASIC has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20240126756
    Abstract: A method and one or more non-transitory storage media are provided to train and implement a one-hot encoder. During a training phase, computation of an encoder state is performed by executing a set of relational statements to extract unique categories in a first training data set, associate each unique category with a unique index, and generate a one-hot encoding for each unique category. The set of relational statements are executed by a query optimization engine. Execution of the set of relational statements is postponed until a result of each relational statement is needed, and the query optimization engine implements one or more optimizations when executing the set of relational statements. During an encoding phase, a set of categorical features in a second training data set are encoded based on the encoder state to form a set of encoded categorical features.
    Type: Application
    Filed: October 12, 2022
    Publication date: April 18, 2024
    Inventors: FELIX SCHMIDT, MATTEO CASSERINI, MILOS VASIC, MARIJA NIKOLIC
  • Publication number: 20240126798
    Abstract: In an embodiment, a computer stores, in memory or storage, many explanation profiles, many log entries, and definitions of many features that log entries contain. Some features may contain a logic statement such as a database query, and these are specially aggregated based on similarity. Based on the entity specified by an explanation profile, statistics are materialized for some or all features. Statistics calculation may be based on scheduled batches of log entries or a stream of live log entries. At runtime, an inference that is based on a new log entry is received. Based on an entity specified in the new log entry, a particular explanation profile is dynamically selected. Based on the new log entry and statistics of features for the selected explanation profile, a local explanation of the inference is generated. In an embodiment, an explanation text template is used to generate the local explanation.
    Type: Application
    Filed: May 30, 2023
    Publication date: April 18, 2024
    Inventors: Arno Schneuwly, Desislava Wagenknecht-Dimitrova, Felix Schmidt, Marija Nikolic, Matteo Casserini, Milos Vasic, Renata Khasanova
  • Publication number: 20230376743
    Abstract: The present invention avoids overfitting in deep neural network (DNN) training by using multitask learning (MTL) and self-supervised learning (SSL) techniques when training a multi-branch DNN to encode a sequence. In an embodiment, a computer first trains the DNN to perform a first task. The DNN contains: a first encoder in a first branch, a second encoder in a second branch, and an interpreter layer that combines data from the first branch and the second branch. The DNN second trains to perform a second task. After the first and second trainings, production encoding and inferencing occur. The first encoder encodes a sparse feature vector into a dense feature vector from which an inference is inferred. In an embodiment, a sequence of log messages is encoded into an encoded trace. An anomaly detector infers whether the sequence is anomalous. In an embodiment, the log messages are database commands.
    Type: Application
    Filed: May 19, 2022
    Publication date: November 23, 2023
    Inventors: Marija Nikolic, Nikola Milojkovic, Arno Schneuwly, Matteo Casserini, Milos Vasic, Renata Khasanova, Felix Schmidt
  • Publication number: 20230368054
    Abstract: The present invention relates to threshold estimation and calibration for anomaly detection. Herein are machine learning (ML) and extreme value theory (EVT) techniques for normalizing and thresholding anomaly scores without presuming a values distribution. In an embodiment, a computer receives many unnormalized anomaly scores and, according to peak over threshold (POT), selects a highest subset of the unnormalized anomaly scores that exceed a tail threshold. Based on the highest subset of the unnormalized anomaly scores, parameters of a probability density function are trained according to EVT. After training and in a production environment, a normalized anomaly score is generated based on an unnormalized anomaly score and the trained parameters of the probability density function. Anomaly detection compares the normalized anomaly score to an optimized anomaly threshold.
    Type: Application
    Filed: May 16, 2022
    Publication date: November 16, 2023
    Inventors: Marija Nikolic, Matteo Casserini, Arno Schneuwly, Nikola Milojkovic, Milos Vasic, Renata Khasanova, Felix Schmidt
  • Publication number: 20230362180
    Abstract: Techniques for implementing a semi-supervised framework for purpose-oriented anomaly detection are provided. In one technique, a data item in inputted into an unsupervised anomaly detection model, which generates first output. Based on the first output, it is determined whether the data item represents an anomaly. In response to determining that the data item represents an anomaly, the data item is inputted into a supervised classification model, which generates second output that indicates whether the data item is unknown. In response to determining that the data item is unknown, a training instance is generated based on the data item. The supervised classification model is updated based on the training instance.
    Type: Application
    Filed: May 9, 2022
    Publication date: November 9, 2023
    Inventors: Milos Vasic, Saeid Allahdadian, Matteo Casserini, Felix Schmidt, Andrew Brownsword
  • Patent number: 11704386
    Abstract: Herein are feature extraction mechanisms that receive parsed log messages as inputs and transform them into numerical feature vectors for machine learning models (MLMs). In an embodiment, a computer extracts fields from a log message. Each field specifies a name, a text value, and a type. For each field, a field transformer for the field is dynamically selected based the field's name and/or the field's type. The field transformer converts the field's text value into a value of the field's type. A feature encoder for the value of the field's type is dynamically selected based on the field's type and/or a range of the field's values that occur in a training corpus of an MLM. From the feature encoder, an encoding of the value of the field's typed is stored into a feature vector. Based on the MLM and the feature vector, the log message is detected as anomalous.
    Type: Grant
    Filed: March 12, 2021
    Date of Patent: July 18, 2023
    Assignee: Oracle International Corporation
    Inventors: Amin Suzani, Saeid Allahdadian, Milos Vasic, Matteo Casserini, Hamed Ahmadi, Felix Schmidt, Andrew Brownsword, Nipun Agarwal
  • Publication number: 20220318684
    Abstract: Techniques are provided for sparse ensembling of unsupervised machine learning models. In an embodiment, the proposed architecture is composed of multiple unsupervised machine learning models that each produce a score as output and a gating network that analyzes the inputs and outputs of the unsupervised machine learning models to select an optimal ensemble of unsupervised machine learning models. The gating network is trained to choose a minimal number of the multiple unsupervised machine learning models whose scores are combined to create a final score that matches or closely resembles a final score that is computed using all the scores of the multiple unsupervised machine learning models.
    Type: Application
    Filed: April 2, 2021
    Publication date: October 6, 2022
    Inventors: SAEID ALLAHDADIAN, AMIN SUZANI, MILOS VASIC, MATTEO CASSERINI, ANDREW BROWNSWORD, FELIX SCHMIDT, NIPUN AGARWAL
  • Publication number: 20220292304
    Abstract: Herein are feature extraction mechanisms that receive parsed log messages as inputs and transform them into numerical feature vectors for machine learning models (MLMs). In an embodiment, a computer extracts fields from a log message. Each field specifies a name, a text value, and a type. For each field, a field transformer for the field is dynamically selected based the field's name and/or the field's type. The field transformer converts the field's text value into a value of the field's type. A feature encoder for the value of the field's type is dynamically selected based on the field's type and/or a range of the field's values that occur in a training corpus of an MLM. From the feature encoder, an encoding of the value of the field's typed is stored into a feature vector. Based on the MLM and the feature vector, the log message is detected as anomalous or not.
    Type: Application
    Filed: March 12, 2021
    Publication date: September 15, 2022
    Inventors: AMIN SUZANI, SAEID ALLAHDADIAN, MILOS VASIC, MATTEO CASSERINI, HAMED AHMADI, FELIX SCHMIDT, ANDREW BROWNSWORD, NIPUN AGARWAL
  • Publication number: 20220188694
    Abstract: Approaches herein relate to model decay of an anomaly detector due to concept drift. Herein are machine learning techniques for dynamically self-tuning an anomaly score threshold. In an embodiment in a production environment, a computer receives an item in a stream of items. A machine learning (ML) model hosted by the computer infers by calculation an anomaly score for the item. Whether the item is anomalous or not is decided based on the anomaly score and an adaptive anomaly threshold that dynamically fluctuates. A moving standard deviation of anomaly scores is adjusted based on a moving average of anomaly scores. The moving average of anomaly scores is then adjusted based on the anomaly score. The adaptive anomaly threshold is then adjusted based on the moving average of anomaly scores and the moving standard deviation of anomaly scores.
    Type: Application
    Filed: December 15, 2020
    Publication date: June 16, 2022
    Inventors: Amin Suzani, Matteo Casserini, Milos Vasic, Saeid Allahdadian, Andrew Brownsword, Hamed Ahmadi, Felix Schmidt, Nipun Agarwal
  • Publication number: 20220188410
    Abstract: Approaches herein relate to reconstructive models such as an autoencoder for anomaly detection. Herein are machine learning techniques that detect and suppress any feature that causes model decay by concept drift. In an embodiment in a production environment, a computer initializes an unsuppressed subset of features with a plurality of features that an already-trained reconstructive model can process. A respective reconstruction error of each feature of the unsuppressed subset of features is calculated. The computer detects that a respective moving average based on the reconstruction error of a particular feature of the unsuppressed subset of features exceeds a respective feature suppression threshold of the particular feature, which causes removal of the particular feature from the unsuppressed subset of features.
    Type: Application
    Filed: December 15, 2020
    Publication date: June 16, 2022
    Inventors: SAEID ALLAHDADIAN, ANDREW BROWNSWORD, MILOS VASIC, MATTEO CASSERINI, AMIN SUZANI, HAMED AHMADI, FELIX SCHMIDT, NIPUN AGARWAL
  • Publication number: 20220156578
    Abstract: Approaches herein relate to reconstructive models such as an autoencoder for anomaly detection. Herein are machine learning techniques that measure inference confidence based on reconstruction error trends. In an embodiment, a computer hosts a reconstructive model that encodes and decodes features. Based on that decoding, the following are automatically calculated: a respective reconstruction error of each feature, a respective moving average of reconstruction errors of each feature, an average of the moving averages of the reconstruction errors of all features, a standard deviation of the moving averages of the reconstruction errors of all features, and a confidence of decoding the features that is based on a ratio of the average of the moving averages of the reconstruction errors to the standard deviation of the moving averages of the reconstruction errors. The computer detects and indicates that a threshold exceeds the confidence of decoding, which may cause important automatic reactions herein.
    Type: Application
    Filed: November 16, 2020
    Publication date: May 19, 2022
    Inventors: SAEID ALLAHDADIAN, MATTEO CASSERINI, ANDREW BROWNSWORD, AMIN SUZANI, MILOS VASIC, FELIX SCHMIDT, NIPUN AGARWAL
  • Publication number: 20220138504
    Abstract: In an embodiment based on computer(s), an ML model is trained to detect outliers. The ML model calculates anomaly scores that include a respective anomaly score for each item in a validation dataset. The anomaly scores are automatically organized by sorting and/or clustering. Based on the organized anomaly scores, a separation is measured that indicates fitness of the ML model. In an embodiment, a computer performs two-clustering of anomaly scores into a first organization that consists of a first normal cluster of anomaly scores and a first anomaly cluster of anomaly scores. The computer performs three-clustering of the same anomaly scores into a second organization that consists of a second normal cluster of anomaly scores, a second anomaly cluster of anomaly scores, and a middle cluster of anomaly scores. A distribution difference between the first organization and the second organization is measured. An ML model is processed based on the distribution difference.
    Type: Application
    Filed: October 29, 2020
    Publication date: May 5, 2022
    Inventors: Hesam Fathi Moghadam, Anatoly Yakovlev, Sandeep Agrawal, Venkatanathan Varadarajan, Robert Hopkins, Matteo Casserini, Milos Vasic, Sanjay Jinturkar, Nipun Agarwal
  • Publication number: 20220108181
    Abstract: A multilayer perceptron herein contains an already-trained combined sequence of residual blocks that contains a semantic sequence of residual blocks and a contextual sequence of residual blocks. The semantic sequence of residual blocks contains a semantic sequence of layers of an autoencoder. The contextual sequence of residual blocks contains a contextual sequence of layers of a recurrent neural network. Each residual block of the combined sequence of residual blocks is used based on a respective survival probability. By the autoencoder and based on the using each residual block of the semantic sequence, a previous entry of a log is semantically encoded. By the recurrent neural network and based on the using each residual block of the contextual sequence, a next entry of the log is predicted. In an embodiment during training, survival probabilities are hyperparameters that are learned and used to probabilistically skip residual blocks such that the multilayer perceptron has stochastic depth.
    Type: Application
    Filed: October 7, 2020
    Publication date: April 7, 2022
    Inventors: HAMED AHMADI, SAEID ALLAHDADIAN, MATTEO CASSERINI, MILOS VASIC, AMIN SUZANI, FELIX SCHMIDT, ANDREW BROWNSWORD, NIPUN AGARWAL