Patents by Inventor Sanjay Jinturkar

Sanjay Jinturkar has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20240086763
    Abstract: Techniques for computing global feature explanations using adaptive sampling are provided. In one technique, first and second samples from an dataset are identified. A first set of feature importance values (FIVs) is generated based on the first sample and a machine-learned model. A second set of FIVs is generated based on the second sample and the model. If a result of a comparison between the first and second FIV sets does not satisfy criteria, then: (i) an aggregated set is generated based on the last two FIV sets; (ii) a new sample that is double the size of a previous sample is identified from the dataset; (iii) a current FIV set is generated based on the new sample and the model; (iv) determine whether a result of a comparison between the current and aggregated FIV sets satisfies criteria; repeating (i)-(iv) until the result of the last comparison satisfies the criteria.
    Type: Application
    Filed: September 14, 2022
    Publication date: March 14, 2024
    Inventors: Jeremy Plassmann, Anatoly Yakovlev, Sandeep R. Agrawal, Ali Moharrer, Sanjay Jinturkar, Nipun Agarwal
  • Patent number: 11720751
    Abstract: A model-agnostic global explainer for textual data processing (NLP) machine learning (ML) models, “NLP-MLX”, is described herein. NLP-MLX explains global behavior of arbitrary NLP ML models by identifying globally-important tokens within a textual dataset containing text data. NLP-MLX accommodates any arbitrary combination of training dataset pre-processing operations used by the NLP ML model. NLP-MLX includes four main stages. A Text Analysis stage converts text in documents of a target dataset into tokens. A Token Extraction stage uses pre-processing techniques to efficiently pre-filter the complete list of tokens into a smaller set of candidate important tokens. A Perturbation Generation stage perturbs tokens within documents of the dataset to help evaluate the effect of different tokens, and combinations of tokens, on the model's predictions.
    Type: Grant
    Filed: January 11, 2021
    Date of Patent: August 8, 2023
    Assignee: Oracle International Corporation
    Inventors: Zahra Zohrevand, Tayler Hetherington, Karoon Rashedi Nia, Yasha Pushak, Sanjay Jinturkar, Nipun Agarwal
  • Patent number: 11687540
    Abstract: Techniques are described for fast approximate conditional sampling by randomly sampling a dataset and then performing a nearest neighbor search on the pre-sampled dataset to reduce the data over which the nearest neighbor search must be performed and, according to an embodiment, to effectively reduce the number of nearest neighbors that are to be found within the random sample. Furthermore, KD-Tree-based stratified sampling is used to generate a representative sample of a dataset. KD-Tree-based stratified sampling may be used to identify the random sample for fast approximate conditional sampling, which reduces variance in the resulting data sample. As such, using KD-Tree-based stratified sampling to generate the random sample for fast approximate conditional sampling ensures that any nearest neighbor selected, for a target data instance, from the random sample is likely to be among the nearest neighbors of the target data instance within the unsampled dataset.
    Type: Grant
    Filed: February 18, 2021
    Date of Patent: June 27, 2023
    Assignee: Oracle International Corporation
    Inventors: Yasha Pushak, Tayler Hetherington, Karoon Rashedi Nia, Zahra Zohrevand, Sanjay Jinturkar, Nipun Agarwal
  • Publication number: 20230153394
    Abstract: Herein are timeseries preprocessing, model selection, and hyperparameter tuning techniques for forecasting development based on temporal statistics of a timeseries and a single feed-forward pass through a machine learning (ML) pipeline. In an embodiment, a computer hosts and operates the ML pipeline that automatically measures temporal statistic(s) of a timeseries. ML algorithm selection, cross validation, and hyperparameters tuning is based on the temporal statistics of the timeseries. The result from the ML pipeline is a rigorously trained and production ready ML model that is validated to have increased accuracy for multiple prediction horizons. Based on the temporal statistics, efficiency is achieved by asymmetry of investment of computer resources in the tuning and training of the most promising ML algorithm(s). Compared to other approaches, this ML pipeline produces a more accurate ML model for a given amount of computer resources and consumes fewer computer resources to achieve a given accuracy.
    Type: Application
    Filed: November 17, 2021
    Publication date: May 18, 2023
    Inventors: Ritesh Ahuja, Anatoly Yakovlev, Venkatanathan Varadarajan, Sandeep R. Agrawal, Hesam Fathi Moghadam, Sanjay Jinturkar, Nipun Agarwal
  • Patent number: 11514697
    Abstract: Herein is a probabilistic indexing technique for searching semi-structured text documents in columnar storage formats such as Parquet, using columnar input/output (I/O) avoidance, and needing minimal storage overhead. In an embodiment, a computer associates columns with text strings that occur in semi-structured documents. Text words that occur in the text strings are detected. Respectively for each text word, a bitmap, of a plurality of bitmaps, that contains a respective bit for each column is generated. Based on at least one of the bitmaps, some of the columns or some of the semi-structured documents are accessed.
    Type: Grant
    Filed: July 15, 2020
    Date of Patent: November 29, 2022
    Assignee: Oracle International Corporation
    Inventors: Jian Wen, Hamed Ahmadi, Sanjay Jinturkar, Nipun Agarwal, Lijian Wan, Shrikumar Hariharasubrahmanian
  • Publication number: 20220366297
    Abstract: In an embodiment, a computer hosts a machine learning (ML) model that infers a particular inference for a particular tuple that is based on many features. For each feature, and for each of many original tuples, the computer: a) randomly selects many perturbed values from original values of the feature in the original tuples, b) generates perturbed tuples that are based on the original tuple and a respective perturbed value, c) causes the ML model to infer a respective perturbed inference for each perturbed tuple, and d) measures a respective difference between each perturbed inference of the perturbed tuples and the particular inference. For each feature, a respective importance of the feature is calculated based on the differences measured for the feature. Feature importances may be used to rank features by influence and/or generate a local ML explainability (MLX) explanation.
    Type: Application
    Filed: May 13, 2021
    Publication date: November 17, 2022
    Inventors: Yasha Pushak, Zahra Zohrevand, Tayler Hetherington, Karoon Rashedi Nia, Sanjay Jinturkar, Nipun Agarwal
  • Publication number: 20220335255
    Abstract: In an embodiment, a computer assigns a respective probability distribution to each of many features that include a first feature and a second feature that are assigned different probability distributions. For each original tuple that are based on the features, a machine learning (ML) model infers a respective original inference. For each feature, and for each original tuple, the computer: a) generates perturbed values based on the probability distribution of the feature, b) generates perturbed tuples that are based on the original tuple and a respective perturbed value, c) causes the ML model to infer a respective perturbed inference for each perturbed tuple, and d) measures a respective difference between each perturbed inference and the original inference. A respective importance of each feature is calculated based on the differences measured for the feature. Feature importances may be used to rank features by influence and/or generate a global or local ML explainability (MLX) explanation.
    Type: Application
    Filed: April 16, 2021
    Publication date: October 20, 2022
    Inventors: ZAHRA ZOHREVAND, YASHA PUSHAK, TAYLER HETHERINGTON, KAROON RASHEDI NIA, SANJAY JINTURKAR, NIPUN AGARWAL
  • Publication number: 20220309360
    Abstract: Herein are techniques for topic modeling and content perturbation that provide machine learning (ML) explainability (MLX) for natural language processing (NLP). A computer hosts an ML model that infers an original inference for each of many text documents that contain many distinct terms. To each text document (TD) is assigned, based on terms in the TD, a topic that contains a subset of the distinct terms. In a perturbed copy of each TD, a perturbed subset of the distinct terms is replaced. For the perturbed copy of each TD, the ML model infers a perturbed inference. For TDs of a topic, the computer detects that a difference between original inferences of the TDs of the topic and perturbed inferences of the TDs of the topic exceeds a threshold. Based on terms in the TDs of the topic, the topic is replaced with multiple, finer-grained new topics. After sufficient topic modeling, a regional explanation of the ML model is generated.
    Type: Application
    Filed: March 25, 2021
    Publication date: September 29, 2022
    Inventors: Zahra Zohrevand, Tayler Hetherington, Karoon Rashedi Nia, Yasha Pushak, Sanjay Jinturkar, Nipun Agarwal
  • Patent number: 11451670
    Abstract: Herein are machine learning (ML) techniques for unsupervised training with a corpus of signaling system 7 (SS7) messages having a diversity of called and calling parties, operation codes (opcodes) and transaction types, numbering plans and nature of address indicators, and mobile country codes and network codes. In an embodiment, a computer stores SS7 messages that are not labeled as anomalous or non-anomalous. Each SS7 message contains an opcode and other fields. For each SS7 message, the opcode of the SS7 message is stored into a respective feature vector (FV) of many FVs that are based on respective unlabeled SS7 messages. The FVs contain many distinct opcodes. Based on the FVs that contain many distinct opcodes and that are based on respective unlabeled SS7 messages, an ML model such as a reconstructive model such as an autoencoder is unsupervised trained to detect an anomalous SS7 message.
    Type: Grant
    Filed: December 16, 2020
    Date of Patent: September 20, 2022
    Assignee: Oracle International Corporation
    Inventors: Hamed Ahmadi, Ali Moharrer, Venkatanathan Varadarajan, Vaseem Akram, Nishesh Rai, Reema Hingorani, Sanjay Jinturkar, Nipun Agarwal
  • Publication number: 20220261400
    Abstract: Techniques are described for fast approximate conditional sampling by randomly sampling a dataset and then performing a nearest neighbor search on the pre-sampled dataset to reduce the data over which the nearest neighbor search must be performed and, according to an embodiment, to effectively reduce the number of nearest neighbors that are to be found within the random sample. Furthermore, KD-Tree-based stratified sampling is used to generate a representative sample of a dataset. KD-Tree-based stratified sampling may be used to identify the random sample for fast approximate conditional sampling, which reduces variance in the resulting data sample. As such, using KD-Tree-based stratified sampling to generate the random sample for fast approximate conditional sampling ensures that any nearest neighbor selected, for a target data instance, from the random sample is likely to be among the nearest neighbors of the target data instance within the unsampled dataset.
    Type: Application
    Filed: February 18, 2021
    Publication date: August 18, 2022
    Inventors: Yasha Pushak, Tayler Hetherington, Karoon Rashedi Nia, Zahra Zohrevand, Sanjay Jinturkar, Nipun Agarwal
  • Publication number: 20220229983
    Abstract: A model-agnostic global explainer for textual data processing (NLP) machine learning (ML) models, “NLP-MLX”, is described herein. NLP-MLX explains global behavior of arbitrary NLP ML models by identifying globally-important tokens within a textual dataset containing text data. NLP-MLX accommodates any arbitrary combination of training dataset pre-processing operations used by the NLP ML model. NLP-MLX includes four main stages. A Text Analysis stage converts text in documents of a target dataset into tokens. A Token Extraction stage uses pre-processing techniques to efficiently pre-filter the complete list of tokens into a smaller set of candidate important tokens. A Perturbation Generation stage perturbs tokens within documents of the dataset to help evaluate the effect of different tokens, and combinations of tokens, on the model's predictions.
    Type: Application
    Filed: January 11, 2021
    Publication date: July 21, 2022
    Inventors: Zahra Zohrevand, Tayler Hetherington, Karoon Rashedi Nia, Yasha Pushak, Sanjay Jinturkar, Nipun Agarwal
  • Publication number: 20220198277
    Abstract: Herein are generative adversarial networks to ensure realistic local samples and surrogate models to provide machine learning (ML) explainability (MLX). Based on many features, an embodiment trains an ML model. The ML model inferences an original inference for original feature values respectively for many features. Based on the same features, a generator model is trained to generate realistic local samples that are distinct combinations of feature values for the features. A surrogate model is trained based on the generator model and based on the original inference by the ML model and/or the original feature values that the original inference is based on. Based on the surrogate model, the ML model is explained. The local samples may be weighted based on semantic similarity to the original feature values, which may facilitate training the surrogate model and/or ranking the relative importance of the features. Local sample weighting may be based on populating a random forest with the local samples.
    Type: Application
    Filed: December 22, 2020
    Publication date: June 23, 2022
    Inventors: Karoon Rashedi Nia, Tayler Hetherington, Zahra Zohrevand, Yasha Pushak, Sanjay Jinturkar, Nipun Agarwal
  • Publication number: 20220191332
    Abstract: Herein are machine learning (ML) techniques for unsupervised training with a corpus of signaling system 7 (SS7) messages having a diversity of called and calling parties, operation codes (opcodes) and transaction types, numbering plans and nature of address indicators, and mobile country codes and network codes. In an embodiment, a computer stores SS7 messages that are not labeled as anomalous or non-anomalous. Each SS7 message contains an opcode and other fields. For each SS7 message, the opcode of the SS7 message is stored into a respective feature vector (FV) of many FVs that are based on respective unlabeled SS7 messages. The FVs contain many distinct opcodes. Based on the FVs that contain many distinct opcodes and that are based on respective unlabeled SS7 messages, an ML model such as a reconstructive model such as an autoencoder is unsupervised trained to detect an anomalous SS7 message.
    Type: Application
    Filed: December 16, 2020
    Publication date: June 16, 2022
    Inventors: Hamed Ahmadi, Ali Moharrer, Venkatanathan Varadarajan, Vaseem Akram, Nishesh Rai, Reema Hingorani, Sanjay Jinturkar, Nipun Agarwal
  • Publication number: 20220188645
    Abstract: Herein are counterfactual explanations of machine learning (ML) inferencing provided by generative adversarial networks (GANs) that ensure realistic counterfactuals and use latent spaces to optimize perturbations. In an embodiment, a first computer trains a generator model in a GAN. A same or second computer hosts a classifier model that inferences an original label for original feature values respectively for many features. Runtime ML explainability (MLX) occurs on the first or second or a third computer as follows. The generator model from the GAN generates a sequence of revised feature values that are based on noise. The noise is iteratively optimized based on a distance between the original feature values and current revised feature values in the sequence of revised feature values. The classifier model inferences a current label respectively for each counterfactual in the sequence of revised feature values.
    Type: Application
    Filed: December 16, 2020
    Publication date: June 16, 2022
    Inventors: Karoon Rashedi Nia, Tayler Hetherington, Zahra Zohrevand, Yasha Pushak, Sanjay Jinturkar, Nipun Agarwal
  • Publication number: 20220172105
    Abstract: End-to-end explanation techniques, which efficiently explain the behavior (feature importance) of any machine learning model on large tabular datasets, are disclosed. These techniques comprise two down-sampling methods to efficiently select a small set of representative samples of a high-dimensional dataset for explaining a machine learning model by making use of the characteristics of the dataset or of an explainer of a machine learning model to optimize the explanation quality. These techniques significantly improve the explanation speed while maintaining the explanation quality of a full dataset evaluation.
    Type: Application
    Filed: November 30, 2020
    Publication date: June 2, 2022
    Inventors: KAROON RASHEDI NIA, TAYLER HETHERINGTON, ZAHRA ZOHREVAND, SANJAY JINTURKAR, NIPUN AGARWAL
  • Publication number: 20220138504
    Abstract: In an embodiment based on computer(s), an ML model is trained to detect outliers. The ML model calculates anomaly scores that include a respective anomaly score for each item in a validation dataset. The anomaly scores are automatically organized by sorting and/or clustering. Based on the organized anomaly scores, a separation is measured that indicates fitness of the ML model. In an embodiment, a computer performs two-clustering of anomaly scores into a first organization that consists of a first normal cluster of anomaly scores and a first anomaly cluster of anomaly scores. The computer performs three-clustering of the same anomaly scores into a second organization that consists of a second normal cluster of anomaly scores, a second anomaly cluster of anomaly scores, and a middle cluster of anomaly scores. A distribution difference between the first organization and the second organization is measured. An ML model is processed based on the distribution difference.
    Type: Application
    Filed: October 29, 2020
    Publication date: May 5, 2022
    Inventors: Hesam Fathi Moghadam, Anatoly Yakovlev, Sandeep Agrawal, Venkatanathan Varadarajan, Robert Hopkins, Matteo Casserini, Milos Vasic, Sanjay Jinturkar, Nipun Agarwal
  • Publication number: 20220129791
    Abstract: A systematic explainer is described herein, which comprises local, model-agnostic, surrogate ML model-based explanation techniques that faithfully explain predictions from any machine learning classifier or regressor. The systematic explainer systematically generates local data samples around a given target data sample, which improves on exhaustive or random data sample generation algorithms. Specifically, using principles of locality and approximation of local decision boundaries, techniques described herein identify a hypersphere (or data sample neighborhood) over which to train the surrogate ML model such that the surrogate ML model produces valuable, high-quality information explaining data samples in the neighborhood of the target data sample.
    Type: Application
    Filed: October 28, 2020
    Publication date: April 28, 2022
    Inventors: Karoon Rashedi Nia, Tayler Hetherington, Zahra Zohrevand, Sanjay Jinturkar, Nipun Agarwal
  • Publication number: 20220121955
    Abstract: Herein, a computer generates and evaluates many preprocessor configurations for a window preprocessor that transforms a training timeseries dataset for an ML model. With each preprocessor configuration, the window preprocessor is configured. The window preprocessor then converts the training timeseries dataset into a configuration-specific point-based dataset that is based on the preprocessor configuration. The ML model is trained based on the configuration-specific point-based dataset to calculate a score for the preprocessor configuration. Based on the scores of the many preprocessor configurations, an optimal preprocessor configuration is selected for finally configuring the window preprocessor, after which, the window preprocessor can optimally transform a new timeseries dataset such as in an offline or online production environment such as for real-time processing of a live streaming timeseries.
    Type: Application
    Filed: October 15, 2020
    Publication date: April 21, 2022
    Inventors: Nikan Chavoshi, Anatoly Yakovlev, Hesam Fathi Moghadam, Venkatanathan Varadarajan, Sandeep Agrawal, Ali Moharrer, Jingxiao Cai, Sanjay Jinturkar, Nipun Agarwal
  • Publication number: 20220043681
    Abstract: Herein, a computer receives a new training dataset for a target ML model. Proven or unproven respective values of hyperparameters of the target ML model are selected. An already-trained ML metamodel predicts an amount of memory that the target ML model will need, when configured with the respective values of the hyperparameters, to train with the new training dataset. In an embodiment, supervised training of the ML metamodel is as follows. The ML metamodel receives feature vectors that each contains distinct details of a respective past training of the target ML model of many and varied trainings of the target ML model. Those distinct details of each past training includes: respective values of the hyperparameters, and respective values of metafeatures of a respective training dataset of many training datasets. Each feature vector is labeled with a respective amount of memory that the target ML model needed during the respective past training.
    Type: Application
    Filed: August 4, 2020
    Publication date: February 10, 2022
    Inventors: Ali Moharrer, Sandeep R. Agrawal, Venkatanathan Varadarajan, Sanjay Jinturkar, Nipun Agarwal
  • Patent number: 11238035
    Abstract: Techniques are described herein for indexing personal information in columnar data storage format based files. In an embodiment, row groups of rows that comprise a plurality of columns are stored in a set of files. Each column of a row group is stored in a chunk of column pages in the set of files. A regular expression index that indexes a particular column in the set of files is stored for each row group. The regular expression index identifies column pages in the chunk of the particular column that include a particular column value that satisfies a regular expression specified in a query. The regular expression specified in the query in evaluated against the particular column using the regular expression index.
    Type: Grant
    Filed: March 10, 2020
    Date of Patent: February 1, 2022
    Assignee: ORACLE INTERNATIONAL CORPORATION
    Inventors: Hamed Ahmadi, Jian Wen, Shrikumar Hariharasubrahmanian, Sanjay Jinturkar, Nipun Agarwal