Patents by Inventor Amit Srivastava
Amit Srivastava has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 12045279Abstract: A system and method and for retrieving one or more visual assets includes receiving a search query for the one or more visual assets, the search query including textual data, encoding the textual data into one or more text embedding representations via a trained text representation machine-learning (ML) model, transmitting the one or more text embedding representations to a matching and selection unit, providing visual embedding representations of one or more visual assets to the matching and selection unit, comparing, by the matching and selection unit, the one or more text embedding representations to the visual embedding representations to identify one or more visual asset search results, and providing the one or more visual asset search results for display.Type: GrantFiled: November 30, 2021Date of Patent: July 23, 2024Assignee: Microsoft Technology Licensing, LLCInventors: Ji Li, Adit Krishnan, Amit Srivastava, Han Hu, Qi Dai, Yixuan Wei, Yue Cao
-
Patent number: 12047704Abstract: The present disclosure relates to application of artificial intelligence (AI) processing that adapts one or more video feeds relative to presentation content. Trained AI processing automatically generates a combined representation comprising one or more video feeds and presentation content. An exemplary combined representation is the result of contextual analysis by one or more trained AI models that are adapted to consider how to adapt presentation of a video feed relative to displayable presentation content (or visa-versa). A combined representation of one or more video feeds and presentation content is automatically generated (and subsequently rendered) based on a result of contextual evaluation of data associated with a video feed and data attributes of presentation content. A combined representation may comprise a modification of the one or more video feeds, objects of presentation content or a combination thereof.Type: GrantFiled: August 26, 2021Date of Patent: July 23, 2024Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Fatima Zohra Daha, Robert Fernand Gordan, Amit Srivastava, Joshua Alexander Doctors
-
Patent number: 12041047Abstract: Disclosed are various approaches for performing biometric authentication of users using an application running on a client device. A biometric model can be trained using biometric data from a population of users. The biometric model can be used by the client application to authenticate users and can be separate from system-level biometric authentication capabilities of the client device.Type: GrantFiled: September 23, 2021Date of Patent: July 16, 2024Assignee: VMware LLCInventors: Amit Gupta, Gaurav Arora, Vinayak Srivastava, Nitish Kumar Rai
-
Patent number: 12039435Abstract: An apparatus to facilitate acceleration of machine learning operations is disclosed. The apparatus comprises at least one processor to perform operations to implement a neural network and accelerator logic to perform communicatively coupled to the processor to perform compute operations for the neural network.Type: GrantFiled: June 21, 2022Date of Patent: July 16, 2024Assignee: INTEL CORPORATIONInventors: Amit Bleiweiss, Anavai Ramesh, Asit Mishra, Deborah Marr, Jeffrey Cook, Srinivas Sridharan, Eriko Nurvitadhi, Elmoustapha Ould-Ahmed-Vall, Dheevatsa Mudigere, Mohammad Ashraf Bhuiyan, Md Faijul Amin, Wei Wang, Dhawal Srivastava, Niharika Maheshwari
-
Patent number: 12032922Abstract: Automatic generation of intelligent content is created using a system of computers including a user device and a cloud-based component that processes the user information. The system performs a process that includes receiving an input document and parsing the input document to generate inputs for a natural language generation model using a text analysis model. The natural language generation model generates one or more candidate presentation scripts based on the inputs. A presentation script is selected from the candidate presentation scripts and displayed. A text-to-speech model may be used to generate a synthesized audio presentation of the presentation script. A final presentation may be generated that includes a visual display of the input document and the corresponding audio presentation in sync with the visual display.Type: GrantFiled: May 12, 2021Date of Patent: July 9, 2024Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Ji Li, Konstantin Seleskerov, Huey-Ru Tsai, Muin Barkatali Momin, Ramya Tridandapani, Sindhu Vigasini Jambunathan, Amit Srivastava, Derek Martin Johnson, Gencheng Wu, Sheng Zhao, Xinfeng Chen, Bohan Li
-
Patent number: 12026948Abstract: Techniques performed by a data processing system include establishing an online presentation session for conducting an online presentation, receiving first media streams comprising presentation content from the first computing device, receiving second media streams from the second computing devices of a subset of the plurality of participants, the second media streams including audio content, video content, or both of the subset of the plurality of participants, analyzing the first media streams using first machine learning models to generate feedback results, analyzing the set of second media streams to identify first reactions by the participants to obtain reaction information, automatically analyzing the feedback results and the reactions to identify discrepancies between the feedback results and the reactions, and automatically updating one or more parameters of the machine learning models based on the discrepancies to improve the suggestions for improving the online presentation.Type: GrantFiled: October 30, 2020Date of Patent: July 2, 2024Assignee: Microsoft Technology Licensing, LLCInventors: Konstantin Seleskerov, Amit Srivastava, Derek Martin Johnson, Priyanka Vikram Sinha, Gencheng Wu, Brittany Elizabeth Mederos
-
Patent number: 12020701Abstract: Methods, systems, and computer programs are presented for detecting a mission changes in a conversation. A user utterance from a user device is received. The user utterance is part of a conversation with an intelligent assistant. The conversation includes preceding user utterances in pursuit of a first mission. It is determined that the user utterance indicates a mission change from the first mission to a second mission based on an application of a machine-learned model to the user utterance and the preceding user utterances. The machine-learned model has been trained repeatedly with past utterances of other users over a time period, the determining based on a certainty of the indication satisfying a certainty threshold. Responsive to the determining that the user utterance indicates the mission change from the first mission to a second mission, a reply to the user utterance is generated to further the second mission rather than the first mission.Type: GrantFiled: September 30, 2021Date of Patent: June 25, 2024Assignee: EBAY INC.Inventors: Stefan Schoenmackers, Amit Srivastava, Lawrence William Colagiovanni, Sanjika Hewavitharana, Ajinkya Gorakhnath Kale, Vinh Khuc
-
Patent number: 12020683Abstract: A real-time name mispronunciation detection feature can enable a user to receive instant feedback anytime they have mispronounced another person's name in an online meeting. The feature can receive audio input of a speaker and obtain a transcript of the audio input; identify a name from text of the transcript based on names of meeting participants; and extract a portion of the audio input corresponding to the name identified from the text of the transcript. The feature can obtain a reference pronunciation for the name using a user identifier associated with the name; and can obtain a pronunciation score for the name based on a comparison between the reference pronunciation for the name and the portion of the audio input corresponding to the name. The feature can then determine whether the pronunciation score is below a threshold; and in response, notify the speaker of a pronunciation error.Type: GrantFiled: October 28, 2021Date of Patent: June 25, 2024Assignee: Microsoft Technology Licensing, LLCInventors: Tapan Bohra, Akshay Mallipeddi, Amit Srivastava, Ana Karen Parra
-
Patent number: 12001514Abstract: The present disclosure relates to processing operations that execute image classification training for domain-specific traffic, where training operations are entirely compliant with data privacy regulations and policies. Image classification model training, as described herein, is configured to classify meaningful image categories in domain-specific scenarios where there is unknown data traffic and strict data compliance requirements that result in privacy-limited image data sets. Iterative image classification training satisfies data compliance requirements through a combination of online image classification training and offline image classification training. This results in tuned image recognition classifiers that have improved accuracy and efficiency over general image recognition classifiers when working with domain-specific data traffic. One or more image recognition classifiers are independently trained and tuned to detect an image class for image classification.Type: GrantFiled: October 18, 2022Date of Patent: June 4, 2024Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Ji Li, Youjun Liu, Amit Srivastava
-
Publication number: 20240169151Abstract: Understanding emojis in the context of online experiences is described. In at least some embodiments, text input is received and a vector representation of the text input is computed. Based on the vector representation, one or more emojis that correspond to the vector representation of the text input are ascertained and a response is formulated that includes at least one of the one or more emojis. In other embodiments, input from a client machine is received. The input includes at least one emoji. A computed vector representation of the emoji is used to look for vector representations of words or phrases that are close to the computed vector representation of the emoji. At least one of the words or phrases is selected and at least one task is performed using the selected word(s) or phrase(s).Type: ApplicationFiled: January 31, 2024Publication date: May 23, 2024Applicant: eBay Inc.Inventors: Dishan Gupta, Ajinkya Gorakhnath Kale, Stefan Boyd Schoenmackers, Amit Srivastava
-
Publication number: 20240095490Abstract: Aspect pre-selection techniques using machine learning are described. In one example, an artificial assistant system is configured to implement a chat bot. A user then engages in a first natural-language conversation. As part of this first natural-language conversation, a communication is generated by the chat bot to prompt the user to specify an aspect of a category that is a subject of a first natural-language conversation and user data is received in response. Data that describes this first natural-language conversation is used to train a model using machine learning. Data is then received by the chat bot as part of a second natural-language conversation. This data, from the second natural-language conversation, is processed using the model as part of machine learning to generate the second search query to include the aspect of the category automatically and without user intervention.Type: ApplicationFiled: November 29, 2023Publication date: March 21, 2024Applicant: eBay Inc.Inventors: Farah Abdallah, Robert Enyedi, Amit Srivastava, Elaine Lee, Braddock Craig Gaskill, Tomer Lancewicki, Xinyu Zhang, Jayanth Vasudevan, Dominique Jean Bouchon
-
Patent number: 11928428Abstract: Understanding emojis in the context of online experiences is described. In at least some embodiments, text input is received and a vector representation of the text input is computed. Based on the vector representation, one or more emojis that correspond to the vector representation of the text input are ascertained and a response is formulated that includes at least one of the one or more emojis. In other embodiments, input from a client machine is received. The input includes at least one emoji. A computed vector representation of the emoji is used to look for vector representations of words or phrases that are close to the computed vector representation of the emoji. At least one of the words or phrases is selected and at least one task is performed using the selected word(s) or phrase(s).Type: GrantFiled: March 13, 2023Date of Patent: March 12, 2024Assignee: eBay Inc.Inventors: Dishan Gupta, Ajinkya Gorakhnath Kale, Stefan Boyd Schoenmackers, Amit Srivastava
-
Patent number: 11909922Abstract: The present disclosure relates to processing operations configured to provide processing that automatically analyzes acoustic signals from attendees of a live presentation and automatically triggers corresponding reaction indications from results of analysis thereof. Exemplary reaction indications provide feedback for live presentations that can be presented in real-time (or near real-time) without requiring a user to manually take action to provide any feedback. As a non-limiting example, reaction indications may be presented in a form that is easy to visualize and understand such as emojis or icons. Another example of a reaction indication is a graphical user interface (GUI) notification that provides a predictive indication of user intent derived from analysis of acoustic signals.Type: GrantFiled: January 18, 2023Date of Patent: February 20, 2024Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Ji Li, Amit Srivastava, Derek Martin Johnson, Priyanka Vikram Sinha, Konstantin Seleskerov, Gencheng Wu
-
Publication number: 20240050572Abstract: The present disclosure provides polysorbate 20 compositions with particular fatty acid ester concentrations. In some embodiments, they may be used in pharmaceutical formulations, for example, to improve stability.Type: ApplicationFiled: October 26, 2023Publication date: February 15, 2024Applicant: Genentech, Inc.Inventors: Sandeep Yadav, Nidhi Doshi, Tomanna Shobha, Anthony Tomlinson, Amit Srivastava
-
Patent number: 11900052Abstract: The present disclosure applies trained artificial intelligence (AI) processing adapted to automatically generating transformations of formatted templates. Pre-existing formatted templates (e.g., slide-based presentation templates) are leveraged by the trained AI processing to automatically generate a plurality of high-quality template transformations. In transforming a formatted template, the trained AI processing not only generates feature transformation of objects thereof but may also provide style transformations where attributes associated with a presentation theme may be modified for a formatted template or set of formatted templates. The trained AI processing is novel in that it is tailored for analysis of feature data of a specific type of formatted template.Type: GrantFiled: November 11, 2020Date of Patent: February 13, 2024Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Ji Li, Amit Srivastava, Mingxi Cheng
-
Patent number: 11875241Abstract: Aspect pre-selection techniques using machine learning are described. In one example, an artificial assistant system is configured to implement a chat bot. A user then engages in a first natural-language conversation. As part of this first natural-language conversation, a communication is generated by the chat bot to prompt the user to specify an aspect of a category that is a subject of a first natural-language conversation and user data is received in response. Data that describes this first natural-language conversation is used to train a model using machine learning. Data, is then be received by the chat bot as part of a second natural-language conversation. This data, from the second natural-language conversation, is processed using the model as part of machine learning to generate the second search query to include the aspect of the category automatically and without user intervention.Type: GrantFiled: August 31, 2021Date of Patent: January 16, 2024Assignee: eBay Inc.Inventors: Farah Abdallah, Robert Enyedi, Amit Srivastava, Elaine Lee, Braddock Craig Gaskill, Tomer Lancewicki, Xinyu Zhang, Jayanth Vasudevan, Dominique Jean Bouchon
-
Publication number: 20240013790Abstract: A method and system for enhancing pronunciation during a speech, the method including receiving audio data, the audio data including a speech, performing at least one of acoustic scoring and language scoring on the speech, determining a pronunciation score of one or more words of the speech based on the acoustic scoring and the language scoring, determining that the pronunciation score for the word does not satisfy a threshold score, responsive to determining that the pronunciation score does satisfy the threshold score, identifying the word as mispronounced, and responsive to identifying the word as mispronounced, outputting the word and the pronunciation score thereof.Type: ApplicationFiled: May 28, 2021Publication date: January 11, 2024Applicant: Microsoft Technology Licensing, LLCInventors: Runnan LI, Sheng ZHAO, Amit SRIVASTAVA, Huakai LIAO, Ana PARRA, Tapan BOHRA, Akshay MALLIPEDDI, Siliang KANG, Lisha MA, Yinhe WEI
-
Patent number: 11841911Abstract: A data processing system implements receiving query text for a search query for textual content recommendation. The query text includes one or more words indicating a type of textual content items being sought. The system implements analyzing the query text using a first machine learning (ML) model to obtain encoded query text, where the first ML model is trained to identify features within the query text and to generate the encoded query text by mapping the features to a hyper-dimensional latent space (HDLS). The system implements identifying one or more content items in a database of encoded content items mapped to the HDLS that satisfy the search query by comparing attributes of the encoded query text with attributes of the encoded content items to identify content items that are closest to the encoded query text within the HDLS, and causing the one or more content items to be displayed.Type: GrantFiled: November 19, 2021Date of Patent: December 12, 2023Assignee: Microsoft Technology Licensing, LLCInventors: Ji Li, Amit Srivastava, Adit Krishnan, Aman Malik
-
Publication number: 20230395064Abstract: A computing apparatus comprises one or more computer readable storage media, one or more processors operatively coupled with the one or more computer readable storage media, and program instructions stored on the one or more computer readable storage media. The program instructions, when executed by the one or more processors, direct the computing apparatus to at least generate an audio recording of speech, extract features from the audio recording indicative of vocal patterns in the speech, determine a register classification of the speech based at least on the features, and display an indication of the register classification in a user interface.Type: ApplicationFiled: June 7, 2022Publication date: December 7, 2023Inventors: Huakai LIAO, Ana PARRA, Gaurav Vinayak TENDOLKAR, Amit SRIVASTAVA, Siliang KANG
-
Patent number: 11833206Abstract: The present disclosure provides polysorbate 20 compositions with particular fatty acid ester concentrations. In some embodiments, they may be used in pharmaceutical formulations, for example, to improve stability.Type: GrantFiled: September 23, 2021Date of Patent: December 5, 2023Assignee: Genentech, Inc.Inventors: Sandeep Yadav, Nidhi Doshi, Tamanna Shobha, Anthony Tomlinson, Amit Srivastava