Patents by Inventor Sandeep Reddy Goli
Sandeep Reddy Goli has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20230289225Abstract: A scalable Internet of Things (IoT) system may include multiple instances of an IoT manager, each instance respectively configured to connect to a respective edge system of multiple edge systems. The IoT system may further include a containerized system configured to allow any instance of the IoT manager to deploy data pipelines to any edge system of the multiple edge systems in delta communications. Any instance of the IoT manager may send a change message to any edge system via a publish/subscribe notification method. In some examples, a centralized IoT manager may form a secure communication with an edge system, synchronize an object model with an edge object model for the edge system, and maintain the edge system using delta change communications. The IoT system may facilitate any instance of the IoT manager to subscribe a communication channel with an associated edge system for receiving update notification.Type: ApplicationFiled: March 16, 2023Publication date: September 14, 2023Applicant: NUTANIX, INC.Inventors: Naorem Khogendro Singh, Sandeep Reddy Goli, Satyam B. Vaghani, Shyan-Ming Perng
-
Patent number: 11635990Abstract: A scalable Internet of Things (IoT) system may include multiple instances of an IoT manager, each instance respectively configured to connect to a respective edge system of multiple edge systems. The IoT system may further include a containerized system configured to allow any instance of the IoT manager to deploy data pipelines to any edge system of the multiple edge systems in delta communications. Any instance of the IoT manager may send a change message to any edge system via a publish/subscribe notification method. In some examples, a centralized IoT manager may form a secure communication with an edge system, synchronize an object model with an edge object model for the edge system, and maintain the edge system using delta change communications. The IoT system may facilitate any instance of the IoT manager to subscribe a communication channel with an associated edge system for receiving update notification.Type: GrantFiled: October 28, 2019Date of Patent: April 25, 2023Assignee: Nutanix, Inc.Inventors: Naorem Khogendro Singh, Sandeep Reddy Goli, Satyam B. Vaghani, Shyan-Ming Perng
-
Publication number: 20220083389Abstract: Systems and methods described herein generally relate to compute node resource scheduling. AI inference services described herein may receive a request to execute a machine learning model in a clustered edge system. To determine which hardware resource comprising computing nodes of the clustered edge system on which to execute the machine learning model, AI inference services may compare the computational workload of the machine learning model, with the computational abilities and functions of the hardware resources. In examples, the comparison is based on a scheduling algorithm, including an identification stage to identify candidate hardware resources capable of executing the machine learning model, and a scoring stage to select the best candidate hardware resource for executing the machine learning model. A scheduler may assign the machine learning model to the selected hardware resource for execution by the AI inference services.Type: ApplicationFiled: June 17, 2021Publication date: March 17, 2022Applicant: Nutanix, Inc.Inventors: Gaurav Poothia, Sandeep Reddy Goli, Deepak Muley, Pranav Desai
-
Publication number: 20210160338Abstract: A centralized Internet of Things (IoT) manager of an IoT system is configured generate and deploy a data pipeline application to an edge system. The centralized IoT manager is configured to receive a request for a data category and identification of a data transformation function, identify a data source of the IoT system belonging to the data category, and identify an edge system of the IoT system associated with the data source. The centralized IoT manager is further configured to generate a containerized data pipeline application based on a configuration of the edge system that is configured apply the data transformation function to input data to provide transformed data, and provide the containerized data pipeline application to the edge system.Type: ApplicationFiled: January 5, 2021Publication date: May 27, 2021Applicant: Nutanix, Inc.Inventors: Heiko Friedrich Koehler, Keshav Sai Srinivas Nanduri, Sandeep Reddy Goli, Satyam Vaghani
-
Patent number: 10893116Abstract: A centralized Internet of Things (IoT) manager of an IoT system is configured generate and deploy a data pipeline application to an edge system. The centralized IoT manager is configured to receive a request for a data category and identification of a data transformation function, identify a data source of the IoT system belonging to the data category, and identify an edge system of the IoT system associated with the data source. The centralized IoT manger is further configured to generate a containerized data pipeline application based on a configuration of the edge system that is configured apply the data transformation function to input data to provide transformed data, and provide the containerized data pipeline application to the edge system.Type: GrantFiled: July 30, 2019Date of Patent: January 12, 2021Assignee: Nutanix, Inc.Inventors: Heiko Friedrich Koehler, Keshav Sai Srinivas Nanduri, Sandeep Reddy Goli, Satyam Vaghani
-
Publication number: 20210006636Abstract: A centralized Internet of Things (IoT) manager of an IoT system is configured generate and deploy a data pipeline application to an edge system. The centralized IoT manager is configured to receive a request for a data category and identification of a data transformation function, identify a data source of the IoT system belonging to the data category, and identify an edge system of the IoT system associated with the data source. The centralized IoT manger is further configured to generate a containerized data pipeline application based on a configuration of the edge system that is configured apply the data transformation function to input data to provide transformed data, and provide the containerized data pipeline application to the edge system.Type: ApplicationFiled: July 30, 2019Publication date: January 7, 2021Applicant: Nutanix, Inc.Inventors: Heiko Friedrich Koehler, Keshav Sai Srinivas Nanduri, Sandeep Reddy Goli, Satyam Vaghani
-
Publication number: 20210004270Abstract: A scalable Internet of Things (IoT) system may include multiple instances of an IoT manager, each instance respectively configured to connect to a respective edge system of multiple edge systems, The IoT system may further include a containerized system configured to allow any instance of the IoT manager to deploy data pipelines to any edge system of the multiple edge systems in delta communications. Any instance of the IoT manager may send a change message to any edge system via a publish/subscribe notification method. In some examples, a centralized IoT manager may form a secure communication with an edge system, synchronize an object model with an edge object model for the edge system, and maintain the edge system using delta change communications. The IoT system may facilitate any instance of the IoT manager to subscribe a communication channel with an associated edge system for receiving update notification.Type: ApplicationFiled: October 28, 2019Publication date: January 7, 2021Applicant: NUTANIX, INC.Inventors: Naorem Khogendro Singh, Sandeep Reddy Goli, Satyam B. Vaghani, Shyan-Ming Perng
-
Publication number: 20200356415Abstract: An example edge system of an Internet of Things system may include a memory configured to store a machine learning (ML) model application having a ML model a machine, and a processor configured to cause a ML inference service to receive a request for an inference from a ML model application having a ML model, and load the ML model application from the memory into an inference engine in response to the request. The processor is further configured to cause the MT inference service to select a runtime environment from the ML model application to execute the ML model based on a hardware configuration of the edge system, and execute the ML model using the selected to provide inference results. The inference results are provided at an output, such as to a data plane or to be stored in the memory.Type: ApplicationFiled: July 25, 2019Publication date: November 12, 2020Applicant: Nutanix, Inc.Inventor: Sandeep Reddy Goli
-
Publication number: 20180121237Abstract: Performance of a virtual machine system is improved by avoiding and/or eliminating bottlenecks in read and write operations. The system analyzes current virtualized workloads and provides working set estimates for individual VMs, hosts, and clusters. The working set estimate data is then utilized to make specific recommendations for different types of backend storage technologies. After procuring a storage device, the system provides a variety of information to aid in the operation of the system. From this information, the system can detect various scenarios and proactively make recommendations to the user about ways in which to improve storage performance at a host level and at a per-VM level. In some embodiments, these recommendations may be implemented automatically without user involvement.Type: ApplicationFiled: October 24, 2017Publication date: May 3, 2018Inventors: Bryan Jeffrey Crowe, Satyam B. Vaghani, Akhilesh Joshi, Shyan-Ming Perng, Snehal Mundle, Chethan Kumar, Sandeep Reddy Goli, Prashant Saxena
-
Publication number: 20140149264Abstract: The present invention provides an apparatus and process to share digital images and videos of a user wearing virtual apparel. The invention comprises of a camera 201 for capturing images and videos; a central processing unit (CPU) that obtains camera media feed and processes it to augment digital imagery of apparel; the CPU being configured to track a user in the media feed; a display screen 203 that displays the processed media feed; an interne adapter 204 capable of connecting to the interne; the CPU configured to upload the processed media feed online to a server 205 and further send the web location of the uploaded image or video, preferably in a text message using a cellular network 206 to the user's mobile phone 207. The user 251 can share the text message with others enabling them to view the uploaded content on an interne enabled device through a private link or through a social networking platform.Type: ApplicationFiled: June 14, 2012Publication date: May 29, 2014Inventors: Hemanth Kumar Satyanarayana, Sandeep Reddy Goli