Patents by Inventor Sai Rahul Chalamalasetti
Sai Rahul Chalamalasetti has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20240111993Abstract: Systems and methods are provided for performing object store offloading. A user query can be received from a client device to access a data object. The semantic structure associated with the data object can be identified, as well as one or more relationships associated with the semantic structure of the data object. A view of the data object can be determined based on the one or more relationships and said view can be provided to a user interface.Type: ApplicationFiled: September 30, 2022Publication date: April 4, 2024Inventors: DARIO KOROLIJA, Kun Wu, Sai Rahul Chalamalasetti, Lance Mackimmie Evans, Dejan S. Milojicic
-
Publication number: 20240111970Abstract: In some examples, a device includes a first processing core comprising a resistive memory array to perform an analog computation, and a digital processing core comprising a digital memory programmable with different values to perform different computations responsive to respective different conditions. The device further includes a controller to selectively apply input data to the first processing core and the digital processing core.Type: ApplicationFiled: December 4, 2023Publication date: April 4, 2024Inventors: John Paul Strachan, Dejan S. Milojicic, Martin Foltin, Sai Rahul Chalamalasetti, Amit S. Sharma
-
Publication number: 20240112029Abstract: A crossbar array includes a number of memory elements. An analog-to-digital converter (ADC) is electronically coupled to the vector output register. A digital-to-analog converter (DAC) is electronically coupled to the vector input register. A processor is electronically coupled to the ADC and to the DAC. The processor may be configured to determine whether division of input vector data by output vector data from the crossbar array is within a threshold value, and if not within the threshold value, determine changed data values as between the output vector data and the input vector data, and write the changed data values to the memory elements of the crossbar array.Type: ApplicationFiled: December 5, 2023Publication date: April 4, 2024Inventors: Sai Rahul Chalamalasetti, Paolo Faraboschi, Martin Foltin, Catherine Graves, Dejan S. Milojicic, John Paul Strachan, Sergey Serebryakov
-
Patent number: 11947928Abstract: Systems and methods are provided for a multi-die dot-product engine (DPE) to provision large-scale machine learning inference applications. The multi-die DPE leverages a multi-chip architecture. For example, a multi-chip interface can include a plurality of DPE chips, where each DPE chip performs inference computations for performing deep learning operations. A hardware interface between a memory of a host computer and the plurality of DPE chips communicatively connects the plurality of DPE chips to the memory of the host computer system during an inference operation such that the deep learning operations are spanned across the plurality of DPE chips. Due to the multi-die architecture, multiple silicon devices are allowed to be used for inference, thereby enabling power-efficient inference for large-scale machine learning applications and complex deep neural networks.Type: GrantFiled: September 10, 2020Date of Patent: April 2, 2024Assignee: Hewlett Packard Enterprise Development LPInventors: Craig Warner, Eun Sub Lee, Sai Rahul Chalamalasetti, Martin Foltin
-
Patent number: 11923899Abstract: Examples described herein relate to a method for synchronizing a wavelength of light in an optical device. In some examples, a heater voltage may be predicted for a heater disposed adjacent to the optical device in a photonic chip. The predicted heater voltage may be applied to the heater to cause a change in the wavelength of the light inside the optical device. In response to applying the heater voltage, an optical power inside the optical device may be measured. Further, a check may be performed to determine whether the measured optical power is a peak optical power. If it is determined that measured optical power is the peak optical power, the application of the predicted heater voltage to the heater may be continued.Type: GrantFiled: December 1, 2021Date of Patent: March 5, 2024Assignee: Hewlett Packard Enterprise Development LPInventors: Hyunmin Jeong, Sai Rahul Chalamalasetti, Marco Fiorentino, Peter Jin Rhim
-
Patent number: 11868855Abstract: In exemplary aspects, a golden data structure can be used to validate the stability of machine learning (ML) models and weights. The golden data structure includes golden input data and corresponding golden output data. The golden output data represents the known correct results that should be output by a ML model when it is run with the golden input data as inputs. The golden data structure can be stored in a secure memory and retrieved for validation separately or together with the deployment of the ML model for a requested ML operation. If the golden data structure is used to validate the model and/or weights concurrently with the performance of the requested operation, the golden input data is combined with the input data for the requested operation and run through the model. Relevant outputs are compared with the golden output data to validate the stability of the model and weights.Type: GrantFiled: November 4, 2019Date of Patent: January 9, 2024Assignee: Hewlett Packard Enterprise Development LPInventors: Sai Rahul Chalamalasetti, Sergey Serebryakov, Dejan S. Milojicic
-
Patent number: 11861429Abstract: In some examples, a device includes a first processing core comprising a resistive memory array to perform an analog computation, and a digital processing core comprising a digital memory programmable with different values to perform different computations responsive to respective different conditions. The device further includes a controller to selectively apply input data to the first processing core and the digital processing core.Type: GrantFiled: April 30, 2018Date of Patent: January 2, 2024Assignee: Hewlett Packard Enterprise Development LPInventors: John Paul Strachan, Dejan S. Milojicic, Martin Foltin, Sai Rahul Chalamalasetti, Amit S. Sharma
-
Patent number: 11853846Abstract: A crossbar array includes a number of memory elements. An analog-to-digital converter (ADC) is electronically coupled to the vector output register. A digital-to-analog converter (DAC) is electronically coupled to the vector input register. A processor is electronically coupled to the ADC and to the DAC. The processor may be configured to determine whether division of input vector data by output vector data from the crossbar array is within a threshold value, and if not within the threshold value, determine changed data values as between the output vector data and the input vector data, and write the changed data values to the memory elements of the crossbar array.Type: GrantFiled: April 30, 2018Date of Patent: December 26, 2023Assignee: Hewlett Packard Enterprise Development LPInventors: Sai Rahul Chalamalasetti, Paolo Faraboschi, Martin Foltin, Catherine Graves, Dejan S. Milojicic, John Paul Strachan, Sergey Serebryakov
-
Publication number: 20230170991Abstract: Examples described herein relate to a method for synchronizing a wavelength of light in an optical device. In some examples, a heater voltage may be predicted for a heater disposed adjacent to the optical device in a photonic chip. The predicted heater voltage may be applied to the heater to cause a change in the wavelength of the light inside the optical device. In response to applying the heater voltage, an optical power inside the optical device may be measured. Further, a check may be performed to determine whether the measured optical power is a peak optical power. If it is determined that measured optical power is the peak optical power, the application of the predicted heater voltage to the heater may be continued.Type: ApplicationFiled: December 1, 2021Publication date: June 1, 2023Inventors: Hyunmin Jeong, Sai Rahul Chalamalasetti, Marco Fiorentino, Peter Jin Rhim
-
Patent number: 11644882Abstract: One embodiment provides a system and method for predicting network power usage associated with workloads. During operation, the system configures a simulator to simulate operations of a plurality of network components, which comprises embedding one or more event counters in each simulated network component. A respective event counter is configured to count a number of network-power-related events. The system collects, based on values of the event counters, network-power-related performance data associated with one or more sample workloads applied to the simulator; and trains a machine-learning model with the collected network-power-related performance data and characteristics of the sample workloads as training data 1, thereby facilitating prediction of network-power-related performance associated with a to-be-evaluated workload.Type: GrantFiled: June 2, 2021Date of Patent: May 9, 2023Assignee: Hewlett Packard Enterprise Development LPInventors: Harumi Kuno, Alan Davis, Torsten Wilde, Daniel William Dauwe, Duncan Roweth, Ryan Dean Menhusen, Sergey Serebryakov, John L. Byrne, Vipin Kumar Kukkala, Sai Rahul Chalamalasetti
-
Publication number: 20220390999Abstract: One embodiment provides a system and method for predicting network power usage associated with workloads. During operation, the system configures a simulator to simulate operations of a plurality of network components, which comprises embedding one or more event counters in each simulated network component. A respective event counter is configured to count a number of network-power-related events. The system collects, based on values of the event counters, network-power-related performance data associated with one or more sample workloads applied to the simulator; and trains a machine-learning model with the collected network-power-related performance data and characteristics of the sample workloads as training data 1, thereby facilitating prediction of network-power-related performance associated with a to-be-evaluated workload.Type: ApplicationFiled: June 2, 2021Publication date: December 8, 2022Inventors: Harumi Kuno, Alan Davis, Torsten Wilde, Daniel William Dauwe, Duncan Roweth, Ryan Dean Menhusen, Sergey Serebryakov, John L. Byrne, Vipin Kumar Kukkala, Sai Rahul Chalamalasetti
-
Patent number: 11443036Abstract: In some examples, an apparatus includes a management controller for use in a computer system having a processing resource for executing an operating system (OS) of the computer system, the management controller being separate from the processing resource and to perform, based on operation of the management controller within a cryptographic boundary, management of components of the computer system, the management of components comprising power control of the computer system. The management controller is to receive sensor data, perform facial recognition based on the sensor data, and determine whether to initiate a security action responsive to the facial recognition.Type: GrantFiled: July 30, 2019Date of Patent: September 13, 2022Assignee: Hewlett Packard Enterprise Development LPInventors: Naysen Robertson, Sai Rahul Chalamalasetti, William James Walker
-
Patent number: 11385863Abstract: Disclosed techniques provide for dynamically changing precision of a multi-stage compute process. For example, changing neural network (NN) parameters on a per-layer basis depending on properties of incoming data streams and per-layer performance of an NN among other considerations. NNs include multiple layers that may each be calculated with a different degree of accuracy and therefore, compute resource overhead (e.g., memory, processor resources, etc.). NNs are usually trained with 32-bit or 16-bit floating-point numbers. Once trained, an NN may be deployed in production. One approach to reduce compute overhead is to reduce parameter precision of NNs to 16 or 8 for deployment. The conversion to an acceptable lower precision is usually determined manually before deployment and precision levels are fixed while deployed.Type: GrantFiled: August 1, 2018Date of Patent: July 12, 2022Assignee: Hewlett Packard Enterprise Development LPInventors: Sai Rahul Chalamalasetti, Paolo Faraboschi, Martin Foltin, Catherine Graves, Dejan S. Milojicic, Sergey Serebryakov, John Paul Strachan
-
Publication number: 20220121885Abstract: Testing for bias in a machine learning (ML) model in a manner that is independent of the code/weights deployment path is described. If bias is detected, an alert for bias is generated, and optionally, the ML model can be incrementally re-trained to mitigate the detected bias. Re-training the ML model to mitigate the bias may include enforcing a bias cost function to maintain a level of bias in the ML model below a threshold bias level. One or more statistical metrics representing the level of bias present in the ML model may be determined and compared against one or more threshold values. If one or more metrics exceed corresponding threshold value(s), the level of bias in the ML model may be deemed to exceed a threshold level of bias, and re-training of the ML model to mitigate the bias may be initiated.Type: ApplicationFiled: October 19, 2020Publication date: April 21, 2022Inventors: Sai Rahul Chalamalasetti, Dejan S. Milojicic, Sergey Serebryakov
-
Publication number: 20220075597Abstract: Systems and methods are provided for a multi-die dot-product engine (DPE) to provision large-scale machine learning inference applications. The multi-die DPE leverages a multi-chip architecture. For example, a multi-chip interface can include a plurality of DPE chips, where each DPE chip performs inference computations for performing deep learning operations. A hardware interface between a memory of a host computer and the plurality of DPE chips communicatively connects the plurality of DPE chips to the memory of the host computer system during an inference operation such that the deep learning operations are spanned across the plurality of DPE chips. Due to the multi-die architecture, multiple silicon devices are allowed to be used for inference, thereby enabling power-efficient inference for large-scale machine learning applications and complex deep neural networks.Type: ApplicationFiled: September 10, 2020Publication date: March 10, 2022Inventors: Craig Warner, Eun Sub Lee, Sai Rahul Chalamalasetti, Martin Foltin
-
Publication number: 20210240945Abstract: In some examples, a device includes a first processing core comprising a resistive memory array to perform an analog computation, and a digital processing core comprising a digital memory programmable with different values to perform different computations responsive to respective different conditions. The device further includes a controller to selectively apply input data to the first processing core and the digital processing core.Type: ApplicationFiled: April 30, 2018Publication date: August 5, 2021Applicant: Hewlett Packard Enterprise Development LPInventors: John Paul STRACHAN, Dejan S. MILOJICIC, Martin FOLTIN, Sai Rahul CHALAMALASETTI, Amit S. SHARMA
-
Publication number: 20210201136Abstract: A crossbar array includes a number of memory elements. An analog-to-digital converter (ADC) is electronically coupled to the vector output register. A digital-to-analog converter (DAC) is electronically coupled to the vector input register. A processor is electronically coupled to the ADC and to the DAC. The processor may be configured to determine whether division of input vector data by output vector data from the crossbar array is within a threshold value, and if not within the threshold value, determine changed data values as between the output vector data and the input vector data, and write the changed data values to the memory elements of the crossbar array.Type: ApplicationFiled: April 30, 2018Publication date: July 1, 2021Inventors: Sai Rahul Chalamalasetti, Paolo Faraboschi, Martin Foltin, Catherine Graves, Dejan S. Milojicic, John Paul Strachan, Sergey Serebryakov
-
Publication number: 20210133624Abstract: In exemplary aspects, a golden data structure can be used to validate the stability of machine learning (ML) models and weights. The golden data structure includes golden input data and corresponding golden output data. The golden output data represents the known correct results that should be output by a ML model when it is run with the golden input data as inputs. The golden data structure can be stored in a secure memory and retrieved for validation separately or together with the deployment of the ML model for a requested ML operation. If the golden data structure is used to validate the model and/or weights concurrently with the performance of the requested operation, the golden input data is combined with the input data for the requested operation and run through the model. Relevant outputs are compared with the golden output data to validate the stability of the model and weights.Type: ApplicationFiled: November 4, 2019Publication date: May 6, 2021Inventors: Sai Rahul Chalamalasetti, Sergey Serebryakov, Dejan S. Milojicic
-
Publication number: 20210034742Abstract: In some examples, an apparatus includes a management controller for use in a computer system having a processing resource for executing an operating system (OS) of the computer system, the management controller being separate from the processing resource and to perform, based on operation of the management controller within a cryptographic boundary, management of components of the computer system, the management of components comprising power control of the computer system. The management controller is to receive sensor data, perform facial recognition based on the sensor data, and determine whether to initiate a security action responsive to the facial recognition.Type: ApplicationFiled: July 30, 2019Publication date: February 4, 2021Inventors: Naysen Robertson, Sai Rahul Chalamalasetti, William James Walker
-
Patent number: 10846138Abstract: A method for allocating resources includes determining that an initial allocation of memory bandwidth for one or more computing jobs fails a performance metric. The memory bandwidth provides access to a global memory pool for multiple legacy processors across a memory fabric. The method also includes determining a new allocation of memory bandwidth for the computing jobs that meets the performance metric. Additionally, the method includes assigning the new allocation of memory bandwidth to the computing jobs. The method further includes executing the computing jobs using the new allocation of memory bandwidth.Type: GrantFiled: August 23, 2018Date of Patent: November 24, 2020Assignee: Hewlett Packard Enterprise Development LPInventors: Zhikui Wang, Antonio Lain, Sai Rahul Chalamalasetti, Anshuman Goswami