Patents by Inventor Venkappa MALA

Venkappa MALA has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20250078337
    Abstract: A method for generating content using a diffusion model of an electronic device, may include: obtaining latent vectors of an input content; inputting the latent vectors into a first lightweight adapter configured for the first application type from among a plurality of lightweight adapters configured individually for application types of the plurality of applications; transforming the latent vectors of the input content into a plurality of intermediate latent vectors using the first lightweight adapter; performing a denoising operation to transform the plurality of intermediate latent vectors into a plurality of next operation vectors; and generating the final content belonging to the application type by decoding the next operation vectors; and outputting the final content.
    Type: Application
    Filed: August 30, 2024
    Publication date: March 6, 2025
    Applicant: SAMSUNG ELECTRONICS., LTD.
    Inventors: Prateek KESERWANI, Sikumar MOHARANA, Alladi Ashok Kumar SENAPATI, Sajith UMMANATH, Azhan ALI, Venkappa MALA
  • Patent number: 12236331
    Abstract: A method of deep neural network (DNN) modularization for optimal loading includes receiving, by an electronic device, a DNN model for execution, obtaining, by the electronic device, a plurality of parameters associated with the electronic device and a plurality of parameters associated with the DNN model, determining, by the electronic device, a number of sub-models of the DNN model and a splitting index, based on the obtained plurality of parameters associated with the electronic device and the obtained plurality of parameters associated with the DNN model, and splitting, by the electronic device, the received DNN model into a plurality of sub-models, based on the determined number of sub-models of the DNN model and the determined splitting index.
    Type: Grant
    Filed: July 9, 2021
    Date of Patent: February 25, 2025
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Brijraj Singh, Mayukh Das, Yash Hemant Jain, Sharan Kumar Allur, Venkappa Mala, Praveen Doreswamy Naidu
  • Publication number: 20240232587
    Abstract: A method and an electronic device for neuro-symbolic learning of an artificial intelligence (AI) model are provided. The method includes receiving input data including various contents and determining in an output of the AI model a predicted probability for each of the contents of the input data, determining a neural loss of the AI model by comparing the predicted probability with a predefined desired probability, determining a symbolic loss for the AI model by comparing the predicted probability with a pre-determined undesired probability, determining weights of a plurality of layers of the AI model, and updating the weights of the plurality of layers of the AI model based on the neural loss and the symbolic loss.
    Type: Application
    Filed: February 1, 2024
    Publication date: July 11, 2024
    Inventors: Srinivas Soumitri MIRIYALA, Efthymia TSAMOURA, Shah Ayub QUADRI, Vikram Nelvoy RAJENDIRAN, Venkappa MALA
  • Patent number: 12020146
    Abstract: A method of processing a neural network model by using a plurality of processors includes allocating at least one slice to each layer from among a plurality of layers included in the neural network model, allocating each layer from among the plurality of layers to the plurality of processors based on respective processing times of the plurality of processors for processing each of the at least one slice, and processing the neural network model by using the plurality of processors based on a result of the allocation.
    Type: Grant
    Filed: August 23, 2019
    Date of Patent: June 25, 2024
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Manas Sahni, Arun Abraham, Sharan Kumar Allur, Venkappa Mala
  • Patent number: 11740941
    Abstract: The present invention describes a method of accelerating execution of one or more application tasks in a computing device using machine learning (ML) based model. According to one embodiment, a neural accelerating engine present in the computing device receives a ML input task for execution on the computing device from a user. The neural accelerating engine further retrieves a trained ML model and a corresponding optimal configuration file based on the received ML input task. Also, the current performance status of the computing device for executing the ML input task is obtained. Then, the neural accelerating engine dynamically schedules and dispatches parts of the ML input task to one or more processing units in the computing device for execution based on the retrieved optimal configuration file and the obtained current performance status of the computing device.
    Type: Grant
    Filed: February 23, 2018
    Date of Patent: August 29, 2023
    Inventors: Arun Abraham, Suhas Parlathaya Kudral, Balaji Srinivas Holur, Sarbojit Ganguly, Venkappa Mala, Suneel Kumar Surimani, Sharan Kumar Allur
  • Publication number: 20230153565
    Abstract: A method of deep neural network (DNN) modularization for optimal loading includes receiving, by an electronic device, a DNN model for execution, obtaining, by the electronic device, a plurality of parameters associated with the electronic device and a plurality of parameters associated with the DNN model, determining, by the electronic device, a number of sub-models of the DNN model and a splitting index, based on the obtained plurality of parameters associated with the electronic device and the obtained plurality of parameters associated with the DNN model, and splitting, by the electronic device, the received DNN model into a plurality of sub-models, based on the determined number of sub-models of the DNN model and the determined splitting index.
    Type: Application
    Filed: July 9, 2021
    Publication date: May 18, 2023
    Applicant: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Brijraj SINGH, Mayukh DAS, Yash Hemant JAIN, Sharan Kumar ALLUR, Venkappa MALA, Praveen Doreswamy NAIDU
  • Publication number: 20230127001
    Abstract: A method for generating an optimal neural network (NN) model may include determining intermediate outputs of the NN model by passing an input dataset through each intermediate exit gate of the plurality of intermediate exit gates, determining an accuracy score for each intermediate exit gate of the plurality of intermediate exit gates based on a comparison of the final output of the NN model with the intermediate output, identifying an earliest intermediate exit gate that produces the intermediate output closer to the final output based on the accuracy score, and generating the optimal NN model by removing remaining layers of the plurality of layers and remaining intermediate exit gates of the plurality of intermediate exit gates located after the determined earliest intermediate exit gate.
    Type: Application
    Filed: December 15, 2022
    Publication date: April 27, 2023
    Applicant: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Mayukh DAS, Brijraj SINGH, Pradeep NELAHONNE SHIVAMURTHAPPA, Aakash KAPOOR, Rajath Elias SOANS, Soham Vijay DIXIT, Sharan Kumar ALLUR, Venkappa MALA
  • Publication number: 20220066829
    Abstract: Disclosed herein is a method and an optimization unit for optimizing and/or improving efficiency of resource utilization in an embedded computing system executing Artificial Intelligence (AI) applications. The method includes: detecting, by an optimization unit comprising processing circuitry and/or executable program instructions configured in the embedded computing system, a launch of an AI application on the embedded computing system; retrieving a runtime profile corresponding to the AI application, the runtime profile indicating resource requirements for executing the AI application; and configuring a runtime environment of the embedded computing system for the AI application based on the runtime profile corresponding to the AI application.
    Type: Application
    Filed: November 8, 2021
    Publication date: March 3, 2022
    Inventors: Ashutosh PAVAGADA VISWESWARA, Pallavi THUMMALA, Chirag GIRDHAR, Alladi Ashok Kumar SENAPATI, Pradeep N S NELAHONNE SHIVAMURTHAPPA, Venkappa MALA
  • Publication number: 20210350203
    Abstract: Embodiments herein provide a NAS method of generating an optimized DNN model for executing a task in an electronic device. The method includes identifying the task to be executed in the electronic device. The method includes estimating a performance parameter to be achieved while executing the task. The method includes determining hardware parameters of the electronic device required to execute the task based on the performance parameter and the task, and determining optimal neural blocks from a plurality of neural blocks based on the performance parameter and the hardware parameter of the electronic device. The method includes generating the optimized DNN model for executing the task based on the optimal neural blocks, and executing the task using the optimized DNN model.
    Type: Application
    Filed: March 24, 2021
    Publication date: November 11, 2021
    Inventors: Mayukh Das, Venkappa Mala, Brijraj Singh, Pradeep Nelahonne Shivamurthappa, Sharan Kumar Allur
  • Publication number: 20210232921
    Abstract: A method, an apparatus, and a system for configuring a neural network across heterogeneous processors are provided. The method includes creating a unified neural network profile for the plurality of processors; receiving at least one request to perform at least one task using the neural network; determining a type of the requested at least one task as one of an asynchronous task and a synchronous task; and parallelizing processing of the neural network across the plurality of processors to perform the requested at least one task, based on the type of the requested at least one task and the created unified neural network profile.
    Type: Application
    Filed: January 27, 2021
    Publication date: July 29, 2021
    Applicant: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Akshay PARASHAR, Arun ABRAHAM, Payal ANAND, Deepthy RAVI, Venkappa MALA, Vikram Nelvoy RAJENDIRAN
  • Publication number: 20200065671
    Abstract: A method of processing a neural network model by using a plurality of processors includes allocating at least one slice to each layer from among a plurality of layers included in the neural network model, allocating each layer from among the plurality of layers to the plurality of processors based on respective processing times of the plurality of processors for processing each of the at least one slice, and processing the neural network model by using the plurality of processors based on a result of the allocation.
    Type: Application
    Filed: August 23, 2019
    Publication date: February 27, 2020
    Applicant: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Manas SAHNI, Arun Abraham, Sharan Kumar Allur, Venkappa Mala
  • Publication number: 20200019854
    Abstract: The present invention describes a method of accelerating execution of one or more application tasks in a computing device using machine learning (ML) based model. According to one embodiment, a neural accelerating engine present in the computing device receives a ML input task for execution on the computing device from a user. The neural accelerating engine further retrieves a trained ML model and a corresponding optimal configuration file based on the received ML input task. Also, the current performance status of the computing device for executing the ML input task is obtained. Then, the neural accelerating engine dynamically schedules and dispatches parts of the ML input task to one or more processing units in the computing device for execution based on the retrieved optimal configuration file and the obtained current performance status of the computing device.
    Type: Application
    Filed: February 23, 2018
    Publication date: January 16, 2020
    Inventors: Arun ABRAHAM, Suhas Parlathaya KUDRAL, Balaji Srinivas HOLUR, Sarbojit GANGULY, Venkappa MALA, Suneel Kumar SURIMANI, Sharan Kumar ALLUR