Patents by Inventor Mehdi Rezagholizadeh
Mehdi Rezagholizadeh has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 11914670Abstract: Methods and systems for compressing a matrix are described. The matrix, having a plurality of rows formed by a respective plurality of vectors, is partitioned into a plurality of submatrices, each submatrix containing sub-vectors from a respective group of one or more contiguous columns of the matrix. For each given submatrix, the sub-vectors are clustered into a plurality of clusters. For each given cluster, a centroid and a variance are computed and stored, based on the sub-vectors belonging to the given cluster. A mapping relating each vector to a respective cluster in each submatrix is stored. The stored centroids, stored variances and stored mapping form a set of compressed data for reconstruction of the matrix.Type: GrantFiled: September 8, 2020Date of Patent: February 27, 2024Assignee: HUAWEI TECHNOLOGIES CO., LTD.Inventors: Krtin Kumar, Mehdi Rezagholizadeh, Peyman Passban
-
Publication number: 20230222326Abstract: Method and system of training a student neural network (SNN) model. A first training phase is performed over a plurality of epochs during which a smoothing factor to teacher neural network (TNN) model outputs to generate smoothed TNN model outputs, a first loss is computed based on the SNN model outputs and the smoothed TNN model outputs, and an updated set of the SNN model parameters is computed with an objective of reducing the first loss in a following epoch of the first training phase. The soothing factor is adjusted over the plurality of epochs of the first training phase to reduce a smoothing effect on the generated smoothed TNN model outputs. A second training phase is performed based on the SNN model outputs and a set of predefined expected outputs for the plurality of input data samples.Type: ApplicationFiled: March 8, 2023Publication date: July 13, 2023Inventors: Aref JAFARI, Mehdi REZAGHOLIZADEH, Ali GHODSI, Pranav SHARMA
-
Publication number: 20230222353Abstract: Method and system of training a student neural network using adversarial learning and knowledge distillation, including: training a generator to generate adversarial data samples for respective training data samples by masking parts of the training data samples with an objective of maximizing a divergence between output predictions generated by the student neural network and a teacher neural network model for the adversarial data samples; and training the student neural network based on objectives of (i) minimizing a divergence between output predictions generated by the student neural network and the teacher neural network model for the adversarial data samples, and (ii) minimizing a divergence between output predictions generated by the student neural network and the teacher neural network model for the training data samples.Type: ApplicationFiled: March 8, 2023Publication date: July 13, 2023Inventors: Vasileios LIOUTAS, Ahmad RASHID, Mehdi REZAGHOLIZADEH
-
Patent number: 11663483Abstract: According to embodiments, an encoder neural network receives a one-hot representation of a real text. The encoder neural network outputs a latent representation of the real text. A decoder neural network receives random noise data or artificial code generated by a generator neural network from random noise data. The decoder neural network outputs softmax representation of artificial text. The decoder neural network receives the latent representation of the real text. The decoder neural network outputs a reconstructed softmax representation of the real text. A hybrid discriminator neural network receives a first combination of the soft-text and the latent representation of the real text and a second combination of the softmax representation of artificial text and the artificial code. The hybrid discriminator neural network outputs a probability indicating whether the second combination is similar to the first combination. Additional embodiments for utilizing latent representation are also disclosed.Type: GrantFiled: October 30, 2018Date of Patent: May 30, 2023Assignee: Huawei Technologies Co., Ltd.Inventors: Md Akmal Haidar, Mehdi Rezagholizadeh
-
Patent number: 11586833Abstract: A method and machine translation system for bi-directional translation of textual sequences between a first language and a second language are described. The machine translation system includes a first autoencoder configured to receive a vector representation of a first textual sequence in the first language and encode the vector representation of the first textual sequence into a first sentence embedding. The machine translation system also includes a sum-product network (SPN) configured to receive the first sentence embedding and generate a second sentence embedding by maximizing a first conditional probability of the second sentence embedding given the first sentence embedding and a second autoencoder receiving the second sentence embedding, the second autoencoder being trained to decode the second sentence embedding into a vector representation of a second textual sequence in the second language.Type: GrantFiled: June 12, 2020Date of Patent: February 21, 2023Assignee: HUAWEI TECHNOLOGIES CO., LTD.Inventors: Mehdi Rezagholizadeh, Vahid Partovi Nia, Md Akmal Haidar, Pascal Poupart
-
Publication number: 20220366226Abstract: Methods and systems for compressing a neural network which performs an inference task and for performing computations of a Kronecker layer of a Kroenke NN are described. A batch of data samples are obtained from a training dataset. The input data of the data samples are inputted into a trained neural network to forward propagate the input data through the trained neural network and generate neural network predictions for the input data. Further, the input data are inputted into a Kronecker neural network to forward propagate the input data through the Kronecker neural network to generate Kronecker neural network predictions for the input data. Afterwards, two losses are computed: a knowledge distillation loss and a loss for Kronecker layer. The knowledge distillation loss is based on outputs generated by a layer of the neural network and a corresponding Kronecker layer of the Kronecker neural network.Type: ApplicationFiled: May 17, 2021Publication date: November 17, 2022Inventors: Marziehsadat TAHAEI, Ali GHODSI, Mehdi REZAGHOLIZADEH, Vahid PARTOVI NIA
-
Publication number: 20220343139Abstract: Methods and systems for training a neural network model using domain mixing and multi-teacher knowledge distillation are described. Tokens, including a unique token, are inputted to an encoder of the neural network model. A unique embedding vector encoded from the unique token is inputted to an adaptor network to generate domain probabilities. A domain mixing embedding vector, determined from the unique embedding vector, is inputted to a predictor of the neural network model, to generate a predicted output. A final loss is computed using a domain mixing loss computed from the domain probabilities and a ground-truth domain of the data sample, and using an output prediction loss computed from the predicted output and a ground-truth label of the data sample. Parameters of the neural network model and adaptor network are updated using the final loss.Type: ApplicationFiled: April 15, 2021Publication date: October 27, 2022Inventors: Peyman PASSBAN, Amirmehdi SHARIFZAD, Mehdi REZAGHOLIZADEH, Khalil BIBI
-
Publication number: 20220343175Abstract: Methods, devices and processor-readable media for re-weighting to improve knowledge distillation are described. A reweighting module may be used to determine relative weights to assign to a ground truth label and dark knowledge distilled from the teacher (i.e. the teacher output logits used as soft labels). A meta-reweighting method is described to optimize the weights for a given labeled data sample.Type: ApplicationFiled: April 15, 2021Publication date: October 27, 2022Inventors: Peng LU, Ahmad RASHID, Mehdi REZAGHOLIZADEH, Abbas GHADDAR
-
Publication number: 20220335303Abstract: Methods, devices and processor-readable media for knowledge distillation using intermediate representations are described. A student model is trained using a Dropout-KD approach in which intermediate layer selection is performed efficiently such that the skip, search, and overfitting problems in intermediate layer KD may be solved. Teacher intermediate layers are selected randomly at each training epoch, with the layer order preserved to avoid breaking information flow. Over the course of multiple training epochs, all of the teacher intermediate layers are used for knowledge distillation. A min-max data augmentation method is also described based on the intermediate layer selection of the Dropout-KD training method.Type: ApplicationFiled: April 16, 2021Publication date: October 20, 2022Inventors: Md Akmal HAIDAR, Mehdi REZAGHOLIZADEH
-
Patent number: 11423282Abstract: In accordance to embodiments, an encoder neural network is configured to receive a one-hot representation of a real text and output a latent representation of the real text generated from the one-hot representation of the real text. A decoder neural network is configured to receive the latent representation of the real text, and output a reconstructed softmax representation of the real text from the latent representation of the real text, the reconstructed softmax representation of the real text is a soft-text. A generator neural network is configured to generate artificial text based on random noise data. A discriminator neural network is configured to receive the soft-text and receive a softmax representation of the artificial text, and output a probability indicating whether the softmax representation of the artificial text received by the discriminator neural network is not from the generator neural network.Type: GrantFiled: October 30, 2018Date of Patent: August 23, 2022Assignee: Huawei Technologies Co., Ltd.Inventors: Md Akmal Haidar, Mehdi Rezagholizadeh
-
Publication number: 20220075843Abstract: Methods and systems for compressing a matrix are described. The matrix, having a plurality of rows formed by a respective plurality of vectors, is partitioned into a plurality of submatrices, each submatrix containing sub-vectors from a respective group of one or more contiguous columns of the matrix. For each given submatrix, the sub-vectors are clustered into a plurality of clusters. For each given cluster, a centroid and a variance are computed and stored, based on the sub-vectors belonging to the given cluster. A mapping relating each vector to a respective cluster in each submatrix is stored. The stored centroids, stored variances and stored mapping form a set of compressed data for reconstruction of the matrix.Type: ApplicationFiled: September 8, 2020Publication date: March 10, 2022Inventors: Krtin KUMAR, Mehdi REZAGHOLIZADEH, Peyman PASSBAN
-
Publication number: 20220076136Abstract: An agnostic combinatorial knowledge distillation (CKD) method for transferring trained knowledge of neural model from a complex model (teacher) to a less complex model (student) is described. In addition to training the student to generate a final output that approximates both the teacher's final output and a ground truth of a training input, the method further maximizes knowledge transfer by training hidden layers of the student to generate outputs that approximate a representation of a subset of teacher hidden layers are mapped to each of the student hidden layers for a given training input.Type: ApplicationFiled: September 8, 2021Publication date: March 10, 2022Inventors: Peyman PASSBAN, Yimeng WU, Mehdi REZAGHOLIZADEH
-
Publication number: 20210390269Abstract: A method and machine translation system for bi-directional translation of textual sequences between a first language and a second language are described. The machine translation system includes a first autoencoder configured to receive a vector representation of a first textual sequence in the first language and encode the vector representation of the first textual sequence into a first sentence embedding. The machine translation system also includes a sum-product network (SPN) configured to receive the first sentence embedding and generate a second sentence embedding by maximizing a first conditional probability of the second sentence embedding given the first sentence embedding and a second autoencoder receiving the second sentence embedding, the second autoencoder being trained to decode the second sentence embedding into a vector representation of a second textual sequence in the second language.Type: ApplicationFiled: June 12, 2020Publication date: December 16, 2021Inventors: Mehdi REZAGHOLIZADEH, Vahid PARTOVI NIA, Md Akmal HAIDAR, Pascal POUPART
-
Publication number: 20210383238Abstract: A method of and system for compressing a deep neural network model using knowledge distillation.Type: ApplicationFiled: June 25, 2021Publication date: December 9, 2021Inventors: Aref JAFARI, Mehdi REZAGHOLIZADEH, Ali GHODSI
-
Patent number: 11151334Abstract: In at least one broad aspect, described herein are systems and methods in which a latent representation shared between two languages is built and/or accessed, and then leveraged for the purpose of text generation in both languages. Neural text generation techniques are applied to facilitate text generation, and in particular the generation of sentences (i.e., sequences of words or subwords) in both languages, in at least some embodiments.Type: GrantFiled: September 26, 2018Date of Patent: October 19, 2021Assignee: HUAWEI TECHNOLOGIES CO., LTD.Inventors: Mehdi Rezagholizadeh, Md Akmal Haidar, Alan Do-Omri, Ahmad Rashid
-
Patent number: 11120337Abstract: A method and system for augmenting a training dataset for a generative adversarial network (GAN). The training dataset includes labelled data samples and unlabelled data samples. The method includes: receiving generated samples generated using a first neural network of the GAN and the unlabelled samples of training dataset; determining a decision value for a sample from a decision function, wherein the sample is a generated sample of the generated samples or an unlabelled sample of the unlabelled samples of the training dataset; comparing the decision value to a threshold; in response to determining that the decision value exceeds the threshold: predicting a label for a sample; assigning the label to the sample; and augmenting the training dataset to include the sample with the assigned label as a labelled sample.Type: GrantFiled: October 20, 2017Date of Patent: September 14, 2021Assignee: HUAWEI TECHNOLOGIES CO., LTD.Inventors: Dalei Wu, Md Akmal Haidar, Mehdi Rezagholizadeh, Alan Do-Omri
-
Patent number: 11003995Abstract: Method and system for performing semi-supervised regression with a generative adversarial network (GAN) that includes a generator comprising a first neural network and a discriminator comprising a second neural network, comprising: outputting, from the first neural network, generated samples derived from a random noise vector; inputting, to the second neural network, the generated samples, a plurality of labelled training samples, and a plurality of unlabelled training samples; and outputting, from the second neural network, a predicted continuous label for each of a plurality of the generated samples and unlabelled samples.Type: GrantFiled: October 20, 2017Date of Patent: May 11, 2021Assignee: HUAWEI TECHNOLOGIES CO., LTD.Inventors: Mehdi Rezagholizadeh, Md Akmal Haidar, Dalei Wu
-
Publication number: 20200134415Abstract: In accordance to embodiments, an encoder neural network is configured to receive a one-hot representation of a real text and output a latent representation of the real text generated from the one-hot representation of the real text. A decoder neural network is configured to receive the latent representation of the real text, and output a reconstructed softmax representation of the real text from the latent representation of the real text, the reconstructed softmax representation of the real text is a soft-text. A generator neural network is configured to generate artificial text based on random noise data. A discriminator neural network is configured to receive the soft-text and receive a softmax representation of the artificial text, and output a probability indicating whether the softmax representation of the artificial text received by the discriminator neural network is not from the generator neural network.Type: ApplicationFiled: October 30, 2018Publication date: April 30, 2020Inventors: Md Akmal Haidar, Mehdi Rezagholizadeh
-
Publication number: 20200134463Abstract: According to embodiments, an encoder neural network receives a one-hot representation of a real text. The encoder neural network outputs a latent representation of the real text. A decoder neural network receives random noise data or artificial code generated by a generator neural network from random noise data. The decoder neural network outputs softmax representation of artificial text. The decoder neural network receives the latent representation of the real text. The decoder neural network outputs a reconstructed softmax representation of the real text. A hybrid discriminator neural network receives a first combination of the soft-text and the latent representation of the real text and a second combination of the softmax representation of artificial text and the artificial code. The hybrid discriminator neural network outputs a probability indicating whether the second combination is similar to the first combination. Additional embodiments for utilizing latent representation are also disclosed.Type: ApplicationFiled: October 30, 2018Publication date: April 30, 2020Inventors: Md Akmal Haidar, Mehdi Rezagholizadeh
-
Patent number: 10607525Abstract: A system and method for color retargeting of an image includes applying a color appearance model to the image to be displayed based in part on a first luminance level. The color appearance model outputs a first set of color responses representing a simulated version of the image at the first luminance level. A color compensation model is further applied to the first set of color responses based in part on a second luminance level. The color compensation model outputs a second set of color responses representing a compensated version of the image. The compensated version of the image may be displayed on a display device set at the second luminance level. At least one of the color appearance model and the color compensation model applies rod-intrusion correction.Type: GrantFiled: May 19, 2016Date of Patent: March 31, 2020Assignee: Irystec Software Inc.Inventors: Tara Akhavan, Mehdi Rezagholizadeh, Afsoon Soudi