Patents by Inventor Guillaume Konrad SAUTIÈRE
Guillaume Konrad SAUTIÈRE has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20250090963Abstract: A device includes a memory and one or more processors coupled to the memory and configured to execute instructions from the memory. Execution of the instructions causes the one or more processors to combine two or more data portions to generate input data for a decoder network. A first data portion of the two or more data portions is based on a first encoding of a data sample by a multiple description coding network and content of a second data portion of the two or more data portions depends on whether data based on a second encoding of the data sample by the multiple description coding network is available. Execution of the instructions also causes the one or more processors to obtain, from the decoder network, output data based on the input data and to generate a representation of the data sample based on the output data.Type: ApplicationFiled: September 8, 2022Publication date: March 20, 2025Inventors: Zisis Iason SKORDILIS, Vivek RAJENDRAN, Guillaume Konrad SAUTIERE, Duminda DEWASURENDRA, Daniel Jared SINDER
-
Publication number: 20240428813Abstract: Systems and techniques are described for coding audio signals. For example, a voice decoder can generate, using a first neural network, an excitation signal for at least one sample of an audio signal at least in part by performing a non-linear operation based on one or more inputs to the first neural network, the excitation signal being configured to excite a learned linear filter. The voice decoder can further generate, using the learned linear filter and the excitation signal, at least one sample of a reconstructed audio signal. For example, a second neural network can be used to generate coefficients for one or more learned linear filters, which receive as input the excitation signal generated by the first neural network trained to perform the non-linear operation.Type: ApplicationFiled: October 10, 2022Publication date: December 26, 2024Inventors: Guillaume Konrad SAUTIERE, Duminda DEWASURENDRA, Zisis Iason SKORDILIS, Vivek RAJENDRAN
-
Publication number: 20240428814Abstract: Systems and techniques are described for coding audio signals. For example, a voice decoder can generate, using a neural network, an excitation signal for at least one sample of an audio signal based on one or more inputs to the neural network, the excitation signal being configured to excite a linear predictive coding (LPC) filter. The voice decoder can further generate, using the LPC filter based on the excitation signal, at least one sample of a reconstructed audio signal. For example, the neural network can generate coefficients for one or more linear time-varying filters (e.g., a linear time-varying harmonic filter and a linear time-varying noise filter). The voice decoder can use the one or more linear time-varying filters including the generated coefficients to generate the excitation signal.Type: ApplicationFiled: October 10, 2022Publication date: December 26, 2024Inventors: Duminda DEWASURENDRA, Guillaume Konrad SAUTIERE, Zisis Iason SKORDILIS, Vivek RAJENDRAN
-
Publication number: 20240378698Abstract: Systems and techniques are provided for processing image data. According to some aspects, a computing device can determine an optical flow between a current frame having a first resolution and a first previous frame having the first resolution. The computing device can warp a second previous frame having a second resolution based on the determined optical flow to generate a warped previous frame having the second resolution, the second resolution being higher than the first resolution. The computing device can process, using a diffusion machine learning model, a noise frame, the current frame, and the warped previous frame to generate an output frame having the second resolution.Type: ApplicationFiled: July 31, 2023Publication date: November 14, 2024Inventors: Jens PETERSEN, Michal Jakub STYPULKOWSKI, Noor Fathima Khanum MOHAMED GHOUSE, Auke Joris WIGGERS, Guillaume Konrad SAUTIERE
-
Publication number: 20240371384Abstract: Systems and techniques are described for audio coding. An audio system receives feature(s) corresponding an audio signal, for example from an encoder and/or a speech synthesis engine. The audio system generates an excitation signal, such as a harmonic signal and/or a noise signal, based on the feature(s). The audio system uses a filterbank to generate band-specific signals from the excitation signal. The band-specific signals correspond to frequency bands. The audio system inputs the feature(s) into a machine learning (ML) filter estimator to generate parameter(s) associated with linear filter(s). The audio system inputs the feature(s) into a voicing estimator to generate gain value(s). The audio system generates an output audio signal based on modification of the band-specific signals, application of the linear filter(s) according to the parameter(s), and amplification using the gain amplifier(s) according to the gain value(s).Type: ApplicationFiled: October 10, 2022Publication date: November 7, 2024Inventors: Zisis Iason SKORDILIS, Vivek RAJENDRAN, Duminda DEWASURENDRA, Guillaume Konrad SAUTIERE
-
Publication number: 20240364925Abstract: Systems and techniques are described herein for processing video data. For example, a machine-learning based stereo video coding system can obtain video data including at least a right-view image of a right view of a scene and a left-view image of a left view of the scene. The machine-learning based stereo video coding system can compress the right-view image and the left-view image in parallel to generate a latent representation of the right-view image and the left-view image. The right-view image and the left-view image can be compressed in parallel based on inter-view information between the right-view image and the left-view image, determined using one or more parallel autoencoders.Type: ApplicationFiled: April 15, 2024Publication date: October 31, 2024Inventors: Hoang Cong Minh LE, Qiqi HOU, Farzad FARHADZADEH, Amir SAID, Auke Joris WIGGERS, Guillaume Konrad SAUTIERE, Reza POURREZA
-
Publication number: 20240323415Abstract: Certain aspects of the present disclosure provide techniques and apparatus for encoding content using a neural network. An example method generally includes encoding video content into a latent space representation through an encoder implemented by a first machine learning model. A code is generated by upsampling the latent space representation of the video content. A prior is calculated based on a conditional probability of obtaining the upsampled latent space representation conditioned by the latent space representation of the video content. A compressed version of the video content is generated based on a probabilistic model implemented by a second machine learning model, the generated code, and the calculated prior, and the compressed version of the video content is output for transmission.Type: ApplicationFiled: March 22, 2023Publication date: September 26, 2024Inventors: David Wilson ROMERO GUZMAN, Gabriele CESA, Guillaume Konrad SAUTIERE, Yunfan ZHANG, Taco Sebastiaan COHEN, Auke Joris WIGGERS
-
Publication number: 20240305785Abstract: An example computing device may include memory and one or more processors. The one or more processors may be configured to parallel entropy decode encoded video data from a received bitstream to generate entropy decoded data. The one or more processors may be configured to predict a motion vector based on the entropy decoded data. The one or more processors may be configured to decode a motion vector residual from the entropy decoded data. The one or more processors may be configured to add the motion vector residual and motion vector. The one or more processors may be configured to warp previous reconstructed video data with an overlapped block-based warp function using the motion vector to generate predicted current video data. The one or more processors may be configured to sum the predicted current video data with a residual block to generate current reconstructed video data.Type: ApplicationFiled: August 28, 2023Publication date: September 12, 2024Inventors: Ties Jehan Van Rozendaal, Hoang Cong Minh Le, Tushar Singhal, Amir Said, Krishna Buska, Guillaume Konrad Sautiere, Anjuman Raha, Auke Joris Wiggers, Frank Steven Mayer, Liang Zhang, Abhijit Khobare, Muralidhar Reddy Akula
-
Publication number: 20240121398Abstract: Systems and techniques are described for processing image data using a residual model that can be configured with an adjustable number of sampling steps. For example, a process can include obtaining a latent representation of an image and processing, using a decoder of a machine learning model, the latent representation of the image to generate an initial reconstructed image. The process can further include processing, using the residual model, the initial reconstructed image and noise data to predict a plurality of predictions of a residual over a number of sampling steps. The residual represents a difference between the image and the initial reconstructed image. The process can include obtaining, from the plurality of predictions of the residual, a final residual representing the difference between the image and the initial reconstructed image. The process can further include combining the initial reconstructed image and the residual to generate a final reconstructed image.Type: ApplicationFiled: August 29, 2023Publication date: April 11, 2024Inventors: Noor Fathima Khanum MOHAMED GHOUSE, Jens PETERSEN, Tianlin XU, Guillaume Konrad SAUTIERE, Auke Joris WIGGERS
-
Patent number: 11526734Abstract: A device includes one or more processors configured to generate, at an encoder portion of an autoencoder, first output data at least partially based on first input data and to generate, at a decoder portion or the autoencoder, a representation of the first input data at least partially based on the first output data. The one or more processors are configured to generate, at the encoder portion, second output data based on second input data and first state data and to generate, at the decoder portion, a representation of the second input data based on the second output data and second state data. Each of the first state data and the second state data correspond to the state of the decoder portion resulting from generation of the representation of the first input data. The first and second input data correspond to sequential values of a signal to be encoded.Type: GrantFiled: April 6, 2020Date of Patent: December 13, 2022Assignee: QUALCOMM IncorporatedInventors: Yang Yang, Guillaume Konrad Sautière, Jongha Ryu, Taco Sebastiaan Cohen
-
Patent number: 11437050Abstract: Techniques are described for coding audio signals. For example, using a neural network, a residual signal is generated for a sample of an audio signal based on inputs to the neural network. The residual signal is configured to excite a long-term prediction filter and/or a short-term prediction filter. Using the long-term prediction filter and/or the short-term prediction filter, a sample of a reconstructed audio signal is determined. The sample of the reconstructed audio signal is determined based on the residual signal generated using the neural network for the sample of the audio signal.Type: GrantFiled: December 10, 2019Date of Patent: September 6, 2022Assignee: QUALCOMM IncorporatedInventors: Zisis Iason Skordilis, Vivek Rajendran, Guillaume Konrad Sautière, Daniel Jared Sinder
-
Patent number: 11405626Abstract: Techniques are described herein for coding video content using recurrent-based machine learning tools. A device can include a neural network system including encoder and decoder portions. The encoder portion can generate output data for the current time step of operation of the neural network system based on an input video frame for a current time step of operation of the neural network system, reconstructed motion estimation data from a previous time step of operation, reconstructed residual data from the previous time step of operation, and recurrent state data from at least one recurrent layer of a decoder portion of the neural network system from the previous time step of operation. A decoder portion of the neural network system can generate, based on the output data and recurrent state data from the previous time step of operation, a reconstructed video frame for the current time step of operation.Type: GrantFiled: November 6, 2020Date of Patent: August 2, 2022Assignee: QUALCOMM IncorporatedInventors: Adam Waldemar Golinski, Yang Yang, Reza Pourreza, Guillaume Konrad Sautiere, Ties Jehan Van Rozendaal, Taco Sebastiaan Cohen
-
Publication number: 20210281867Abstract: Techniques are described herein for coding video content using recurrent-based machine learning tools. A device can include a neural network system including encoder and decoder portions. The encoder portion can generate output data for the current time step of operation of the neural network system based on an input video frame for a current time step of operation of the neural network system, reconstructed motion estimation data from a previous time step of operation, reconstructed residual data from the previous time step of operation, and recurrent state data from at least one recurrent layer of a decoder portion of the neural network system from the previous time step of operation. A decoder portion of the neural network system can generate, based on the output data and recurrent state data from the previous time step of operation, a reconstructed video frame for the current time step of operation.Type: ApplicationFiled: November 6, 2020Publication date: September 9, 2021Inventors: Adam Waldemar GOLINSKI, Yang YANG, Reza POURREZA, Guillaume Konrad SAUTIERE, Ties Jehan VAN ROZENDAAL, Taco Sebastiaan COHEN
-
Publication number: 20210089863Abstract: A device includes one or more processors configured to generate, at an encoder portion of an autoencoder, first output data at least partially based on first input data and to generate, at a decoder portion or the autoencoder, a representation of the first input data at least partially based on the first output data. The one or more processors are configured to generate, at the encoder portion, second output data based on second input data and first state data and to generate, at the decoder portion, a representation of the second input data based on the second output data and second state data. Each of the first state data and the second state data correspond to the state of the decoder portion resulting from generation of the representation of the first input data. The first and second input data correspond to sequential values of a signal to be encoded.Type: ApplicationFiled: April 6, 2020Publication date: March 25, 2021Inventors: Yang YANG, Guillaume Konrad SAUTIÈRE, Jongha RYU, Taco Sebastiaan COHEN
-
Publication number: 20210074308Abstract: Techniques are described for coding audio signals. For example, using a neural network, a residual signal is generated for a sample of an audio signal based on inputs to the neural network. The residual signal is configured to excite a long-term prediction filter and/or a short-term prediction filter. Using the long-term prediction filter and/or the short-term prediction filter, a sample of a reconstructed audio signal is determined. The sample of the reconstructed audio signal is determined based on the residual signal generated using the neural network for the sample of the audio signal.Type: ApplicationFiled: December 10, 2019Publication date: March 11, 2021Inventors: Zisis Iason SKORDILIS, Vivek RAJENDRAN, Guillaume Konrad SAUTIÈRE, Daniel Jared SINDER