Patents by Inventor Nicholas Johnston
Nicholas Johnston has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20240104786Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for reliably performing data compression and data decompression across a wide variety of hardware and software platforms by using integer neural networks. In one aspect, there is provided a method for entropy encoding data which defines a sequence comprising a plurality of components, the method comprising: for each component of the plurality of components: processing an input comprising: (i) a respective integer representation of each of one or more components of the data which precede the component in the sequence, (ii) an integer representation of one or more respective latent variables characterizing the data, or (iii) both, using an integer neural network to generate data defining a probability distribution over the predetermined set of possible code symbols for the component of the data.Type: ApplicationFiled: November 28, 2023Publication date: March 28, 2024Inventors: Nicholas Johnston, Johannes Balle
-
Publication number: 20240078712Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for compressing and decompressing data. In one aspect, a method comprises: processing data using an encoder neural network to generate a latent representation of the data; processing the latent representation of the data using a hyper-encoder neural network to generate a latent representation of an entropy model; generating an entropy encoded representation of the latent representation of the entropy model; generating an entropy encoded representation of the latent representation of the data using the latent representation of the entropy model; and determining a compressed representation of the data from the entropy encoded representations of: (i) the latent representation of the data and (ii) the latent representation of the entropy model used to entropy encode the latent representation of the data.Type: ApplicationFiled: April 25, 2023Publication date: March 7, 2024Inventors: David Charles Minnen, Saurabh Singh, Johannes Balle, Troy Chinen, Sung Jin Hwang, Nicholas Johnston, George Dan Toderici
-
Patent number: 11869221Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for reliably performing data compression and data decompression across a wide variety of hardware and software platforms by using integer neural networks. In one aspect, there is provided a method for entropy encoding data which defines a sequence comprising a plurality of components, the method comprising: for each component of the plurality of components: processing an input comprising: (i) a respective integer representation of each of one or more components of the data which precede the component in the sequence, (ii) an integer representation of one or more respective latent variables characterizing the data, or (iii) both, using an integer neural network to generate data defining a probability distribution over the predetermined set of possible code symbols for the component of the data.Type: GrantFiled: September 18, 2019Date of Patent: January 9, 2024Assignee: Google LLCInventors: Nicholas Johnston, Johannes Balle
-
Publication number: 20230186082Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for processing a network input using a neural network to generate a network output for the network input. One of the methods includes maintaining, for each of the plurality of neural network layers, a respective look-up table that maps each possible combination of a quantized input index and a quantized weight index to a multiplication result; and generating a network output from a network input, comprising, for each of the neural network layers: receiving data specifying a quantized input to the neural network layer, the quantized input comprising a plurality of quantized input values; and generating a layer output for the neural network layer from the quantized input to the neural network layer using the respective look-up table for the neural network layer.Type: ApplicationFiled: October 31, 2022Publication date: June 15, 2023Inventors: Michele Covell, David Marwood, Shumeet Baluja, Nicholas Johnston
-
Patent number: 11670010Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for compressing and decompressing data. In one aspect, a method comprises: processing data using an encoder neural network to generate a latent representation of the data; processing the latent representation of the data using a hyper-encoder neural network to generate a latent representation of an entropy model; generating an entropy encoded representation of the latent representation of the entropy model; generating an entropy encoded representation of the latent representation of the data using the latent representation of the entropy model; and determining a compressed representation of the data from the entropy encoded representations of: (i) the latent representation of the data and (ii) the latent representation of the entropy model used to entropy encode the latent representation of the data.Type: GrantFiled: January 19, 2022Date of Patent: June 6, 2023Assignee: Google LLCInventors: David Charles Minnen, Saurabh Singh, Johannes Balle, Troy Chinen, Sung Jin Hwang, Nicholas Johnston, George Dan Toderici
-
Patent number: 11610124Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for receiving, by a neural network (NN), a dataset for generating features from the dataset. A first set of features is computed from the dataset using at least a feature layer of the NN. The first set of features i) is characterized by a measure of informativeness; and ii) is computed such that a size of the first set of features is compressible into a second set of features that is smaller in size than the first set of features and that has a same measure of informativeness as the measure of informativeness of the first set of features. The second set of features if generated from the first set of features using a compression method that compresses the first set of features to generate the second set of features.Type: GrantFiled: October 29, 2019Date of Patent: March 21, 2023Assignee: Google LLCInventors: Abhinav Shrivastava, Saurabh Singh, Johannes Balle, Sami Ahmad Abu-El-Haija, Nicholas Johnston, George Dan Toderici
-
Patent number: 11488016Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for processing a network input using a neural network to generate a network output for the network input. One of the methods includes maintaining, for each of the plurality of neural network layers, a respective look-up table that maps each possible combination of a quantized input index and a quantized weight index to a multiplication result; and generating a network output from a network input, comprising, for each of the neural network layers: receiving data specifying a quantized input to the neural network layer, the quantized input comprising a plurality of quantized input values; and generating a layer output for the neural network layer from the quantized input to the neural network layer using the respective look-up table for the neural network layer.Type: GrantFiled: January 23, 2020Date of Patent: November 1, 2022Assignee: Google LLCInventors: Michele Covell, David Marwood, Shumeet Baluja, Nicholas Johnston
-
Patent number: 11354822Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for image compression and reconstruction. A request to generate an encoded representation of an input image is received. The encoded representation of the input image is then generated. The encoded representation includes a respective set of binary codes at each iteration. Generating the set of binary codes for the iteration from an initial set of binary includes: for any tiles that have already been masked off during any previous iteration, masking off the tile. For any tiles that have not yet been masked off during any of the previous iterations, a determination is made as to whether a reconstruction error of the tile when reconstructed from binary codes at the previous iterations satisfies an error threshold. When the reconstruction quality satisfies the error threshold, the tile is masked off.Type: GrantFiled: May 16, 2018Date of Patent: June 7, 2022Assignee: Google LLCInventors: Michele Covell, Damien Vincent, David Charles Minnen, Saurabh Singh, Sung Jin Hwang, Nicholas Johnston, Joel Eric Shor, George Dan Toderici
-
Publication number: 20220138991Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for compressing and decompressing data. In one aspect, a method comprises: processing data using an encoder neural network to generate a latent representation of the data; processing the latent representation of the data using a hyper-encoder neural network to generate a latent representation of an entropy model; generating an entropy encoded representation of the latent representation of the entropy model; generating an entropy encoded representation of the latent representation of the data using the latent representation of the entropy model; and determining a compressed representation of the data from the entropy encoded representations of: (i) the latent representation of the data and (ii) the latent representation of the entropy model used to entropy encode the latent representation of the data.Type: ApplicationFiled: January 19, 2022Publication date: May 5, 2022Inventors: David Charles Minnen, Saurabh Singh, Johannes Balle, Troy Chinen, Sung Jin Hwang, Nicholas Johnston, George Dan Toderici
-
Patent number: 11306831Abstract: A tap assembly includes a flow control mechanism, a spindle which rotates about its principal axis to operate the flow control mechanism, and a handle connectable to the spindle. The location on the handle where the spindle connects coincides with the spindle's principal axis but does not with the centroid of the handle's planform shape. When the handle is turned by a user, the handle rotates about the spindle's principal axis. The handle also moves or translates relative the spindle's principal axis. When the handle is in an initial “fully off” position, there is an area that is obscured from a user's view by the handle, but when the handle is initially turned from the initial position towards a final “fully on” position, the area begins to be revealed, and with further rotation of the handle towards the final position, more of, or different parts of, the area become revealed.Type: GrantFiled: May 11, 2017Date of Patent: April 19, 2022Assignees: RAMTAPS PTY LTD, ROGERS SELLER & MYHILL PTY LTDInventor: Nicholas Johnston
-
Patent number: 11257254Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for compressing and decompressing data. In one aspect, a method comprises: processing data using an encoder neural network to generate a latent representation of the data; processing the latent representation of the data using a hyper-encoder neural network to generate a latent representation of an entropy model; generating an entropy encoded representation of the latent representation of the entropy model; generating an entropy encoded representation of the latent representation of the data using the latent representation of the entropy model; and determining a compressed representation of the data from the entropy encoded representations of: (i) the latent representation of the data and (ii) the latent representation of the entropy model used to entropy encode the latent representation of the data.Type: GrantFiled: July 18, 2019Date of Patent: February 22, 2022Assignee: Google LLCInventors: David Charles Minnen, Saurabh Singh, Johannes Balle, Troy Chinen, Sung Jin Hwang, Nicholas Johnston, George Dan Toderici
-
Patent number: 11250595Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for image compression and reconstruction. An image encoder system receives a request to generate an encoded representation of an input image that has been partitioned into a plurality of tiles and generates the encoded representation of the input image. To generate the encoded representation, the system processes a context for each tile using a spatial context prediction neural network that has been trained to process context for an input tile and generate an output tile that is a prediction of the input tile. The system determines a residual image between the particular tile and the output tile generated by the spatial context prediction neural network by process the context for the particular tile and generates a set of binary codes for the particular tile by encoding the residual image using an encoder neural network.Type: GrantFiled: May 29, 2018Date of Patent: February 15, 2022Assignee: Google LLCInventors: Michele Covell, Damien Vincent, David Charles Minnen, Saurabh Singh, Sung Jin Hwang, Nicholas Johnston, Joel Eric Shor, George Dan Toderici
-
Publication number: 20210358180Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for reliably performing data compression and data decompression across a wide variety of hardware and software platforms by using integer neural networks. In one aspect, there is provided a method for entropy encoding data which defines a sequence comprising a plurality of components, the method comprising: for each component of the plurality of components: processing an input comprising: (i) a respective integer representation of each of one or more components of the data which precede the component in the sequence, (ii) an integer representation of one or more respective latent variables characterizing the data, or (iii) both, using an integer neural network to generate data defining a probability distribution over the predetermined set of possible code symbols for the component of the data.Type: ApplicationFiled: September 18, 2019Publication date: November 18, 2021Inventors: Nicholas Johnston, Johannes Balle
-
Publication number: 20210335017Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for image compression and reconstruction. A request to generate an encoded representation of an input image is received. The encoded representation of the input image is then generated. The encoded representation includes a respective set of binary codes at each iteration. Generating the set of binary codes for the iteration from an initial set of binary includes: for any tiles that have already been masked off during any previous iteration, masking off the tile. For any tiles that have not yet been masked off during any of the previous iterations, a determination is made as to whether a reconstruction error of the tile when reconstructed from binary codes at the previous iterations satisfies an error threshold. When the reconstruction quality satisfies the error threshold, the tile is masked off.Type: ApplicationFiled: May 16, 2018Publication date: October 28, 2021Inventors: Michele Covell, Damien Vincent, David Charles Minnen, Saurabh Singh, Sung Jin Hwang, Nicholas Johnston, Joel Eric Shor, George Dan Toderici
-
Patent number: 11080175Abstract: A scalable enterprise platform for automated functional and integration regression testing is provided. Embodiments of the disclosed system facilitate the testing of any number of different software systems in development, even where the systems have unique dataset formats. Embodiments of the present invention provide a common method to generate logging and results reports across the platform, thereby providing simpler results analysis. Embodiments may also standardize the query set and facilitate the capability to analyze large results sets. Furthermore, embodiments of the disclosed system may combine the original data to the validated data to allow testers to analyze the testing results. In addition, embodiments of the present invention supports secured separation of testing domains. In at least one embodiment, the system includes a centralized user interface system that provides users with different domains to securely access one or more testing domains.Type: GrantFiled: May 13, 2019Date of Patent: August 3, 2021Assignee: JPMORGAN CHASE BANK, N.A.Inventors: Mark R Wilson, Nicholas Johnston, Pollawat Poonjiradejma, Hani El Sayyed, Thomas Williams
-
Publication number: 20200311548Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for receiving, by a neural network (NN), a dataset for generating features from the dataset. A first set of features is computed from the dataset using at least a feature layer of the NN. The first set of features i) is characterized by a measure of informativeness; and ii) is computed such that a size of the first set of features is compressible into a second set of features that is smaller in size than the first set of features and that has a same measure of informativeness as the measure of informativeness of the first set of features. The second set of features if generated from the first set of features using a compression method that compresses the first set of features to generate the second set of features.Type: ApplicationFiled: October 29, 2019Publication date: October 1, 2020Inventors: Abhinav Shrivastava, Saurabh Singh, Johannes Balle, Sami Ahmad Abu-El-Haija, Nicholas Johnston, George Dan Toderici
-
Publication number: 20200234126Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for processing a network input using a neural network to generate a network output for the network input. One of the methods includes maintaining, for each of the plurality of neural network layers, a respective look-up table that maps each possible combination of a quantized input index and a quantized weight index to a multiplication result; and generating a network output from a network input, comprising, for each of the neural network layers: receiving data specifying a quantized input to the neural network layer, the quantized input comprising a plurality of quantized input values; and generating a layer output for the neural network layer from the quantized input to the neural network layer using the respective look-up table for the neural network layer.Type: ApplicationFiled: January 23, 2020Publication date: July 23, 2020
-
Publication number: 20200233792Abstract: A scalable enterprise platform for automated functional and integration regression testing is provided. Embodiments of the disclosed system facilitate the testing of any number of different software systems in development, even where the systems have unique dataset formats. Embodiments of the present invention provide a common method to generate logging and results reports across the platform, thereby providing simpler results analysis. Embodiments may also standardize the query set and facilitate the capability to analyze large results sets. Furthermore, embodiments of the disclosed system may combine the original data to the validated data to allow testers to analyze the testing results. In addition, embodiments of the present invention supports secured separation of testing domains. In at least one embodiment, the system includes a centralized user interface system that provides users with different domains to securely access one or more testing domains.Type: ApplicationFiled: May 13, 2019Publication date: July 23, 2020Inventors: Mark R. Wilson, Nicholas Johnston, Pollawat Poonjiradejma, Hani El Sayyed, Thomas Williams
-
Patent number: 10713818Abstract: Methods, and systems, including computer programs encoded on computer storage media for compressing data items with variable compression rate. A system includes an encoder sub-network configured to receive a system input image and to generate an encoded representation of the system input image, the encoder sub-network including a first stack of neural network layers including one or more LSTM neural network layers and one or more non-LSTM neural network layers, the first stack configured to, at each of a plurality of time steps, receive an input image for the time step that is derived from the system input image and generate a corresponding first stack output, and a binarizing neural network layer configured to receive a first stack output as input and generate a corresponding binarized output.Type: GrantFiled: January 28, 2019Date of Patent: July 14, 2020Assignee: Google LLCInventors: George Dan Toderici, Sean O'Malley, Rahul Sukthankar, Sung Jin Hwang, Damien Vincent, Nicholas Johnston, David Charles Minnen, Joel Shor, Michele Covell
-
Publication number: 20200111238Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for image compression and reconstruction. An image encoder system receives a request to generate an encoded representation of an input image that has been partitioned into a plurality of tiles and generates the encoded representation of the input image. To generate the encoded representation, the system processes a context for each tile using a spatial context prediction neural network that has been trained to process context for an input tile and generate an output tile that is a prediction of the input tile. The system determines a residual image between the particular tile and the output tile generated by the spatial context prediction neural network by process the context for the particular tile and generates a set of binary codes for the particular tile by encoding the residual image using an encoder neural network.Type: ApplicationFiled: May 29, 2018Publication date: April 9, 2020Inventors: Michele Covell, Damien Vincent, David Charles Minnen, Saurabh Singh, Sung Jin Hwang, Nicholas Johnston, Joel Eric Shor, George Dan Toderici