Patents by Inventor Randy Huang
Randy Huang has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 12314837Abstract: Provided are systems, methods, and integrated circuits for neural network processing. In various implementations, an integrated circuit for neural network processing can include a plurality of memory banks storing weight values for a neural network. The memory banks can be on the same chip as an array of processing engines. Upon receiving input data, the circuit can be configured to use the set of weight values to perform a task defined for the neural network. Performing the task can include reading weight values from the memory banks, inputting the weight values into the array of processing engines, and computing a result using the array of processing engines, where the result corresponds to an outcome of performing the task.Type: GrantFiled: June 22, 2023Date of Patent: May 27, 2025Assignee: Amazon Technologies, Inc.Inventors: Randy Huang, Ron Diamant
-
Patent number: 12067492Abstract: Disclosed herein are techniques for performing multi-layer neural network processing for multiple contexts. In one embodiment, a computing engine is set in a first configuration to implement a second layer of a neural network and to process first data related to a first context to generate first context second layer output. The computing engine can be switched from the first configuration to a second configuration to implement a first layer of the neural network. The computing engine can be used to process second data related to a second context to generate second context first layer output. The computing engine can be set to a third configuration to implement a third layer of the neural network to process the first context second layer output and the second context first layer output to generate a first processing result of the first context and a second processing result of the second context.Type: GrantFiled: May 5, 2023Date of Patent: August 20, 2024Assignee: Amazon Technologies, Inc.Inventors: Dana Michelle Vantrease, Ron Diamant, Thomas A. Volpe, Randy Huang
-
Patent number: 12008466Abstract: In various implementations, provided are systems and methods for operating a neural network that includes conditional structures. In some implementations, an integrated circuit can compute a result using a set of intermediate results, where the intermediate results are computed from the outputs of a hidden layer of the neural network. The integrated circuit can further test the result against a condition. The outcome of the test can determine a next layer that the integrated circuit is to execute, or can be used to determine that further execution of the neural network can be terminated.Type: GrantFiled: March 23, 2018Date of Patent: June 11, 2024Assignee: Amazon Technologies, Inc.Inventors: Randy Huang, Ron Diamant, Thomas A. Volpe
-
Patent number: 11960997Abstract: Disclosed herein are techniques for classifying data with a data processing circuit. In one embodiment, the data processing circuit includes a probabilistic circuit configurable to generate a decision at a pre-determined probability, and an output generation circuit including an output node and configured to receive input data and a weight, and generate output data at the output node for approximating a product of the input data and the weight. The generation of the output data includes propagating the weight to the output node according a first decision of the probabilistic circuit. The probabilistic circuit is configured to generate the first decision at a probability determined based on the input data.Type: GrantFiled: January 7, 2022Date of Patent: April 16, 2024Assignee: Amazon Technologies, Inc.Inventors: Randy Huang, Ron Diamant
-
Publication number: 20240072472Abstract: A connector comprises an insulation housing, a shielding shell, an inner sheath, and an outer sheath. The shielding shell is sheathed on the insulation housing. The inner sheath is formed on the shielding shell by injection molding. The inner sheath includes a front end wrapped on the shielding shell and a retaining portion formed on the front end of the inner sheath and spaced from the shielding shell to form a gap therebetween. The outer sheath is formed on the shielding shell and the inner sheath by injection molding. The outer sheath includes an engaging portion embedded in the gap between the retaining portion and the shielding shell. The retaining portion holds the engaging portion on the shielding shell to prevent the formation of a gap between the front end of the outer sheath and the shielding shell.Type: ApplicationFiled: August 29, 2023Publication date: February 29, 2024Applicants: SIBAS Electronics (Xiamen) Co. Ltd., Tyco Electronics (Shanghai) Co., Ltd.Inventors: Daokuan (Jeremy) Zhang, Weidong (Randy) Huang, Yong (Chris) Wang, Hua He, Xinzheng (Tony) Fan, Jianfei (Fencer) Yu
-
Patent number: 11868878Abstract: Disclosed herein are techniques for implementing a large fully-connected layer in an artificial neural network. The large fully-connected layer is grouped into multiple fully-connected subnetworks. Each fully-connected subnetwork is configured to classify an object into an unknown class or a class in a subset of target classes. If the object is classified as the unknown class by a fully-connected subnetwork, a next fully-connected subnetwork may be used to further classify the object. In some embodiments, the fully-connected layer is grouped based on a ranking of target classes.Type: GrantFiled: March 23, 2018Date of Patent: January 9, 2024Assignee: Amazon Technologies, Inc.Inventors: Randy Huang, Ron Diamant
-
Publication number: 20230351186Abstract: Disclosed herein are techniques for performing multi-layer neural network processing for multiple contexts. In one embodiment, a computing engine is set in a first configuration to implement a second layer of a neural network and to process first data related to a first context to generate first context second layer output. The computing engine can be switched from the first configuration to a second configuration to implement a first layer of the neural network. The computing engine can be used to process second data related to a second context to generate second context first layer output. The computing engine can be set to a third configuration to implement a third layer of the neural network to process the first context second layer output and the second context first layer output to generate a first processing result of the first context and a second processing result of the second context.Type: ApplicationFiled: May 5, 2023Publication date: November 2, 2023Inventors: Dana Michelle Vantrease, Ron Diamant, Thomas A. Volpe, Randy Huang
-
Patent number: 11797853Abstract: Disclosed herein are techniques for performing multi-layer neural network processing for multiple contexts. In one embodiment, a computing engine is set in a first configuration to implement a second layer of a neural network and to process first data related to a first context to generate first context second layer output. The computing engine can be switched from the first configuration to a second configuration to implement a first layer of the neural network. The computing engine can be used to process second data related to a second context to generate second context first layer output. The computing engine can be set to a third configuration to implement a third layer of the neural network to process the first context second layer output and the second context first layer output to generate a first processing result of the first context and a second processing result of the second context.Type: GrantFiled: September 22, 2022Date of Patent: October 24, 2023Assignee: Amazon Technologies, Inc.Inventors: Dana Michelle Vantrease, Ron Diamant, Thomas A. Volpe, Randy Huang
-
Publication number: 20230334294Abstract: Provided are systems, methods, and integrated circuits for neural network processing. In various implementations, an integrated circuit for neural network processing can include a plurality of memory banks storing weight values for a neural network. The memory banks can be on the same chip as an array of processing engines. Upon receiving input data, the circuit can be configured to use the set of weight values to perform a task defined for the neural network. Performing the task can include reading weight values from the memory banks, inputting the weight values into the array of processing engines, and computing a result using the array of processing engines, where the result corresponds to an outcome of performing the task.Type: ApplicationFiled: June 22, 2023Publication date: October 19, 2023Inventors: Randy Huang, Ron Diamant
-
Patent number: 11741345Abstract: Provided are systems, methods, and integrated circuits for a neural network processing system. In various implementations, the system can include a first array of processing engines coupled to a first set of memory banks and a second array of processing engines coupled to a second set of memory banks. The first and second set of memory banks be storing all the weight values for a neural network, where the weight values are stored before any input data is received. Upon receiving input data, the system performs a task defined for the neural network. Performing the task can include computing an intermediate result using the first array of processing engines, copying the intermediate result to the second set of memory banks, and computing a final result using the second array of processing engines, where the final result corresponds to an outcome of performing the task.Type: GrantFiled: September 25, 2020Date of Patent: August 29, 2023Assignee: Amazon Technologies, Inc.Inventors: Randy Huang, Ron Diamant
-
Publication number: 20230048221Abstract: A connector includes an insulator, a plurality of pairs of terminals provided on the insulator, and a resistance element. The plurality of pairs of terminals include at least one pair of cable terminals electrically connected with a cable and at least one pair of vacant terminals that are unused. The resistance element is electrically connected between each pair of vacant terminals.Type: ApplicationFiled: August 10, 2022Publication date: February 16, 2023Applicants: Tyco Electronics (Shanghai) Co. Ltd., SIBAS Electronics (Xiamen) Co. Ltd.Inventors: Daokuan (Jeremy) Zhang, Weidong (Randy) Huang, Yong (Chris) Wang, Yu (Rain) Wang, Hua He, Yingchun (David) Wang, Liqiang (Gino) Yao
-
Publication number: 20230014783Abstract: Disclosed herein are techniques for performing multi-layer neural network processing for multiple contexts. In one embodiment, a computing engine is set in a first configuration to implement a second layer of a neural network and to process first data related to a first context to generate first context second layer output. The computing engine can be switched from the first configuration to a second configuration to implement a first layer of the neural network. The computing engine can be used to process second data related to a second context to generate second context first layer output. The computing engine can be set to a third configuration to implement a third layer of the neural network to process the first context second layer output and the second context first layer output to generate a first processing result of the first context and a second processing result of the second context.Type: ApplicationFiled: September 22, 2022Publication date: January 19, 2023Inventors: Dana Michelle Vantrease, Ron Diamant, Thomas A. Volpe, Randy Huang
-
Patent number: 11475306Abstract: Disclosed herein are techniques for performing multi-layer neural network processing for multiple contexts. In one embodiment, a computing engine is set in a first configuration to implement a second layer of a neural network and to process first data related to a first context to generate first context second layer output. The computing engine can be switched from the first configuration to a second configuration to implement a first layer of the neural network. The computing engine can be used to process second data related to a second context to generate second context first layer output. The computing engine can be set to a third configuration to implement a third layer of the neural network to process the first context second layer output and the second context first layer output to generate a first processing result of the first context and a second processing result of the second context.Type: GrantFiled: March 22, 2018Date of Patent: October 18, 2022Assignee: Amazon Technologies, Inc.Inventors: Dana Michelle Vantrease, Ron Diamant, Thomas A. Volpe, Randy Huang
-
Patent number: 11461631Abstract: Disclosed herein are techniques for scheduling and executing multi-layer neural network computations for multiple contexts. In one embodiment, a method comprises determining a set of computation tasks to be executed, the set of computation tasks including a first computation task and a second computation task, as well as a third computation task and a fourth computation task to provide input data for the first and second computation tasks; determining a first execution batch comprising the first and second computation tasks; determining a second execution batch comprising at least the third computation task to be executed before the first execution batch; determining whether to include the fourth computation task in the second execution batch based on whether the memory device has sufficient capacity to hold input data and output data of both of the third and fourth computation; executing the second execution batch followed by the first execution batch.Type: GrantFiled: March 22, 2018Date of Patent: October 4, 2022Assignee: Amazon Technologies, Inc.Inventors: Dana Michelle Vantrease, Ron Diamant, Thomas A. Volpe, Randy Huang
-
Patent number: 11416736Abstract: Systems and methods are related to improving throughput of neural networks in integrated circuits by combining values in operands to increase compute density. A system includes an integrated circuit (IC) having multiplier circuitry. The IC receives a first value and a second value in a first operand. The IC performs a multiplication operation, via the multiplier circuitry, on the first operand and a second operand to produce a first multiplied product based at least in part on the first value and a second multiplied product based at least in part on the second value.Type: GrantFiled: December 27, 2017Date of Patent: August 16, 2022Assignee: Intel CorporationInventors: Kevin Nealis, Randy Huang
-
Patent number: 11263517Abstract: Disclosed herein are techniques for obtain weights for neural network computations. In one embodiment, an integrated circuit may include an arithmetic circuit configured to perform arithmetic operations for a neural network. The integrated circuit may also include a weight processing circuit configured to: acquire data from a memory device; receive configuration information indicating a size of each quantized weight of a set of quantized weights; extract the set of quantized weights from the data based on the size of the each weight indicated by the configuration information; perform de-quantization processing on the set of quantized weights to generate a set of de-quantized weights; and provide the set of de-quantized weights to the arithmetic circuit to enable the arithmetic circuit to perform the arithmetic operations. The memory device may be part of or external to the integrated circuit.Type: GrantFiled: February 28, 2018Date of Patent: March 1, 2022Assignee: Amazon Technologies, Inc.Inventors: Ron Diamant, Randy Huang
-
Patent number: 11250319Abstract: Disclosed herein are techniques for classifying data with a data processing circuit. In one embodiment, the data processing circuit includes a probabilistic circuit configurable to generate a decision at a pre-determined probability, and an output generation circuit including an output node and configured to receive input data and a weight, and generate output data at the output node for approximating a product of the input data and the weight. The generation of the output data includes propagating the weight to the output node according a first decision of the probabilistic circuit. The probabilistic circuit is configured to generate the first decision at a probability determined based on the input data.Type: GrantFiled: September 25, 2017Date of Patent: February 15, 2022Assignee: Amazon Technologies, Inc.Inventors: Randy Huang, Ron Diamant
-
Patent number: 10983754Abstract: Disclosed herein are techniques for accelerating convolution operations or other matrix multiplications in applications such as neural network. In one example, an apparatus comprises a first circuit, a second circuit, and a third circuit. The first circuit is configured to: receive first values in a first format, the first values being generated from one or more asymmetric quantization operations of second values in a second format, and generate difference values based on subtracting a third value from each of the first values, the third value representing a zero value in the first format. The second circuit is configured to generate a sum of products in the first format using the difference values. The third circuit is configured to convert the sum of products from the first format to the second format based on scaling the sum of products with a scaling factor.Type: GrantFiled: June 2, 2020Date of Patent: April 20, 2021Assignee: Amazon Technologies, Inc.Inventors: Dana Michelle Vantrease, Randy Huang, Ron Diamant, Thomas Elmer, Sundeep Amirineni
-
Patent number: 10943167Abstract: Disclosed herein are techniques for performing neural network computations. In one embodiment, an apparatus includes an array of processing elements, the array having configurable dimensions. The apparatus further includes a controller configured to set the dimensions of the array of processing elements based on at least one of: a first number of input data sets to be received by the array, or a second number of output data sets to be output by the array.Type: GrantFiled: August 12, 2019Date of Patent: March 9, 2021Assignee: Amazon Technologies, Inc.Inventors: Sundeep Amirineni, Ron Diamant, Randy Huang, Thomas A. Volpe
-
Publication number: 20210019600Abstract: Provided are systems, methods, and integrated circuits for a neural network processing system. In various implementations, the system can include a first array of processing engines coupled to a first set of memory banks and a second array of processing engines coupled to a second set of memory banks. The first and second set of memory banks be storing all the weight values for a neural network, where the weight values are stored before any input data is received. Upon receiving input data, the system performs a task defined for the neural network. Performing the task can include computing an intermediate result using the first array of processing engines, copying the intermediate result to the second set of memory banks, and computing a final result using the second array of processing engines, where the final result corresponds to an outcome of performing the task.Type: ApplicationFiled: September 25, 2020Publication date: January 21, 2021Inventors: Randy Huang, Ron Diamant