Patents by Inventor Larry Roberts
Larry Roberts has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 11951534Abstract: Systems and methods are disclosed for an event detection system that captures data associated with events while a DC casting system forms an ingot, determines characteristics of the events, and improves the casting system based on the events. Example systems and methods may include initiating a casting operation using one or more pieces of equipment of a casting system including a casting apparatus; capturing sensor data associated with one or more acoustic signals captured relative to the one or more pieces of equipment performing the casting operation; comparing the sensor data with a set of acoustic profiles; determining whether a particular type of event has occurred; causing an adjustment to the casting system or to the casting operation based on whether the particular type of event has occurred; and initiating a second casting operation using the adjusted casting system or casting operation.Type: GrantFiled: July 23, 2021Date of Patent: April 9, 2024Assignee: Novelis Inc.Inventors: John Robert Buster McCallum, Phillip Wilson, Larry Hudson, Richard Shockley, Todd F. Bischoff, Adwait Thakur
-
Patent number: 11938372Abstract: An improved exercise assembly structured to be struck by a user includes a base supporting a shaft on a supporting surface. A support shaft extends outwardly from the supporting surface. At least one target is connected to the support shaft and extends outwardly. The at least one target is connected to a breakaway mechanism, which is connected to a rotational fitting, which connects to the support shaft. The target is able to be struck and rotate around the shaft via the rotational fitting. Upon a misdirected striking (force) inflicting on the target, the breakaway mechanism mitigates forces away from the shaft and rotational fitting in order to reduce total stress on the assembly.Type: GrantFiled: September 21, 2021Date of Patent: March 26, 2024Inventors: Larry Roberts, Bertho Boman
-
Patent number: 11847550Abstract: A method, computer program product, and system perform computations using a processor. A first instruction including a first index vector operand and a second index vector operand is received and the first index vector operand is decoded to produce first coordinate sets for a first array, each first coordinate set including at least a first coordinate and a second coordinate of a position of a non-zero element in the first array. The second index vector operand is decoded to produce second coordinate sets for a second array, each second coordinate set including at least a third coordinate and a fourth coordinate of a position of a non-zero element in the second array. The first coordinate sets are summed with the second coordinate sets to produce output coordinate sets and the output coordinate sets are converted into a set of linear indices.Type: GrantFiled: December 4, 2020Date of Patent: December 19, 2023Assignee: NVIDIA CorporationInventors: William J. Dally, Angshuman Parashar, Joel Springer Emer, Stephen William Keckler, Larry Robert Dennison
-
Patent number: 11799799Abstract: A switch architecture enables ports to stash packets in unused buffers on other ports, exploiting excess internal bandwidth that may exist, for example, in a tiled switch. This architecture leverages unused port buffer memory to improve features such as congestion handling and error recovery.Type: GrantFiled: July 16, 2021Date of Patent: October 24, 2023Assignee: NVIDIA Corp.Inventors: Matthias Augustin Blumrich, Nan Jiang, Larry Robert Dennison
-
Publication number: 20230327996Abstract: Aggregation of small payloads from multiple packets may improve bandwidth efficiency of a network, particularly a high-performance compute cluster with thousands of network endpoints and distributed data. Aggregation is context-based and a packet header is reduced because the common components that are shared by the aggregated messages are included once within the header. Execution contexts are explicitly created and destroyed by application programs. Each participating endpoint stores context-specific properties until the context is destroyed, so that the properties are not included in the header. Aggregation may be performed at different hierarchical levels by switches and/or endpoints.Type: ApplicationFiled: January 4, 2023Publication date: October 12, 2023Inventors: Benjamin Klenk, Alan Lynn Davis, Larry Robert Dennison
-
Patent number: 11770215Abstract: Packet flows between a transmitter and a receiver in an unreliable and unordered switched packet network may be established as a result of receiving a second packet comprising a second memory operation on a memory address. The transmission of memory load command packets followed by memory store command packets in the packet flow may be serialized, and a synchronization operation may be executed between the transmitter and the receiver when a packet count at the receiver satisfies a number of data packets in the packet flow.Type: GrantFiled: February 17, 2022Date of Patent: September 26, 2023Assignee: NVIDIA CORP.Inventors: Hans Eberle, Larry Robert Dennison, John Martin Snyder
-
Publication number: 20230261794Abstract: Packet flows between a transmitter and a receiver in an unreliable and unordered switched packet network may be established as a result of receiving a second packet comprising a second memory operation on a memory address. The transmission of memory load command packets followed by memory store command packets in the packet flow may be serialized, and a synchronization operation may be executed between the transmitter and the receiver when a packet count at the receiver satisfies a number of data packets in the packet flow.Type: ApplicationFiled: February 17, 2022Publication date: August 17, 2023Applicant: NVIDIA Corp.Inventors: Hans Eberle, Larry Robert Dennison, John Martin Snyder
-
Publication number: 20220417176Abstract: A method is provided for operating a network switch comprising a plurality of input ports and a plurality of output ports. The method comprises receiving a first data packet received via a first input port and a second data packet received via a second input port to be delivered to an egress endpoint connected to a first output port, configuring a plurality of crossbar switch units arranged in a tiled architecture to pass the first data packet to the first output port via a primary path and pass the second data packet to the first output port via a secondary path, and transmitting the first data packet and the second data packet to the egress endpoint. The first data packet and the second data packet pass through the plurality of crossbar switch units simultaneously.Type: ApplicationFiled: June 23, 2022Publication date: December 29, 2022Inventors: Glenn Alan Dearth, Nan Jiang, Mark D. Hummel, Gregory Michael Thorson, Karan Gupta, Dane Thomas Mrazek, Eric Anderson, Larry Robert Dennison
-
Patent number: 11502867Abstract: A network device configured to perform scalable, in-network computations is described. The network device is configured to process pull requests and/or push requests from a plurality of endpoints connected to the network. A collective communication primitive from a particular endpoint can be received at a network device. The collective communication primitive is associated with a multicast region of a shared global address space and is mapped to a plurality of participating endpoints. The network device is configured to perform an in-network computation based on information received from the participating endpoints before forwarding a response to the collective communication primitive back to one or more of the participating endpoints.Type: GrantFiled: July 24, 2020Date of Patent: November 15, 2022Assignee: NVIDIA CorporationInventors: Benjamin Klenk, Nan Jiang, Larry Robert Dennison
-
Patent number: 11470394Abstract: A communication method between a source device and a target device utilizes speculative connection setup between the source device and the target device, target-device-side packet ordering, and fine-grained ordering to remove packet dependencies.Type: GrantFiled: July 21, 2020Date of Patent: October 11, 2022Assignee: NVIDIA CORP.Inventors: Hans Eberle, Larry Robert Dennison
-
Patent number: 11463272Abstract: A network device configured to perform scalable, in-network computations is described. The network device is configured to process pull requests and/or push requests from a plurality of endpoints connected to the network. A collective communication primitive from a particular endpoint can be received at a network device. The collective communication primitive is associated with a multicast region of a shared global address space and is mapped to a plurality of participating endpoints. The network device is configured to perform an in-network computation based on information received from the participating endpoints before forwarding a response to the collective communication primitive back to one or more of the participating endpoints. The endpoints can inject pull requests (e.g., load commands) and/or push requests (e.g., store commands) into the network. A multicast capability enables tasks, such as a reduction operation, to be offloaded to hardware in the network device.Type: GrantFiled: October 6, 2021Date of Patent: October 4, 2022Assignee: NVIDIA CorporationInventors: Benjamin Klenk, Nan Jiang, Larry Robert Dennison, Gregory M. Thorson
-
Patent number: 11363339Abstract: A communication method between a source device and a target device utilizes speculative connection setup between the source device and the target device, target-device-side packet ordering, and fine-grained ordering to remove packet dependencies.Type: GrantFiled: July 20, 2020Date of Patent: June 14, 2022Assignee: NVIDIA Corp.Inventors: Hans Eberle, Larry Robert Dennison
-
Patent number: 11341369Abstract: A technique for performing data parallel training of a neural network model is disclosed that incorporates batch normalization techniques using partial populations to generate normalization parameters. The technique involves processing, by each processor of a plurality of processors in parallel, a first portion of a sub-batch of training samples allocated to the processor to generate activations for the first portion of the sub-batch. Each processor analyzes the activations and transmits statistical measures for the first portion to an additional processor that reduces the statistical measures from multiple processors to generate normalization parameters for a partial population of the training samples that includes the first portion from each of the plurality of processors. The normalization parameters are then transmitted back to each of the processors to normalize the activations for both the first portion and a second portion of the sub-batch of training samples allocated to each processor.Type: GrantFiled: October 31, 2019Date of Patent: May 24, 2022Assignee: NVIDIA CorporationInventors: Larry Robert Dennison, Benjamin Klenk
-
Patent number: 11336476Abstract: A network device configured to perform scalable, in-network computations is described. The network device is configured to process pull requests and/or push requests from a plurality of endpoints connected to the network. A collective communication primitive from a particular endpoint can be received at a network device. The collective communication primitive is associated with a multicast region of a shared global address space and is mapped to a plurality of participating endpoints. The network device is configured to perform an in-network computation based on information received from the participating endpoints before forwarding a response to the collective communication primitive back to one or more of the participating endpoints. The endpoints can inject pull requests (e.g., load commands) and/or push requests (e.g., store commands) into the network. A multicast capability enables tasks, such as a reduction operation, to be offloaded to hardware in the network device.Type: GrantFiled: July 24, 2020Date of Patent: May 17, 2022Assignee: NVIDIA CorporationInventors: Benjamin Klenk, Nan Jiang, Larry Robert Dennison, Gregory M. Thorson
-
Patent number: 11327900Abstract: Multiprocessor clusters in a virtualized environment conventionally fail to provide memory access security, which is frequently a requirement for efficient utilization in multi-client settings. Without adequate access security, a malicious process may access what might be confidential data that belongs to a different client sharing the multiprocessor cluster. Furthermore, an inadvertent programming error in the code for one client process may accidentally corrupt data that belongs to the different client. Neither scenario is acceptable. Embodiments of the present disclosure provide access security by enabling each processing node within a multiprocessor cluster to virtualize and manage local memory access and only process access requests possessing proper access credentials. In this way, different applications executing on a multiprocessor cluster may be isolated from each other while advantageously sharing the hardware resources of the multiprocessor cluster.Type: GrantFiled: July 23, 2020Date of Patent: May 10, 2022Assignee: NVIDIA CorporationInventors: Samuel Hammond Duncan, Sanjeev Jain, Mark Douglas Hummel, Vyas Venkataraman, Olivier Giroux, Larry Robert Dennison, Alexander Toichi Ishii, Hemayet Hossain, Nir Haim Arad
-
Publication number: 20220095017Abstract: A communication method between a source device and a target device utilizes speculative connection setup between the source device and the target device, target-device-side packet ordering, and fine-grained ordering to remove packet dependencies.Type: ApplicationFiled: December 1, 2021Publication date: March 24, 2022Applicant: NVIDIA Corp.Inventors: Hans Eberle, Larry Robert Dennison
-
Patent number: 11243974Abstract: A method of dynamically converting a document from a non-DICOM format to a DICOM format includes receiving a query for use in searching for the document from one or more repositories, searching the one or more repositories for the document that matches the query, receiving from the one or more repositories one or more document results of the searching, receiving from a user a request for retrieval of at least one document from the one or more document results of the searching, determining if the at least one document requested for retrieval is in the non-DICOM format, upon positive determination, converting the at least one document requested for retrieval from the non-DICOM to the DICOM format, and sending the at least one converted document to the user.Type: GrantFiled: December 29, 2016Date of Patent: February 8, 2022Assignee: Hyland Switzerland SarlInventors: Razvan Atanasiu, Otto Hunter Gasser, Larry Robert Sitka
-
Publication number: 20220029845Abstract: A network device configured to perform scalable, in-network computations is described. The network device is configured to process pull requests and/or push requests from a plurality of endpoints connected to the network. A collective communication primitive from a particular endpoint can be received at a network device. The collective communication primitive is associated with a multicast region of a shared global address space and is mapped to a plurality of participating endpoints. The network device is configured to perform an in-network computation based on information received from the participating endpoints before forwarding a response to the collective communication primitive back to one or more of the participating endpoints. The endpoints can inject pull requests (e.g., load commands) and/or push requests (e.g., store commands) into the network. A multicast capability enables tasks, such as a reduction operation, to be offloaded to hardware in the network device.Type: ApplicationFiled: October 6, 2021Publication date: January 27, 2022Inventors: Benjamin Klenk, Nan Jiang, Larry Robert Dennison, Gregory M. Thorson
-
Patent number: 11170263Abstract: A technique utilizing speculative execution and rollback for performing data parallel training of a neural network model is disclosed. Activations for a layer of the neural network model are normalized during a speculative normalization operation using estimated normalization parameters associated with a partial population of a set of training data allocated to a particular processor. Normalization parameters associated with the total population of the set of training data are generated by a distributed reduce operation in parallel with the speculative normalization operation. An optional rollback operation can revert the activations to a pre-normalization state if the estimated normalization parameters for the partial population are subsequently determined to be inaccurate compared to the normalization parameters for the population of the set of training data distributed across a plurality of processors.Type: GrantFiled: October 31, 2019Date of Patent: November 9, 2021Assignee: NVIDIA CorporationInventors: Larry Robert Dennison, Benjamin Klenk
-
Patent number: 11171798Abstract: A network device configured to perform scalable, in-network computations is described. The network device is configured to process pull requests and/or push requests from a plurality of endpoints connected to the network. A collective communication primitive from a particular endpoint can be received at a network device. The collective communication primitive is associated with a multicast region of a shared global address space and is mapped to a plurality of participating endpoints. The network device is configured to perform an in-network computation based on information received from the participating endpoints before forwarding a response to the collective communication primitive back to one or more of the participating endpoints. The endpoints can inject pull requests (e.g., load commands) and/or push requests (e.g., store commands) into the network. A multicast capability enables tasks, such as a reduction operation, to be offloaded to hardware in the network device.Type: GrantFiled: July 24, 2020Date of Patent: November 9, 2021Assignee: NVIDIA CorporationInventors: Benjamin Klenk, Nan Jiang, Larry Robert Dennison, Gregory M. Thorson