Patents by Inventor Luiz M. Franca-Neto

Luiz M. Franca-Neto has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11783176
    Abstract: Embodiments of storage device architecture for processing data using machine learning are disclosed. In some embodiments, the storage device includes a separate I/O core and a neural network core. The storage device can create a copy of data streams in which the data is stored, and the neural network core can process the copy of the data streams in a neural network while the I/O core can perform read or write functions on the data streams.
    Type: Grant
    Filed: March 25, 2019
    Date of Patent: October 10, 2023
    Inventors: Luiz M. Franca-Neto, Viacheslav Dubeyko
  • Patent number: 11769042
    Abstract: Some embodiments include a special-purpose hardware accelerator that can perform specialized machine learning tasks during both training and inference stages. For example, this hardware accelerator uses a systolic array having a number of data processing units (“DPUs”) that are each connected to a small number of other DPUs in a local region. Data from the many nodes of a neural network is pulsed through these DPUs with associated tags that identify where such data was originated or processed, such that each DPU has knowledge of where incoming data originated and thus is able to compute the data as specified by the architecture of the neural network. These tags enable the systolic neural network engine to perform computations during backpropagation, such that the systolic neural network engine is able to support training.
    Type: Grant
    Filed: December 27, 2018
    Date of Patent: September 26, 2023
    Inventor: Luiz M. Franca-Neto
  • Patent number: 11741346
    Abstract: Devices and methods for systolically processing data according to a neural network. In one aspect, a first arrangement of processing units includes at least first, second, third, and fourth processing units. The first and second processing units are connected to systolically pulse data to one another, and the third and fourth processing units are connected to systolically pulse data to one another. A second arrangement of processing units includes at least fifth, sixth, seventh, and eighth processing units. The fifth and sixth processing units are connected to systolically pulse data to one another, and the seventh and eighth processing units are connected to systolically pulse data to one another. The second processing unit is configured to systolically pulse data to the seventh processing unit along a first interconnect and the third processing unit is configured to systolically pulse data to the sixth processing unit along a second interconnect.
    Type: Grant
    Filed: May 16, 2018
    Date of Patent: August 29, 2023
    Assignee: Western Digital Technologies, Inc.
    Inventor: Luiz M. Franca-Neto
  • Patent number: 11551064
    Abstract: A method of performing computations of a neural network is disclosed comprising assigning a first processing unit to perform computations of a first node of a first layer of the neural network and assigning a second processing unit to perform computations of a second node of a second layer of the neural network. Computations of the first node are performed using the first processing unit to generate a first activation output that is transmitted to a first output systolic element of the first processing unit. The first activation output is systolically pulsed to a first input systolic element of the second processing unit and computations of the second node are performed by using the second processing unit to process at least the first activation output.
    Type: Grant
    Filed: May 16, 2018
    Date of Patent: January 10, 2023
    Assignee: Western Digital Technologies, Inc.
    Inventor: Luiz M. Franca-Neto
  • Patent number: 11494620
    Abstract: A method of computer processing is disclosed comprising receiving a data packet at a processing node of a neural network, performing a calculation of the data packet at the processing node to create a processed data packet, attaching a tag to the processed data packet, transmitting the processed data packet from the processing node to a receiving node during a systolic pulse, receiving the processed data packet at the receiving node, performing a clockwise convolution on the processed data packet and a counter clockwise convolution on the processed data packet, performing an adding function and backpropagating results of the performed sigmoid function to each of the processing nodes that originally processed the data packet.
    Type: Grant
    Filed: May 16, 2018
    Date of Patent: November 8, 2022
    Assignee: Western Digital Technologies, Inc.
    Inventor: Luiz M. Franca-Neto
  • Patent number: 11494582
    Abstract: Devices and methods for performing computations of a neural network include determining, for a particular layer of the neural network, a number of input channels and a number of output maps generated for particular pixels of a plurality of pixels. A portion of a processing chip is configured into a plurality of processing units, with the processing chip including a plurality of tensor arrays and a plurality memory cells. A particular processing unit of the plurality of processing units performs computations associated with a particular pixel. Individual processing units of the plurality of processing units each include a number of tensor arrays determined based on the number of input channels and a number of memory cells or pixel arrays corresponding to a number of output maps. The plurality of processing units is assigned to perform computations of the particular layer.
    Type: Grant
    Filed: December 27, 2018
    Date of Patent: November 8, 2022
    Assignee: Western Digital Technologies, Inc.
    Inventor: Luiz M. Franca-Neto
  • Patent number: 11461579
    Abstract: Some embodiments include a special-purpose hardware accelerator that can perform specialized machine learning tasks during both training and inference stages. For example, this hardware accelerator uses a systolic array having a number of data processing units (“DPUs”) that are each connected to a small number of other DPUs in a local region. Data from the many nodes of a neural network is pulsed through these DPUs with associated tags that identify where such data was originated or processed, such that each DPU has knowledge of where incoming data originated and thus is able to compute the data as specified by the architecture of the neural network. These tags enable the systolic neural network engine to perform computations during backpropagation, such that the systolic neural network engine is able to support training.
    Type: Grant
    Filed: December 27, 2018
    Date of Patent: October 4, 2022
    Assignee: Western Digital Technologies, Inc.
    Inventor: Luiz M. Franca-Neto
  • Patent number: 11372577
    Abstract: Embodiments of an improved memory architecture for processing data inside of a device are described. In some embodiments, the device can store neural network layers, such as a systolic flow engine, in non-volatile memory and/or a separate first memory. A processor of a host system can delegate the execution of a neural network to the device. Advantageously, neural network processing in the device can be scalable, with the ability to process large amounts of data.
    Type: Grant
    Filed: January 6, 2021
    Date of Patent: June 28, 2022
    Assignee: Western Digital Technologies, Inc.
    Inventors: Luiz M. Franca-Neto, Viacheslav Dubeyko
  • Patent number: 11218375
    Abstract: Embodiments of the present disclosure generally relate to a cloud computing network and a method of transferring information among processing nodes in a cloud computing network. In one embodiment, a cloud computing network is disclosed herein. The cloud computing network includes a plurality of motherboards arranged in racks. Each individual motherboard includes a central hub and a plurality of processing nodes equipped to the central hub. Each processing node is configured to access memory or storage space of another processing node in the same motherboard by intermediation of the hub. The access is called a communication between a pair of processing nodes. The communication includes a string of information transmitted between processing nodes. The string of information has a plurality of frames. Each frame includes a plurality of time slots, wherein each time slot is allotted a specific node pair.
    Type: Grant
    Filed: March 25, 2020
    Date of Patent: January 4, 2022
    Assignee: WESTERN DIGITAL TECHNOLOGIES, INC.
    Inventor: Luiz M. Franca-Neto
  • Patent number: 11164074
    Abstract: At least a subset of first processing units of a first arrangement of a first systolic processing chip is assigned to a first layer of a neural network and at least a subset of second processing units of a second arrangement of the first systolic processing chip is assigned to a second layer of the neural network. At least a subset of third processing units of a third arrangement of a second systolic processing chip is assigned to a third layer of the neural network. Input data is processed using the subset of the first processing units to generate first activation output values. The first activation output values are systollically pulsed to the subset of the second processing units and processed to generate second activation output values. The second activation output values are processed using the subset of the third processing units of the second systolic processing chip.
    Type: Grant
    Filed: May 16, 2018
    Date of Patent: November 2, 2021
    Assignee: Western Digital Technologies, Inc.
    Inventor: Luiz M. Franca-Neto
  • Patent number: 11164073
    Abstract: At least a subset of first processing units of a first arrangement of an array of processing units is assigned to perform computations of a first layer of a neural network and at least a subset of second processing units of a last arrangement is assigned to perform computations of a second layer. At least a second subset of the first processing units of the first arrangement is re-assigned to perform computations of a third layer. In one aspect, it is determined that a number of layers exceeds a number of arrangements of processing units of a systolic processing chip. A first arrangement of processing units of the number of arrangements is assigned to perform computations according to a first layer for a first set of forward propagations, and the first arrangement is assigned to perform computations according to a different layer for a second set of forward propagations.
    Type: Grant
    Filed: May 16, 2018
    Date of Patent: November 2, 2021
    Assignee: Western Digital Technologies, Inc.
    Inventors: Luiz M. Franca-Neto, Luis V. Cargnini
  • Patent number: 11164072
    Abstract: Devices and methods for systolically processing data according to a neural network. A first processing unit performs computations of a first node of a first layer to generate a first output and attaches a first tag to the first output identifying the first processing unit. A second processing unit performs computations of a second node of the first layer to generate a second output and attaches a second tag to the second output identifying the second processing unit. A third processing unit performs computations of a third node of a second layer including receiving the first and second outputs, using a first convolutional engine to perform a first convolution on the first output using a first weight identified by the first tag, and using a second convolutional engine of to perform a second convolution on the second output using a second weight identified by the second tag.
    Type: Grant
    Filed: May 16, 2018
    Date of Patent: November 2, 2021
    Assignee: Western Digital Technologies, Inc.
    Inventor: Luiz M. Franca-Neto
  • Publication number: 20210124524
    Abstract: Embodiments of an improved memory architecture for processing data inside of a device are described. In some embodiments, the device can store neural network layers, such as a systolic flow engine, in non-volatile memory and/or a separate first memory. A processor of a host system can delegate the execution of a neural network to the device. Advantageously, neural network processing in the device can be scalable, with the ability to process large amounts of data.
    Type: Application
    Filed: January 6, 2021
    Publication date: April 29, 2021
    Inventors: Luiz M. Franca-Neto, Viacheslav Dubeyko
  • Patent number: 10929058
    Abstract: Embodiments of an improved memory architecture by processing data inside of the memory device are described. In some embodiments, the memory device can store neural network layers, such as a systolic flow engine, in non-volatile memory and/or a separate DRAM memory. Central processing unit (CPU) of a host system can delegate the execution of a neural network to the memory device. Advantageously, neural network processing in the memory device can be scalable, with the ability to process large amounts of data.
    Type: Grant
    Filed: March 25, 2019
    Date of Patent: February 23, 2021
    Assignee: Western Digital Technologies, Inc.
    Inventors: Luiz M. Franca-Neto, Viacheslav Dubeyko
  • Patent number: 10860477
    Abstract: A method and a storage system are provided for implementing enhanced solid state storage class memory (eSCM) including a direct attached dual in line memory (DIMM) card containing Dynamic Random Access Memory (DRAM), and at least one 5 non-volatile memory, for example, Phase Change Memory (PCM), Resistive RAM (ReRAM), Spin-Transfer-Torque RAM (STT-RAM), and NAND Flash chips. An eSCM processor controls selectively allocating data among the DRAM, and the at least one non-volatile memory primarily based upon a data set size.
    Type: Grant
    Filed: September 26, 2017
    Date of Patent: December 8, 2020
    Assignee: WESTERN DIGITAL TECNOLOGIES, INC.
    Inventors: Frank R. Chu, Luiz M. Franca-Neto, Timothy K. Tsai, Qingbo Wang
  • Patent number: 10796198
    Abstract: Some embodiments include a special-purpose hardware accelerator that can perform specialized machine learning tasks during both training and inference stages. For example, this hardware accelerator uses a systolic array having a number of data processing units (“DPUs”) that are each connected to a small number of other DPUs in a local region. Data from the many nodes of a neural network is pulsed through these DPUs with associated tags that identify where such data was originated or processed, such that each DPU has knowledge of where incoming data originated and thus is able to compute the data as specified by the architecture of the neural network. These tags enable the systolic neural network engine to perform computations during backpropagation, such that the systolic neural network engine is able to support training.
    Type: Grant
    Filed: December 27, 2018
    Date of Patent: October 6, 2020
    Assignee: Western Digital Technologies, Inc.
    Inventor: Luiz M. Franca-Neto
  • Publication number: 20200311537
    Abstract: Embodiments of storage device architecture for processing data using machine learning are disclosed. In some embodiments, the storage device includes a separate I/O core and a neural network core. The storage device can create a copy of data streams in which the data is stored, and the neural network core can process the copy of the data streams in a neural network while the I/O core can perform read or write functions on the data streams.
    Type: Application
    Filed: March 25, 2019
    Publication date: October 1, 2020
    Inventors: Luiz M. Franca-Neto, Viacheslav Dubeyko
  • Publication number: 20200228411
    Abstract: Embodiments of the present disclosure generally relate to a cloud computing network and a method of transferring information among processing nodes in a cloud computing network. In one embodiment, a cloud computing network is disclosed herein. The cloud computing network includes a plurality of motherboards arranged in racks. Each individual motherboard includes a central hub and a plurality of processing nodes equipped to the central hub. Each processing node is configured to access memory or storage space of another processing node in the same motherboard by intermediation of the hub. The access is called a communication between a pair of processing nodes. The communication includes a string of information transmitted between processing nodes. The string of information has a plurality of frames. Each frame includes a plurality of time slots, wherein each time slot is allotted a specific node pair.
    Type: Application
    Filed: March 25, 2020
    Publication date: July 16, 2020
    Inventor: Luiz M. FRANCA-NETO
  • Patent number: 10644958
    Abstract: Embodiments of the present disclosure generally relate to a cloud computing network and a method of transferring information among processing nodes in a cloud computing network. In one embodiment, a cloud computing network is disclosed herein. The cloud computing network includes a plurality of motherboards arranged in racks. Each individual motherboard includes a central hub and a plurality of processing nodes equipped to the central hub. Each processing node is configured to access memory or storage space of another processing node in the same motherboard by intermediation of the hub. The access is called a communication between a pair of processing nodes. The communication includes a string of information transmitted between processing nodes. The string of information has a plurality of frames. Each frame includes a plurality of time slots, wherein each time slot is allotted a specific node pair.
    Type: Grant
    Filed: January 30, 2016
    Date of Patent: May 5, 2020
    Assignee: WESTERN DIGITAL TECHNOLOGIES, INC.
    Inventor: Luiz M. Franca-Neto
  • Patent number: 10514867
    Abstract: Various aspects directed towards facilitating error management within a shared non-volatile memory (NVM) architecture are disclosed. Data programmed into a plurality NVM cells is encoded prior to programming, and a range of programmability associated with each of the plurality of NVM cells is determined when the plurality of NVM cells are programmed A first error management scheme is then applied to NVM cells identified as limited-range programmable cells, and a second error management scheme is applied to NVM cells identified as full-range programmable cells, such that the second error management scheme is different than the first error management scheme.
    Type: Grant
    Filed: December 30, 2016
    Date of Patent: December 24, 2019
    Assignee: WESTERN DIGITAL TECHNOLOGIES, INC.
    Inventor: Luiz M. Franca-Neto