Patents by Inventor Robert Pawlowski

Robert Pawlowski has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20250110918
    Abstract: Techniques for offloading function streams are described. In some examples, a function is a sequence of instructions and a stream is a sequence of functions. In some examples, a co-processor is to handle functions and/or function streams provided by a main processor. In some examples, the co-processor includes a plurality of execution resources that at least include one or more of a direct memory access (DMA) engine, an atomic engine, and a collectives engine.
    Type: Application
    Filed: September 30, 2023
    Publication date: April 3, 2025
    Inventors: Robert PAWLOWSKI, Vincent CAVE, Fabio CHECCONI, Scott CLINE, Shruti SHARMA
  • Patent number: 12204901
    Abstract: Techniques for operating on an indirect memory access instruction, where the instruction accesses a memory location via at least one indirect address. A pipeline processes the instruction and a memory operation engine generates a first access to the at least one indirect address and a second access to a target address determined by the at least one indirect address. A cache memory used with the pipeline and the memory operation engine caches pointers. In response to a cache hit when executing the indirect memory access instruction, operations dereference a pointer to obtain the at least one indirect address, not set a cache bit, and return data for the instruction without storing the data in the cache memory; and in response to a cache miss, operations set the cache bit, obtain, and store a cache line for a missed pointer, and return data without storing the data in the cache memory.
    Type: Grant
    Filed: June 25, 2021
    Date of Patent: January 21, 2025
    Assignee: Intel Corporation
    Inventors: Robert Pawlowski, Sriram Aananthakrishnan, Jason Howard, Joshua Fryman
  • Patent number: 12158852
    Abstract: Systems, methods, and apparatuses for direct memory access instruction set architecture support for flexible dense compute using a reconfigurable spatial array are described.
    Type: Grant
    Filed: June 25, 2021
    Date of Patent: December 3, 2024
    Assignee: Intel Corporation
    Inventors: Robert Pawlowski, Bharadwaj Krishnamurthy, Shruti Sharma, Byoungchan Oh, Jing Fang, Daniel Klowden, Jason Howard, Joshua Fryman
  • Patent number: 12153932
    Abstract: Examples include techniques for an in-network acceleration of a parallel prefix-scan operation. Examples include configuring registers of a node included in a plurality of nodes on a same semiconductor package. The registers to be configured responsive to receiving an instruction that indicates a logical tree to map to a network topology that includes the node. The instruction associated with a prefix-scan operation to be executed by at least a portion of the plurality of nodes.
    Type: Grant
    Filed: December 21, 2020
    Date of Patent: November 26, 2024
    Assignee: Intel Corporation
    Inventors: Ankit More, Fabrizio Petrini, Robert Pawlowski, Shruti Sharma, Sowmya Pitchaimoorthy
  • Publication number: 20240241645
    Abstract: Systems, apparatuses and methods may provide for technology that includes a plurality of hash management buffers corresponding to a plurality of pipelines, wherein each hash management buffer in the plurality of hash management buffers is adjacent to a pipeline in the plurality of pipelines, and wherein a first hash management buffer is to issue one or more hash packets associated with one or more hash operations on a hash table. The technology may also include a plurality of hash engines corresponding to a plurality of dynamic random access memories (DRAMs), wherein each hash engine in the plurality of hash engines is adjacent to a DRAM in the plurality of DRAMs, and wherein one or more of the hash engines is to initialize a target memory destination associated with the hash table and conduct the one or more hash operations in response to the one or more hash packets.
    Type: Application
    Filed: March 29, 2024
    Publication date: July 18, 2024
    Inventors: Robert Pawlowski, Shruti Sharma, Fabio Checconi, Sriram Aananthakrishnan, Jesmin Jahan Tithi, Jordi Wolfson-Pou, Joshua B. Fryman
  • Publication number: 20240160580
    Abstract: This disclosure describes systems, methods, and devices related to a global address space (VEGAS) approach. The device may execute at least two processes within a device in a computing environment, each process running on a respective compute block of at least two compute blocks. The device may manage allocations of virtual memory spaces for the least two compute blocks using an independent logical system separate from the at least two compute blocks. The device may isolate the virtual memory spaces of the at least two processes by allowing each compute block to access only its own allocated virtual memory space.
    Type: Application
    Filed: June 13, 2023
    Publication date: May 16, 2024
    Inventors: Gurpreet Singh KALSI, Joshua FRYMAN, Jason HOWARD, Robert PAWLOWSKI
  • Patent number: 11960922
    Abstract: In an embodiment, a processor comprises: an execution circuit to execute instructions; at least one cache memory coupled to the execution circuit; and a table storage element coupled to the at least one cache memory, the table storage element to store a plurality of entries each to store object metadata of an object used in a code sequence. The processor is to use the object metadata to provide user space multi-object transactional atomic operation of the code sequence. Other embodiments are described and claimed.
    Type: Grant
    Filed: September 24, 2020
    Date of Patent: April 16, 2024
    Assignee: Intel Corporation
    Inventors: Joshua B. Fryman, Jason M. Howard, Ibrahim Hur, Robert Pawlowski
  • Publication number: 20240119015
    Abstract: Systems, apparatuses and methods may provide for technology that detects a condition in which a plurality of atomic instructions target a common address and different bit positions in a mask, generates a combined read-lock request for the plurality of atomic instructions in response to the condition, and sends the combined read-lock request to a lock buffer coupled to a memory device associated with the common address.
    Type: Application
    Filed: August 30, 2023
    Publication date: April 11, 2024
    Inventors: Shruti Sharma, Robert Pawlowski
  • Publication number: 20240069921
    Abstract: Technology described herein provides a dynamically reconfigurable processing core. The technology includes a plurality of pipelines comprising a core, where the core is reconfigurable into one of a plurality of core modes, a core network to provide inter-pipeline connections for the pipelines, and logic to receive a morph instruction including a target core mode from an application running on the core, determine a present core state for the core, and morph, based on the present core state, the core to the target core mode. In embodiments, to morph the core, the logic is to select, based on the target core mode, which inter-pipeline connections are active, where each pipeline includes at least one multiplexor via which the inter-pipeline connections are selected to be active. In embodiments, to morph the core, the logic is further to select, based on the target core mode, which memory access paths are active.
    Type: Application
    Filed: September 29, 2023
    Publication date: February 29, 2024
    Inventors: Scott Cline, Robert Pawlowski, Joshua Fryman, Ivan Ganev, Vincent Cave, Sebastian Szkoda, Fabio Checconi
  • Publication number: 20240045829
    Abstract: Techniques for multi-dimensional network sorted array merging. A first switch of a plurality of switches of an apparatus may receive a first element of a first array and a first element of a second array. The first switch may determine that the first element of the first array is less than the first element of the second array. The first switch may cause the first element of the first array to be stored as a first element of an output array.
    Type: Application
    Filed: April 5, 2023
    Publication date: February 8, 2024
    Applicant: Intel Corporation
    Inventors: Robert Pawlowski, Sriram Aananthakrishnan
  • Publication number: 20240028555
    Abstract: Techniques for multi-dimensional network sorted array intersection. A first switch of a plurality of switches of an apparatus may receive a first element of a first array from a first compute tile of the plurality of compute tiles and a first element of a second array from a second compute tile of the plurality of compute tiles. The first switch may determine that the first element of the first array is equal to the first element of the second array. The first switch may cause the first element of the first array to be stored as a first element of an output array, the output array to comprise an intersection of the first array and the second array.
    Type: Application
    Filed: September 29, 2023
    Publication date: January 25, 2024
    Inventors: Robert Pawlowski, Sriram Aananthakrishna, Shruti Sharma
  • Publication number: 20240020253
    Abstract: Systems, apparatuses and methods may provide for technology that detects a plurality of sub-instruction requests from a first memory engine in a plurality of memory engines, wherein the plurality of sub-instruction requests are associated with a direct memory access (DMA) data type conversion request from a first pipeline, wherein each sub-instruction request corresponds to a data element in the DMA data type conversion request, and wherein the first memory engine is to correspond to the first pipeline, decodes the plurality of sub-instruction requests to identify one or more arguments, loads a source array from a dynamic random access memory (DRAM) in a plurality of DRAMs, wherein the operation engine is to correspond to the DRAM, and conducts a conversion of the source array from a first data type to a second data type in accordance with the one or more arguments.
    Type: Application
    Filed: September 29, 2023
    Publication date: January 18, 2024
    Inventors: Shruti Sharma, Robert Pawlowski, Fabio Checconi, Jesmin Jahan Tithi
  • Publication number: 20230333998
    Abstract: Systems, apparatuses and methods may provide for technology that includes a plurality of memory engines corresponding to a plurality of pipelines, wherein each memory engine in the plurality of memory engines is adjacent to a pipeline in the plurality of pipelines, and wherein a first memory engine is to request one or more direct memory access (DMA) operations associated with a first pipeline, and a plurality of operation engines corresponding to a plurality of dynamic random access memories (DRAMs), wherein each operation engine in the plurality of operation engines is adjacent to a DRAM in the plurality of DRAMs, and wherein one or more of the plurality of operation engines is to conduct the one or more DMA operations based on one or more bitmaps.
    Type: Application
    Filed: May 5, 2023
    Publication date: October 19, 2023
    Inventors: Shruti Sharma, Robert Pawlowski, Fabio Checconi, Jesmin Jahan Tithi
  • Publication number: 20230315451
    Abstract: Systems, apparatuses and methods may provide for technology that detects, by an operation engine, a plurality of sub-instruction requests from a first memory engine in a plurality of memory engines, wherein the plurality of sub-instruction requests are associated with a direct memory access (DMA) bitmap manipulation request from a first pipeline, wherein each sub-instruction request corresponds to a data element in the DMA bitmap manipulation request, and wherein the first memory engine is to correspond to the first pipeline. The technology also detects, by the operation engine, one or more arguments in the plurality of sub-instruction requests, sends, by the operation engine, one or more load requests to a DRAM in the plurality of DRAMs in accordance with the one or more arguments, and sends, by the operation engine, one or more store requests to the DRAM in accordance with the one or more arguments, wherein the operation engine is to correspond to the DRAM.
    Type: Application
    Filed: May 31, 2023
    Publication date: October 5, 2023
    Inventors: Shruti Sharma, Robert Pawlowski, Fabio Checconi, Jesmin Jahan Tithi
  • Patent number: 11630691
    Abstract: Disclosed embodiments relate to an improved memory system architecture for multi-threaded processors. In one example, a system includes a system comprising a multi-threaded processor core (MTPC), the MTPC comprising: P pipelines, each to concurrently process T threads; a crossbar to communicatively couple the P pipelines; a memory for use by the P pipelines, a scheduler to optimize reduction operations by assigning multiple threads to generate results of commutative arithmetic operations, and then accumulate the generated results, and a memory controller (MC) to connect with external storage and other MTPCs, the MC further comprising at least one optimization selected from: an instruction set architecture including a dual-memory operation; a direct memory access (DMA) engine; a buffer to store multiple pending instruction cache requests; multiple channels across which to stripe memory requests; and a shadow-tag coherency management unit.
    Type: Grant
    Filed: August 24, 2021
    Date of Patent: April 18, 2023
    Assignee: Intel Corporation
    Inventors: Robert Pawlowski, Ankit More, Jason M. Howard, Joshua B. Fryman, Tina C. Zhong, Shaden Smith, Sowmya Pitchaimoorthy, Samkit Jain, Vincent Cave, Sriram Aananthakrishnan, Bharadwaj Krishnamurthy
  • Publication number: 20230095207
    Abstract: A memory architecture may provide support for any number of direct memory access (DMA) operations at least partially independent of the CPU coupled to the memory. DMA operations may involve data movement between two or more memory locations and may involve minor computations. At least some DMA operations may include any number of atomic functions, and at least some of the atomic functions may include a corresponding return value. A system includes a first direct memory access (DMA) engine to request a DMA operation. The DMA operation includes an atomic operation. The system also includes a second DMA engine to receive a return value associated with the atomic operation and store the return value at a source memory.
    Type: Application
    Filed: September 24, 2021
    Publication date: March 30, 2023
    Inventors: Robert Pawlowski, Fabio Checconi, Fabrizio Petrini
  • Publication number: 20220414038
    Abstract: Systems, methods, and apparatuses for direct memory access instruction set architecture support for flexible dense compute using a reconfigurable spatial array are described.
    Type: Application
    Filed: June 25, 2021
    Publication date: December 29, 2022
    Inventors: ROBERT PAWLOWSKI, BHARADWAJ KRISHNAMURTHY, SHRUTI SHARMA, BYOUNGCHAN OH, JING FANG, DANIEL KLOWDEN, JASON HOWARD, JOSHUA FRYMAN
  • Publication number: 20220413855
    Abstract: Techniques for operating on an indirect memory access instruction, where the instruction accesses a memory location via at least one indirect address. A pipeline processes the instruction and a memory operation engine generates a first access to the at least one indirect address and a second access to a target address determined by the at least one indirect address. A cache memory used with the pipeline and the memory operation engine caches pointers. In response to a cache hit when executing the indirect memory access instruction, operations dereference a pointer to obtain the at least one indirect address, not set a cache bit, and return data for the instruction without storing the data in the cache memory; and in response to a cache miss, operations set the cache bit, obtain, and store a cache line for a missed pointer, and return data without storing the data in the cache memory.
    Type: Application
    Filed: June 25, 2021
    Publication date: December 29, 2022
    Inventors: Robert Pawlowski, Sriram Aananthakrishnan, Jason Howard, Joshua Fryman
  • Patent number: 11360809
    Abstract: Embodiments of apparatuses, methods, and systems for scheduling tasks to hardware threads are described. In an embodiment, a processor includes a multiple hardware threads and a task manager. The task manager is to issue a task to a hardware thread. The task manager includes a hardware task queue to store a descriptor for the task. The descriptor is to include a field to store a value to indicate whether the task is a single task, a collection of iterative tasks, and a linked list of tasks.
    Type: Grant
    Filed: June 29, 2018
    Date of Patent: June 14, 2022
    Assignee: Intel Corporation
    Inventors: William Paul Griffin, Joshua Fryman, Jason Howard, Sang Phill Park, Robert Pawlowski, Michael Abbott, Scott Cline, Samkit Jain, Ankit More, Vincent Cave, Fabrizio Petrini, Ivan Ganev
  • Publication number: 20220100508
    Abstract: Embodiments of apparatuses and methods for copying and operating on matrix elements are described. In embodiments, an apparatus includes a hardware instruction decoder to decode a single instruction and execution circuitry, coupled to hardware instruction decoder, to perform one or more operations corresponding to the single instruction. The single instruction has a first operand to reference a base address of a first representation of a source matrix and a second operand to reference a base address of second representation of a destination matrix. The one or more operations include copying elements of the source matrix to corresponding locations in the destination matrix and filling empty elements of the destination matrix with a single value.
    Type: Application
    Filed: December 25, 2020
    Publication date: March 31, 2022
    Applicant: Intel Corporation
    Inventors: Robert Pawlowski, Ankit More, Vincent Cave, Sriram Aananthakrishnan, Jason M. Howard, Joshua B. Fryman