Patents by Inventor Håkan Lars-Göran PERSSON
Håkan Lars-Göran PERSSON has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 11954038Abstract: A data processing system includes a memory system, a processor and a cache system. The cache system includes a cache and a data encoder associated with the cache. The data encoder encodes blocks of uncompressed data having a particular data size for storing in the memory system. The processor is configured, when an array of data has a data size equal to the particular data size or is able to be combined with one or more other arrays of data already written to the cache to provide a plurality of arrays of data having a data size that is equal to the particular data size, to output the array of data from the processor to the data encoder, bypassing the cache, for encoding as or as part of a block of data having the particular data size.Type: GrantFiled: July 19, 2021Date of Patent: April 9, 2024Assignee: Arm LimitedInventors: Olof Henrik Uhrenholt, Håkan Lars-Göran Persson, Jakob Axel Fries
-
Publication number: 20240086340Abstract: A data processing system that comprises a processing unit and a communications bus over which bus transactions to access memory can be performed is disclosed. The system includes a codec, and the processing unit can initiate over the communications bus, bus transactions that comprise the codec accessing the memory.Type: ApplicationFiled: January 24, 2022Publication date: March 14, 2024Inventors: Håkan Lars-Göran PERSSON, Vladimir DOLZHENKO
-
Publication number: 20230236992Abstract: In response to determining circuitry determining that a portion of data to be sent to a recipient over an interconnect has a predetermined value, data sending circuitry performs data elision to: omit sending at least one data FLIT corresponding to the portion of data having the predetermined value; and send a data-elision-specifying FLIT specifying data-elision information indicating to the recipient that sending of the at least one data FLIT has been omitted and that the recipient can proceed assuming the portion of data has the predetermined value. The data-elision-specifying FLIT is a FLIT other than a write request FLIT for initiating a memory write transaction sequence. This helps to conserve data FLIT bandwidth for other data not having the predetermined value.Type: ApplicationFiled: January 21, 2022Publication date: July 27, 2023Inventors: Klas Magnus BRUCE, Jamshed JALAL, Håkan Lars-Göran PERSSON, Phanindra Kumar MANNAVA
-
Patent number: 11550620Abstract: Apparatuses and methods are disclosed for performing data processing operations in main processing circuitry and delegating certain tasks to auxiliary processing circuitry. User-specified instructions executed by the main processing circuitry comprise a task dispatch specification specifying an indication of the auxiliary processing circuitry and multiple data words defining a delegated task comprising at least one virtual address indicator. In response to the task dispatch specification the main processing circuitry performs virtual-to-physical address translation with respect to the at least one virtual address indicator to derive at least one physical address indicator, and issues a task dispatch memory write transaction to the auxiliary processing circuitry comprises the indication of the auxiliary processing circuitry and the multiple data words, wherein the at least one virtual address indicator in the multiple data words is substituted by the at least one physical address indicator.Type: GrantFiled: March 3, 2021Date of Patent: January 10, 2023Assignee: Arm LimitedInventors: Håkan Lars-Göran Persson, Frederic Claude Marie Piry, Matthew Lucien Evans, Albin Pierrick Tonnerre
-
Publication number: 20220398686Abstract: When storing data of an array of data in memory in a graphics processing system, respective memory regions are allocated for storing blocks of the data array, with the allocated region of memory for a block of the data array corresponding to a maximum possible size of the block of the data array when compressed, and being divided into a plurality of memory allocation sub-blocks, having at least one sub-block having a first, larger size and at least one sub-block having a second, smaller size. Blocks of the data array are compressed using a compression scheme, with each compressed block being stored in one or more of the sub-blocks of its allocated memory region.Type: ApplicationFiled: June 3, 2022Publication date: December 15, 2022Inventor: Håkan Lars-Göran PERSSON
-
Publication number: 20220283847Abstract: Apparatuses and methods are disclosed for performing data processing operations in main processing circuitry and delegating certain tasks to auxiliary processing circuitry. User-specified instructions executed by the main processing circuitry comprise a task dispatch specification specifying an indication of the auxiliary processing circuitry and multiple data words defining a delegated task comprising at least one virtual address indicator. In response to the task dispatch specification the main processing circuitry performs virtual-to-physical address translation with respect to the at least one virtual address indicator to derive at least one physical address indicator, and issues a task dispatch memory write transaction to the auxiliary processing circuitry comprises the indication of the auxiliary processing circuitry and the multiple data words, wherein the at least one virtual address indicator in the multiple data words is substituted by the at least one physical address indicator.Type: ApplicationFiled: March 3, 2021Publication date: September 8, 2022Inventors: Håkan Lars-Göran PERSSON, Frederic Claude Marie PIRY, Matthew Lucien EVANS, Albin Pierrick TONNERRE
-
Publication number: 20220027281Abstract: A data processing system includes a memory system, a processor and a cache system. The cache system includes a cache and a data encoder associated with the cache. The data encoder encodes blocks of uncompressed data having a particular data size for storing in the memory system. The processor is configured, when an array of data has a data size equal to the particular data size or is able to be combined with one or more other arrays of data already written to the cache to provide a plurality of arrays of data having a data size that is equal to the particular data size, to output the array of data from the processor to the data encoder, bypassing the cache, for encoding as or as part of a block of data having the particular data size.Type: ApplicationFiled: July 19, 2021Publication date: January 27, 2022Inventors: Olof Henrik Uhrenholt, Håkan Lars-Göran Persson, Jakob Axel Fries
-
Patent number: 10747681Abstract: Apparatuses and methods for address translation invalidation are provided. In an apparatus having address translation storage which stores merged address translation information for multiple address translation stages, a set of counters are provided to hold a set of counter values. Entries in the address translation storage are stored with identifiers of first and second counters selected from the set of counters in dependence on respective context information for a first stage and a second stage of address translation together with a counter value of each counter. In response to an invalidation request specifying a first or second addressing scheme invalidation context a counter of the set of counters is selected in dependence on the first or second addressing scheme invalidation context and its value is modified.Type: GrantFiled: March 18, 2019Date of Patent: August 18, 2020Assignee: Arm LimitedInventor: Håkan Lars-Göran Persson
-
Patent number: 10719632Abstract: A data processing system includes a host processor that executes an operating system and an accelerator operable to process data under the control of the operating system executing on the host processor. The accelerator can be switched between a normal mode of operation and a protected mode of operation in which the side channel information that can be provided by the accelerator to the host processor is restricted. The data processing system also includes a mechanism for switching the accelerator from its normal mode of operation to the protected mode of operation, and from its protected mode of operation to the normal mode of operation.Type: GrantFiled: August 25, 2016Date of Patent: July 21, 2020Assignee: Arm LimitedInventors: Håkan Lars-Göran Persson, Steven John Price, Thomas James Cooksey
-
Patent number: 10664399Abstract: A filter comprises interface circuitry, to intercept coherency protocol transactions exchanged between a master device comprising a first cache and an interconnect for managing coherency between the first cache and at least one other cache or other master device. The filter has filtering circuitry for filtering the coherency protocol transactions in dependence on memory access permission data defining which regions of an address space the master device is allowed to access.Type: GrantFiled: November 29, 2017Date of Patent: May 26, 2020Assignee: ARM LimitedInventors: Håkan Lars-Göran Persson, Ian Rudolf Bratt, Andrew Brookfield Swaine, Bruce James Mathewson
-
Patent number: 10423534Abstract: A cache memory, such as a translation lookaside buffer cache 16, includes a plurality of blocks of bit storage circuits 26 which can operate in either a first mode to store a plurality of shared-tagged data values having a shared tag, which his stored in a tag memory 24, or in a second mode to store a plurality of individual-tag data values and respective individual tags. The tag entries within the tag memory comprise the shared tag value for a given block operating in the first mode and a composite value for a given block operating in the second mode. The composite value includes a discriminator value indicative of the respective individual tags, such as a hash value or a Bloom filter value calculated in dependence upon the individual tags, using which potential matches with the individual tags may be identified from the discriminator value.Type: GrantFiled: December 5, 2016Date of Patent: September 24, 2019Assignee: ARM LimitedInventor: Håkan Lars-Göran Persson
-
Patent number: 10078589Abstract: Interconnect circuitry and a method of operating the interconnect circuitry are provided, where the interconnect circuitry is suitable to couple at least two master devices to a memory, each comprising a local cache. Any access to the memory mediated by the interconnect circuitry is policed by a memory protection controller situated between the interconnect circuitry and the memory. The interconnect circuitry modifies a coherency type associated with a memory transaction received from one of the master devices to a type which ensures that when a modified version of a copy of a transaction target specified by the issuing master device is stored in a local cache of another master device an access to the transaction target in the memory must take place and therefore must be policed by the memory protection controller.Type: GrantFiled: April 30, 2015Date of Patent: September 18, 2018Assignee: ARM LimitedInventors: Daniel Sara, Antony John Harris, Håkan Lars-Göran Persson, Andrew Christopher Rose, Ian Bratt
-
Publication number: 20180157590Abstract: A filter unit comprises interface circuitry, to intercept coherency protocol transactions exchanged between a master device comprising a first cache and an interconnect for managing coherency between the first cache and at least one other cache or other master device. The filter unit has filtering circuitry for filtering the coherency protocol transactions in dependence on memory access permission data defining which regions of an address space the master device is allowed to access.Type: ApplicationFiled: November 29, 2017Publication date: June 7, 2018Inventors: Håkan Lars-Göran PERSSON, Ian Rudolf BRATT, Andrew Brookfield SWAINE, Bruce James MATHEWSON
-
Publication number: 20170192900Abstract: A cache memory, such as a translation lookaside buffer cache 16, includes a plurality of blocks of bit storage circuits 26 which can operate in either a first mode to store a plurality of shared-tagged data values having a shared tag, which his stored in a tag memory 24, or in a second mode to store a plurality of individual-tag data values and respective individual tags. The tag entries within the tag memory comprise the shared tag value for a given block operating in the first mode and a composite value for a given block operating in the second mode. The composite value includes a discriminator value indicative of the respective individual tags, such as a hash value or a Bloom filter value calculated in dependence upon the individual tags, using which potential matches with the individual tags may be identified from the discriminator value.Type: ApplicationFiled: December 5, 2016Publication date: July 6, 2017Inventor: Håkan Lars-Göran PERSSON
-
Publication number: 20170060637Abstract: A data processing system includes a host processor that executes an operating system and an accelerator operable to process data under the control of the operating system executing on the host processor. The accelerator can be switched between a normal mode of operation and a protected mode of operation in which the side channel information that can be provided by the accelerator to the host processor is restricted. The data processing system also includes a mechanism for switching the accelerator from its normal mode of operation to the protected mode of operation, and from its protected mode of operation to the normal mode of operation.Type: ApplicationFiled: August 25, 2016Publication date: March 2, 2017Applicant: ARM LimitedInventors: Håkan Lars-Göran Persson, Steven John Price, Thomas James Cooksey
-
Publication number: 20160321179Abstract: Interconnect circuitry and a method of operating the interconnect circuitry are provided, where the interconnect circuitry is suitable to couple at least two master devices to a memory, each comprising a local cache. Any access to the memory mediated by the interconnect circuitry is policed by a memory protection controller situated between the interconnect circuitry and the memory. The interconnect circuitry modifies a coherency type associated with a memory transaction received from one of the master devices to a type which ensures that when a modified version of a copy of a transaction target specified by the issuing master device is stored in a local cache of another master device an access to the transaction target in the memory must take place and therefore must be policed by the memory protection controller.Type: ApplicationFiled: April 30, 2015Publication date: November 3, 2016Inventors: Daniel SARA, Antony John HARRIS, Håkan Lars-Göran PERSSON, Andrew Christopher ROSE, Ian BRATT