Patents by Inventor Michael P. MESNIER
Michael P. MESNIER has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20220214810Abstract: In one embodiment, a device includes interface circuitry and processing circuitry. The interface circuitry communicates with a plurality of storage devices associated with a storage system. The processing circuitry receives a request to write a data object to the storage system. The data object includes a set of data elements, and the storage system is organized into blocks and shards, which are distributed across the storage devices. The processing circuitry determines a storage layout for the data object, which arranges the set of data elements across a set of blocks and shards with padding to align each data element within block and shard boundaries. The processing circuitry writes the data object to the storage system based on the storage layout.Type: ApplicationFiled: March 24, 2022Publication date: July 7, 2022Inventors: Ian F. Adams, Neha Agrawal, Michael P. Mesnier
-
Publication number: 20220188028Abstract: In one embodiment, a system comprises a host processor and a storage system. The storage system comprises one or more storage devices, and each storage device comprises a non-volatile memory and a compute offload controller. The non-volatile memory stores data, and the compute offload controller performs compute tasks on the data based on compute offload commands from the host processor.Type: ApplicationFiled: March 12, 2020Publication date: June 16, 2022Applicant: Intel CorporationInventors: Michael P. Mesnier, John S. Keys, Ian F. Adams, Yi Zou, Luis Carlos Maria Remis, Daniel Robert McLeran, Mariusz Barczak, Arun Raghunath, Lay Wai Kong
-
Publication number: 20210232605Abstract: In accordance with some embodiments, classification of input/output requests from a database to a storage system may be performed. Each input/output request may be associated with a database class, and each database class may be mapped to a quality of service policy. Thus, quality of service may be enforced such that different data blocks within the storage system of the database may be afforded appropriate quality of service.Type: ApplicationFiled: April 13, 2021Publication date: July 29, 2021Applicant: Intel CorporationInventors: Michael P. Mesnier, Tian Luo, Feng Chen
-
Publication number: 20210117134Abstract: Techniques for storage and processing for distributed file systems are disclosed. In the illustrative embodiment, padding is placed between data elements in a file to be stored on a distributed file system. The file is to be split into several objects in order to be stored in the distributed file system, and the padding is used to prevent a data element from being split across two different objects. The objects are stored on data nodes, which analyze the objects to determine which data elements are present in the object as well at the location of those objects. The location of the objects is saved on the data storage device, and those locations can be used to perform queries on the data elements in the object on the data storage device itself. Such an approach can reduce transfer of data elements from data storage to local memory of the data node.Type: ApplicationFiled: December 23, 2020Publication date: April 22, 2021Inventors: John S. Keys, Daniel R. McLeran, Ian F. Adams, Michael P. Mesnier, Nilesh N. Shah
-
Patent number: 10817176Abstract: Embodiments of the present disclosure may relate to a compute offload controller that may include a parser to parse a received compute offload command, and identify a block-based compute descriptor based at least in part on the compute offload command. In some embodiments, the compute offload controller may further include an offload executor to perform an operation on data in a block-based storage device based at least in part on the block-based compute descriptor. In some embodiments, the block-based compute descriptor may include a virtual input object, a virtual output object, and a compute type identifier. Other embodiments may be described and/or claimed.Type: GrantFiled: June 18, 2018Date of Patent: October 27, 2020Assignee: Intel CorporationInventors: Ian F. Adams, John Keys, Michael P. Mesnier, Dave Minturn
-
Patent number: 10659532Abstract: Technologies for reducing latency variation of stored data object requests include a proxy computing node communicatively coupled to a plurality of storage nodes. The proxy computing node is configured to determine whether to chunk a data object corresponding to a data object request received by the proxy computing node. Accordingly, the proxy computing node is configured to obtain a retrieval latency of the storage node and determine whether to chunk the data object based on the retrieval latency. The proxy computing node is further configured to, subsequent to a determination to chunk the data object, determine a chunk size (i.e., a portion of the data object) to be retrieved from the storage node and a remaining size of the data object at the storage node after the portion of the data object corresponding to the chunk request is received. Other embodiments are described and claimed.Type: GrantFiled: September 26, 2015Date of Patent: May 19, 2020Assignee: Intel CorporationInventors: Arun Raghunath, Michael P. Mesnier
-
Patent number: 10503654Abstract: Methods and apparatus related to framework and/or methodology for selective caching of Erasure Coded fragments in a distributed storage system are described. In one embodiment, a plurality of fragments of a data object are generated. Each of the plurality of fragments is Erasure Coded (EC) prior to storage at a storage node of a plurality of storage nodes. Each of the plurality of fragments is transmitted with a caching hint to indicate whether that fragment is to be cached at the storage node. Other embodiments are also disclosed and claimed.Type: GrantFiled: September 1, 2016Date of Patent: December 10, 2019Assignee: Intel CorporationInventors: Arun Raghunath, Michael P. Mesnier, Yi Zou
-
Patent number: 10313256Abstract: Embodiments of apparatuses and methods for adaptive data compression and associated contextual information are described. In various embodiments, an apparatus may include a context monitoring module to gather contextual information for transmission of data and a policy module to gather user preference on cost associated with transmission of data. The apparatus may further include an analysis module to determine whether to compress data prior to transmission, based at least in part on the contextual information and the user preference. Other embodiments may be described and/or claimed.Type: GrantFiled: May 21, 2015Date of Patent: June 4, 2019Assignee: Intel CorporationInventors: Ren Wang, Weishuang Zhao, Alexander W. Min, Michael P. Mesnier, Richard Chuang, Tsung-Yuan C. Tai, Scott D. Hahn
-
Publication number: 20190042093Abstract: Embodiments of the present disclosure may relate to a compute offload controller that may include a parser to parse a received compute offload command, and identify a block-based compute descriptor based at least in part on the compute offload command. In some embodiments, the compute offload controller may further include an offload executor to perform an operation on data in a block-based storage device based at least in part on the block-based compute descriptor. In some embodiments, the block-based compute descriptor may include a virtual input object, a virtual output object, and a compute type identifier. Other embodiments may be described and/or claimed.Type: ApplicationFiled: June 18, 2018Publication date: February 7, 2019Inventors: Ian F. Adams, John Keys, Michael P. Mesnier, Dave Minturn
-
Publication number: 20180189148Abstract: A multiple data protection scheme logic enables applications to use objects encoded with different data protection schemes in a single namespace. Instead of configuring data protection policy supporting only one data protection scheme in a single namespace, flexible data protection policies that allow different data protection schemes in the single namespace promote more efficient use of storage and processor resources. Smaller objects can use data protection schemes that favor more efficient processor performance over increased storage costs such as replication, whereas larger objects can use data protection schemes that favor decreased storage costs over less efficient processor performance such as erasure coding.Type: ApplicationFiled: December 30, 2016Publication date: July 5, 2018Inventors: Ian F. Adams, Michael P. MESNIER, Arun RAGHUNATH
-
Publication number: 20180060242Abstract: Methods and apparatus related to framework and/or methodology for selective caching of Erasure Coded fragments in a distributed storage system are described. In one embodiment, a plurality of fragments of a data object are generated. Each of the plurality of fragments is Erasure Coded (EC) prior to storage at a storage node of a plurality of storage nodes. Each of the plurality of fragments is transmitted with a caching hint to indicate whether that fragment is to be cached at the storage node. Other embodiments are also disclosed and claimed.Type: ApplicationFiled: September 1, 2016Publication date: March 1, 2018Applicant: Intel CorporationInventors: Arun Raghunath, Michael P. Mesnier, Yi Zou
-
Patent number: 9842156Abstract: In accordance with some embodiments, classification of input/output requests from a database to a storage system may be performed. Each input/output request may be associated with a database class, and each database class may be mapped to a quality of service policy. Thus, quality of service may be enforced such that different data blocks within the storage system of the database may be afforded appropriate quality of service.Type: GrantFiled: December 29, 2011Date of Patent: December 12, 2017Assignee: Intel CorporationInventors: Michael P. Mesnier, Tian Luo, Feng Chen
-
Patent number: 9740635Abstract: Computer-readable storage media, computing devices and methods associated with file cache management are discussed herein. In embodiments, a computing device may include a file cache and a file cache manager coupled with the file cache. The file cache manager may be configured to implement a context-aware eviction policy to identify a candidate file for deletion from the file cache, from a plurality of individual files contained within the file cache, based at least in part on file-level context information associated with the individual files. In embodiments, the file-level context information may include an indication of access recency and access frequency associated with the individual files. In such embodiments, identifying the candidate file for deletion from the file cache may be based, at least in part, on both the access recency and the access frequency of the individual files. Other embodiments may be described and/or claimed.Type: GrantFiled: March 12, 2015Date of Patent: August 22, 2017Assignee: Intel CorporationInventors: Ren Wang, Weishuang Zhao, Wei Shen, Michael P. Mesnier, Tsung-Yuan C. Tai, Mesut A. Ergin
-
Publication number: 20170228170Abstract: Storage class memory may be used in an architecture to achieve high performance, high reliability, high compatibility. In some embodiments, reads may be handled in a conventional way used in a memory based model. However writes do not use a memory based model but instead correspond to a storage based model. The hybrid nature can be achieved by setting the storage class memory to be write protected so that all writes must go through a software based block device interface. In some embodiments, the software based block device interface prevents erroneous writes to the storage class memory.Type: ApplicationFiled: April 5, 2017Publication date: August 10, 2017Inventors: Feng Chen, Michael P. Mesnier
-
Patent number: 9619174Abstract: Storage class memory may be used in an architecture to achieve high performance, high reliability, high compatibility. In some embodiments, reads may be handled in a conventional way used in a memory based model. However writes do not use a memory based model but instead correspond to a storage based model. The hybrid nature can be achieved by setting the storage class memory to be write protected so that all writes must go through a software based block device interface. In some embodiments, the software based block device interface prevents erroneous writes to the storage class memory.Type: GrantFiled: December 30, 2011Date of Patent: April 11, 2017Assignee: Intel CorporationInventors: Feng Chen, Michael P. Mesnier
-
Publication number: 20170093975Abstract: Technologies for managing data object requests in a storage node cluster include a proxy computing node communicatively coupled to the cluster of storage nodes. The proxy computing node is configured to receive data object requests from a communicatively coupled client computing device and identify a plurality of storage nodes of the cluster at which the data object of the data object request is stored. The proxy computing node is further configured to determine which of the identified storage nodes from which to retrieve the stored data object and transmit a request for the data object. Additionally, the proxy computing node is configured to estimate a request completion time based on a service time and a wait time for each of the identified storage nodes, as well as identify which of the storage nodes to retrieve the stored data object from based on the estimated request completion times. Other embodiments are described and claimed.Type: ApplicationFiled: September 26, 2015Publication date: March 30, 2017Inventors: Arun Raghunath, Michael P. Mesnier, Paul Brett
-
Publication number: 20170093976Abstract: Technologies for reducing latency variation of stored data object requests include a proxy computing node communicatively coupled to a plurality of storage nodes. The proxy computing node is configured to determine whether to chunk a data object corresponding to a data object request received by the proxy computing node. Accordingly, the proxy computing node is configured to obtain a retrieval latency of the storage node and determine whether to chunk the data object based on the retrieval latency. The proxy computing node is further configured to, subsequent to a determination to chunk the data object, determine a chunk size (i.e., a portion of the data object) to be retrieved from the storage node and a remaining size of the data object at the storage node after the portion of the data object corresponding to the chunk request is received. Other embodiments are described and claimed.Type: ApplicationFiled: September 26, 2015Publication date: March 30, 2017Inventors: Arun Raghunath, Michael P. Mesnier
-
Publication number: 20160344646Abstract: Embodiments of apparatuses and methods for adaptive data compression and associated contextual information are described. In various embodiments, an apparatus may include a context monitoring module to gather contextual information for transmission of data and a policy module to gather user preference on cost associated with transmission of data. The apparatus may further include an analysis module to determine whether to compress data prior to transmission, based at least in part on the contextual information and the user preference. Other embodiments may be described and/or claimed.Type: ApplicationFiled: May 21, 2015Publication date: November 24, 2016Inventors: Ren Wang, Weishuang Zhao, Alexander W. Min, Michael P. Mesnier, Richard Chuang, Tsung-Yuan C. Tai, Scott D. Hahn
-
Publication number: 20160314188Abstract: In accordance with some embodiments, classification of input/output requests from a database to a storage system may be performed. Each input/output request may be associated with a database class, and each database class may be mapped to a quality of service policy. Thus, quality of service may be enforced such that different data blocks within the storage system of the database may be afforded appropriate quality of service.Type: ApplicationFiled: June 30, 2016Publication date: October 27, 2016Inventors: Michael P. Mesnier, Tian Luo, Feng Chen
-
Publication number: 20160267020Abstract: Computer-readable storage media, computing devices and methods associated with file cache management are discussed herein. In embodiments, a computing device may include a file cache and a file cache manager coupled with the file cache. The file cache manager may be configured to implement a context-aware eviction policy to identify a candidate file for deletion from the file cache, from a plurality of individual files contained within the file cache, based at least in part on file-level context information associated with the individual files. In embodiments, the file-level context information may include an indication of access recency and access frequency associated with the individual files. In such embodiments, identifying the candidate file for deletion from the file cache may be based, at least in part, on both the access recency and the access frequency of the individual files. Other embodiments may be described and/or claimed.Type: ApplicationFiled: March 12, 2015Publication date: September 15, 2016Inventors: Ren Wang, Weishuang Zhao, Wei Shen, Michael P. Mesnier, Tsung-Yuan C. Tai, Mesut A. Ergin