Abstract: Embodiments of the present disclosure provide methods and devices for inserting data in a database. The method can include caching one or more rows of data in a delta table in response to a request to insert the one or more rows in a destination table, wherein the destination table is one of a group of destination tables of a first database, and wherein the delta table contains requested changes of rows of the destination table, the changes being caused by a data operation; determining a priority of the destination table based on runtime information of the destination table relative to the group of destination tables; and scheduling the merging of the inserted rows from the delta table to the destination table based on the determined priority of the destination table. The destination table is in a format of a file storing one or more blocks, each block having a set of sequence of elements clustered from each column of the destination table.
Abstract: Various methods, apparatuses/systems, and media for implementing a data book application module is disclosed. The processor identifies an application that needs to be scanned through a data factory; receives inventories of all servers and databases associated with the data factory; scans the servers and databases for receiving inventories of schema, tables and columns associated with the application; and applies artificial intelligence (AI) and/or machine learning (ML) routines and matching algorithms for matching contents of columns to predefined logical terms. The processor also converts the contents of columns into taxonomies associated with the predefined logical terms; matches the taxonomies with the corresponding predefined logical terms; assigns a probability of accuracy value to the matched terms; and populates a data catalog with the matched terms when the assigned probability of accuracy value satisfies a predetermined threshold value.
Abstract: Automatically extracting lineage data for distributed file systems and databases by generating a file folder including data and having a file path, generating a data table associated with the data and the file path, tracking the lineage of data into the file folder and the lineage of data from the data table, and recording an overall lineage comprising the lineage of data into the file folder and from the data table.
Type:
Grant
Filed:
October 7, 2019
Date of Patent:
March 15, 2022
Assignee:
International Business Machines Corporation
Inventors:
Lukasz Jakub Palus, Grzegorz Piotr Szczepanik, Krzysztof Rudek, Michal Bodziony
Abstract: An analytic model generated in a data scientist model acquisition toolbench is received. An analytic model is translated using a structuring tool. An execution simulation for testing the translated analytic model is performed for executing the model in the data scientist model acquisition toolbench.
Abstract: The present disclosure is directed to a modified GraphQL server that enables application developers to define custom GraphQL schema fragments. For example, a developer may provide the GraphQL server with a GraphQL standard definition language (SDL) file having annotations that indicate that a customized data-handling script, such as a custom data-fetching script or a custom type resolver script, be called for handling certain data types or fields. These customized data-handling scripts are resolved and called at runtime based on the annotations in the GraphQL SDL file. By enabling developers to define schema fragments in the SDL language and implement data-handling scripts in a suitable scripting language, schema portability can be maintained across client instances, as well as instance release versions.
Abstract: A system comprises a data processing unit (DPU) integrated circuit having programmable processor cores and hardware-based accelerators configured for processing streams of data units; and software executing on one or more of the processing cores. In response to a request to perform an operation on a set of one or more data tables, each having one or more columns of data arranged in a plurality of rows, the software configures the DPU to: input at least a portion of the rows of each of the database tables as at least one or more streams of data units, process the one or more streams of data units with the hardware-based accelerators to apply one or more of compression, encoding or encryption to produce a resultant stream of data units; and write the resultant stream of data units to a storage in a tree data structure.
Type:
Grant
Filed:
September 26, 2019
Date of Patent:
March 1, 2022
Assignee:
Fungible, Inc.
Inventors:
Satyanarayana Lakshmipathi Billa, Rajan Goyal
Abstract: A staging engine of a staging server receives a request to change a production database from a client device. The staging engine of the staging server accesses one or more schemas corresponding to the production database and determines one or more migration commands based on the received request and the accessed one or more schemas. The one or more migration commands correspond to a difference between a current structure of the production database and a final structure of the production database after the production database is updated. The staging engine transmits the one or more migration commands to a migration engine, wherein the migration engine asynchronously applies changes to the production database according to the one or more migration commands.
Abstract: The system described herein provides for iteratively executing an algorithm across partitions of a database table, in parallel. An instance of the algorithm may be executed for each partition of a database table across different hosts, in parallel, using predefined control data. The system may receive a result corresponding to each partition in response to executing algorithm on the respective partition. The system may determine a result corresponding to each of the partitions satisfies predefined criteria. The predefined criteria is defined by the user. In response to determining that the result corresponding to a partition fails to satisfy the predefined criteria, the system may modify the control data and execute an instance of the algorithm on each of the partitions using the modified control data.
Abstract: A data storage method includes obtaining at least one data record, each data record includes one carrier identifier and at least one label value, the at least one data record is stored in a data table, and the data table is used to record a correspondence between a carrier identifier and a label value; and creating a bitmap index corresponding to the at least one data record, where the bitmap index includes at least one bitmap; each bitmap corresponds to one label value; each bitmap includes at least one bitmap position, and each bitmap position is used to record whether a carrier corresponding to one carrier identifier has a label value corresponding to a current bitmap.
Type:
Grant
Filed:
January 21, 2020
Date of Patent:
February 15, 2022
Assignee:
HUAWEI TECHOLOGIES CO., LTD.
Inventors:
Jieshan Bi, Chaoqiang Zhong, Ming Lin, Zhenfeng Ji
Abstract: One or more implementations of the present specification disclose a database analysis information generation method, including: reading data in a to-be-merged data level and data in a merging data level; generating a new merging data level by merging the data in the to-be-merged data level and the data in the merging data level in a merging process; and performing an operation of generating new database analysis information based on data in the new merging data level and data in existing database analysis information in the process of generating the new merging data level.
Abstract: Methods, non-transitory computer readable media, and computing devices that manages clusters of a storage system using a cloud resident orchestrator includes receiving an application programming interface (API) request from a volume manager device associated with storage volumes. Control messages in a queue intended for the storage volumes are identified. A message processing operation for each of the identified control messages is determined. The determined message processing operation is performed using the volume manager device associated with storage volumes.
Type:
Grant
Filed:
April 23, 2019
Date of Patent:
February 15, 2022
Assignee:
NETAPP, INC.
Inventors:
Abhishek Naidu, Ravindra Kuramkote, Paul Ngan, Paul Yuedong Mu
Abstract: There are provided a data archiving method and system, which can process primary archiving capable of rapidly searching for compressed and stored data by compressing and storing, in a table of a database or another database, at least some of data stored in the database included in a target system, and can process secondary archiving for data for which search is no longer requested or for which a search request is low by storing at least some of the data compressed and stored in the table in the form of a file according to a preset condition.
Abstract: Embodiments described herein provide a mechanism that translates a natural language question to a database query format that may be applied to a data table to generate an answer to the natural language question. The system may identify key terms from a natural language question and classify the key terms as variable names or operation names. The natural language question is than translated into a format of question template containing variable names and operation names. In this way, the system may map the template question to a database query which can be applied to operate on a relational database to identify a cell value that represents an answer to the natural language question.
Abstract: A method of controlling read and write access to a memory structure involves initiating a read lock by obtaining a reader pool ID for a thread from a fixed pool of readers, waiting for a writer to finish by entering a wait-loop and querying a scheduler to reschedule the thread if current wait time exceeds a threshold value, declaring a resource to be read, checking for an active write lock and returning the reader pool ID for the thread. Initiating a write-lock involves checking for an active write lock flag and an active read lock flag and entering a wait-loop if the active write lock flag or the active read lock flag is present, and querying a scheduler to reschedule the thread if the wait time exceeds the threshold value.
Type:
Grant
Filed:
March 22, 2019
Date of Patent:
January 25, 2022
Assignee:
Substrate Inc.
Inventors:
Christian Beaumont, Behnaz Beaumont, Jouke van der Maas, Jan Drake
Abstract: Systems and techniques for garbage collection are disclosed for concurrently performing a garbage collection cycle in a single traversal of a garbage collection heap while application threads are running. The garbage collection cycle includes marking a first memory object as live. The garbage collection cycle also includes determining that a forwarding pointer of the first memory object points to the first memory object. The garbage collection cycle further includes evacuating the first memory object to a free region based on the determining. The garbage collection cycle additionally includes evacuating a second memory object in the same single traversal of the garbage collection heap in which the first memory object is being marked live.
Abstract: A de-serialized data store and transaction system includes a content addressable memory mapped to an n-tuple-based multimodal data structure, the n-tuple-based multimodal data structure supporting heterogeneous nested data types including a streamable abstract syntax tree and a self-describing operational interface comprising commands transmitted via a plurality of nested s-expressions.
Type:
Grant
Filed:
March 22, 2019
Date of Patent:
January 11, 2022
Assignee:
Substrate Inc.
Inventors:
Christian Beaumont, Behnaz Beaumont, Jouke van der Maas, Jan Drake
Abstract: A system, method, and computer-readable medium for proving feedback on database instructions, identifying, for example, existing patterns and providing suggested replacement instructions. This may have the effect of improving the efficiency of instructions used to create and/or manipulate databases. According to some aspects, these and other benefits may be achieved by parsing received instructions into an organizational structure, traversing the organizational structure for known patterns, and suggesting replacement patterns. In implementation, this may be effected by receiving one or more sets of known patterns and corresponding replacement patterns, parsing received instructions, comparing the known patterns with the parsed instructions, and providing suggested replacement patterns based on one or more known patterns matching the parsed instructions. A benefit of may include reducing Cartesian products during the merging of tables.
Type:
Grant
Filed:
April 14, 2020
Date of Patent:
January 4, 2022
Assignee:
Capital One Services, LLC
Inventors:
Dennis J. Mire, Puneet Goyal, Siddharth Gupta, Srinivas Kumar, Deepak Sundararaj, Oron Hazi
Abstract: Automatic generation and insertion of operations into a timeline of operations via interactions with a data visualization. A timeline structure comprising a plurality of operation records that identify successive operations performed against a table of data in response to user inputs is maintained. A data visualization is presented that presents a portion of the table of data in a particular state after a most recent operation is performed. One or more user inputs are received via user interaction with the data visualization. An operation against the table of data is performed to change the state of the table of data from the particular state to a new state. A new data visualization is presented that presents a portion of the table of data in the new state, and the timeline structure is altered to reflect the operation performed against the table of data.
Abstract: System, method, and apparatus for providing and using a centralized Link Metadata System to identify if navigational links are pop-out links, is provided. The present invention offers a Link Metadata System, available on the Internet, of metadata associated with any given, specific Internet domain and URL or URI path combination. Software browsing user agents on remote window servers are then configured with the ability to request metadata from the Link Metadata System indicating that a link is a pop-out link. Upon receiving metadata, the user agent determines whether to perform navigation as normal or perform the pop-out link functionality of reloading the remote client web browser's top level container with the content provided at the link's destination URL or URI.
Abstract: An object encoding and computation method in a database system, where a global object sample space is constituted based on correlation between a plurality of object sample spaces in a database. A global dictionary table is created in the global object sample space to provide a globally unique encoding value for each object, set the encoding value as an object identifier (ID) of the object, and ensure that the object ID and object description information are in a one-to-one mapping relationship. The database object is encoded and compressed based on the global dictionary table. Further, the object ID is directly used for database query computation.