Abstract:
The invention relates in particular to a system comprising: a database; and a cache layer comprising one or more cache clusters, each of the one or more cache clusters operative to: maintain, in a memory of one or more cache nodes of the cache cluster, at least a portion of a social graph, the social graph comprising a plurality of nodes and a plurality of edges connecting the nodes, each node being uniquely identified by a node identifier, and each edge indicating associations between nodes; process one or more queries received from one or more client systems of users of a social network environment, the processing of each query comprising: if the query is for associations between nodes in the portion of the social graph stored in the cache cluster, then respond to the query by searching the portion of the social graph stored in the memory of the cache cluster; and if the query is not for associations between nodes in the portion of the social graph stored in the cache cluster, then forward the query to the database for processing; and send, to the one or more client systems for display, search results responsive to the received queries; to corresponding methods and to a corresponding cache node and storage medium.
Abstract:
The use of a data model index for assisting in efficient response to a data model query. The data model query identifies the data model itself and references the schema of the data model, rather than referencing the schema of the underlying data source. The index assists by correlating previous model queries on the data model and associated response to those previous model queries. Accordingly, if a similar model query is encountered, the response from the previous model query may still be considered fresh, the index may be used to more efficiently retrieve the results for the similar model query. A data management system may include multiple data models, each having a corresponding data model index.
Abstract:
The invention relates to a computer system for both online transaction processing and online analytical processing, comprising: a processor coupled to a database, the database comprising the database comprising: a main store (116) for storing records, a differential buffer (114) for receiving and buffering added or deleted or modified records, the differential buffer being coupled to the main store, a schema comprising records stored in the main store and records stored in the differential buffer, and a cache store (112) for caching a result of a query against the schema; and a cache controller (106) executable by the processor and communicatively coupled to the database, the cache controller being configured for: storing the result of the query in the cache store; receiving an analytical request; and determining, in response to the received request, an up-to-date result of the query by (216): accessing the cache store to obtain the cached result; determining the records of the schema that have been added or deleted or modified since the step of storing the cached result in the cache store on the basis of the records stored in the differential buffer; and incrementally deriving the up-to-date result from the cached result and from the records determined in the previous step.
Abstract:
A method and apparatus for operating a database system (1) is described, the method comprising analyzing (14) a plurality of database queries (3a, 3b, 3c), generating (17) a data structure object (6) within the database of the database system (1), wherein the data structure object (6) comprises at least one data set indicated by the plurality of database queries (3a, 3b, 3c) and dynamically modifying (23) the data structure object (6) based on hardware loads.
Abstract:
A database query is modified to improve performance by identifying portions of database tables that are likely to be reused and modifying the database query accordingly. According to certain embodiments, the degree of selectivity with which a table is accessed is used to identify tables that are likely to be reused in subsequent database operations.
Abstract:
Techniques for implementing mechanisms for chaining continuous queries are provided. In some examples, results of a first continuous query may be stored in a data object. For example, the first continuous query may be configured to process business event data and store the results in the data object. Additionally, a second continuous query may then be processed based at least in part on the results stored in the data object.
Abstract:
A data cache platform maintains pre-computed database query results computed by a computation platform based on data maintained in the computation platform and is configured to determine probabilities of the pre-computed database query results being outdated, to automatically issue re-computation orders to the computation platform for updating pre- computed database query results on the basis of the determined probabilities of the pre- computed database query results being outdated and to receive the updated pre-computed database query results as results of the re-computation orders. The probability determination depends on a probabilistic model and on the occurrence of asynchronous real-time events. The real-time events are indeterministic with regard to the expiration of the cached database query results and only have a probabilistic influence on the discrepancies between the database query results maintained in the data cache platform and presumed actual database query results.
Abstract:
The invention relates to a computer system for both online transaction processing and online analytical processing, comprising: a processor coupled to a database, the database comprising the database comprising: a main store (116) for storing records, a differential buffer (114) for receiving and buffering added or deleted or modified records, the differential buffer being coupled to the main store, a schema comprising records stored in the main store and records stored in the differential buffer, and a cache store (112) for caching a result of a query against the schema; and a cache controller (106) executable by the processor and communicatively coupled to the database, the cache controller being configured for: storing the result of the query in the cache store; receiving an analytical request; and determining, in response to the received request, an up-to-date result of the query by (216): accessing the cache store to obtain the cached result; determining the records of the schema that have been added or deleted or modified since the step of storing the cached result in the cache store on the basis of the records stored in the differential buffer; and incrementally deriving the up-to-date result from the cached result and from the records determined in the previous step.
Abstract:
For a database query that defines a plurality of separate snowflake schemas, a query optimizer computes separately for each of the snowflake schemas a logical access plan for obtaining from that schema's tables a respective record set that includes the data requested from those tables by that query. The query optimizer also computes a logical access plan for obtaining the query's results from the record sets in which execution of the logical access plans thus computed will result.
Abstract:
The invention relates to a method for providing statistical data from a data warehouse (410) involving one or more data storage means and one or more processor coupled to the data storage means characterized in that it comprises the steps of: • defining a plurality of index fields (114) each index field accepting a plurality of index field values; • creating a plurality of indexed files (432) and hierarchically indexing the files as trees of indexes (300) which includes, for each tree, the steps of: - hierarchically ordering the index fields (201); - defining bins, each bin being associated to an index field and gathering one or more index field values for that index field; - creating indexes by concatenating one or more bins in compliance with the hierarchy of hierarchically ordered index fields to form a sequence of bins, an index comprising only one bin per index field; - hierarchically indexing the files as trees of indexes (300), each index having zero or more children indexes or and having at most one parent index, so that each child index comprises the same sequence of bins as the index of its parent index plus at least an additional bin associated to an additional index field;
• providing each index with a data container (325, 335) configured to store statistical data, so that each data container is indexed and is directly addressable from within the files hierarchically indexed; • receiving (436) one or more input files (434) comprised of raw data and updating the data containers with the input files (434), which includes for each input file using one or more processor arranged for performing the steps of: - identifying and extracting from the raw data at least one attribute to be analyzed through statistics and one or more input file parameters characterizing the attribute; - creating at least an individual record (620) from the input file (434), each individual record (620) comprising at least an attribute and the one or more input file parameters characterizing said attribute; - associating each input file parameter with at least an index field; - establishing a correspondence between each input file parameter and a bin of the at least one index field associated to that input file parameter; - identifying data containers indexed with the one or more bins that all correspond to an input file parameter of said individual record (620); - incrementally updating the identified data containers (325, 335) with the at least one attribute of said individual records (620) to obtain statistical data describing the attribute.