摘要:
Methods and systems are provided for providing search results while maintaining search context. A search system may display a plurality of links related to a received search query, wherein the links are all of the same entity type. The system may display sets of search results related to the plurality of links, while continuing to display the plurality of links.
摘要:
Methods and systems are provided for providing search results while maintaining search context. A search system may display a plurality of links related to a received search query, wherein the links are all of the same entity type. The system may display sets of search results related to the plurality of links, while continuing to display the plurality of links
摘要:
The method of the present invention includes maintaining a first tier 101 and a second tier 102 of devices 30 that have access to a secondary bus 42 that a PCI to PCI bridge 38 services. Each device 30 that has access to secondary PCI bus 42 is categorized into either first tier 101 or a second tier 102. The devices 30 in first tier 101 are provided more frequent opportunities to gain access to secondary PCI bus 42 than devices in low tier 102. Next, a pending transaction is recognized when an initiating device 30 that has been categorized into second tier 102 accesses secondary PCI bus 42 and attempts a transaction that crosses PCI to PCI bridge 38 to primary PCI bus 26. However, PCI to PCI bridge 38 is unable to complete the transaction on primary PCI bus 26. Therefore, PCI to PCI bridge 38 is unable to provide access to any other device 30 on secondary bus 42 until the pending transaction completes. Next, device 30 that initiated the pending transaction is categorized into first tier 101 until the pending transaction is completed.
摘要:
A liveness-based memory allocation module operating so that a program thread invoking the memory allocation module is provided with an allocation of memory including a reserve of free heap slots beyond the immediate requirements of the invoking thread. The module receives a parameter representing a thread execution window from an invoking thread; calculates a liveness metric based upon the parameter; calculates a reserve of memory to be passed to the invoking thread based upon the parameter; returns a block of memory corresponding to the calculated reserve of memory. Equations, algorithms, and sampling strategies for calculating liveness metrics are disclosed, as well as a method for adaptive control of the module to achieve a balance between memory efficiency and potential contention as specified by a single control parameter.
摘要:
Methods and systems are provided for ranking search results and generating a presentation. In some implementations, a search system generates a presentation based on a search query. In some implementations, a search system ranks search results based on data stored in a knowledge graph. In some implementations, a search system identifies a modifying concept such as a superlative in a received search query, and determines ranking properties based on the modifying concept.
摘要:
Methods and systems are provided for ranking search results and generating a presentation. In some implementations, a search system generates a presentation based on a search query. In some implementations, a search system ranks search results based on data stored in a knowledge graph. In some implementations, a search system identifies a modifying concept such as a superlative in a received search query, and determines ranking properties based on the modifying concept.
摘要:
Described is predicting cache locality in a multicore/multithreaded processing environment including when threads share cache data in a non-uniform interleaving manner. Thread execution traces are analyzed to compute a set of per-thread parameters that can then be used to predict cache miss rates for other cache sizes. In one aspect, a model is based upon a probability that the cache reuse distance will increase because of accesses by other threads, and another probability that the reuse distance will decrease because of intercept accesses by other threads to shared data blocks. Estimates of the number of shared data blocks, possibly shared data blocks and private data blocks are used in the computations.
摘要:
A method to quantify a plurality of digital data sharing in a multi-threaded execution includes the steps of: providing at least one processor; providing a computer readable non-transitory storage medium including a computer readable multi-threaded executable code and a computer readable executable code to calculate a plurality of shared footprint values and an average shared footprint value; running the multi-threaded executable code on the at least one computer processor; running the computer readable executable code configured to calculate a plurality of shared footprint values and an average shared footprint value; calculating a plurality of shared footprint values by use of a linear-time process for a corresponding plurality of executable windows in time; and calculating and saving an average shared footprint value based on the plurality of shared footprint values to quantify by a metric the data sharing by the multi-threaded execution. A system to perform the method is also described.
摘要:
A computer-implemented method for providing a definition or a translation is disclosed according to an aspect of the subject technology. The method comprises receiving an input indicating a word selected by a user, sending a definition request for the word to a definition server, receiving a response to the definition request from the definition server, sending a translation request for the word to a translation server, and receiving a response to the translation request from the translation server. The method also comprises determining whether to provide the user with a definition or a translation of the word based on the responses from the definition server and the translation server, and providing the user with the definition or the translation of the word based on the determination.