Abstract:
Embodiments of the present invention disclose a method, an apparatus and a system for joint optimization. The method for joint optimization includes: decomposing the joint optimization of an entire network into the joint optimization performed in each sub-network, regarding a bandwidth requirement for a server off the sub-network as a bandwidth requirement for a virtual server on a port, iteratively performing the joint optimization in each sub-network, and applying results of the joint optimization in the network. In the embodiments of the present invention, the bandwidth requirement for the server off the sub-network is regarded as the bandwidth requirement for the virtual server on the port, the joint optimization is iteratively performed in each sub-network, and the results of the joint optimization are applied in the network, so that the joint optimization of the entire network is performed in parallel.
Abstract:
Embodiments of the present invention provide a data processing system and a data processing method. An MMIO address of a data request is obtained, where data stored in a PCIe storage device can be directly obtained from the MMIO address, and according to the MMIO address, a network adapter can directly read the data from the PCIe storage device of the data processing system, and transmit the data to a second data processing system, or directly write the data received from the second data processing system into the PCIe storage device. Therefore, the processing system can implement transmitting data directly from the PCIe storage device to the network adapter without using memory. During data transmission between the two data processing systems, a usage ratio of a resource, such as memory and a CPU, is reduced, and efficiency of data transmission is improved.
Abstract:
Embodiments of the present invention provide a data processing system and a data processing method. An MMIO address of a data request is obtained, where data stored in a PCIe storage device can be directly obtained from the MMIO address, and according to the MMIO address, a network adapter can directly read the data from the PCIe storage device of the data processing system, and transmit the data to a second data processing system, or directly write the data received from the second data processing system into the PCIe storage device. Therefore, the processing system can implement transmitting data directly from the PCIe storage device to the network adapter without using memory. During data transmission between the two data processing systems, a usage ratio of a resource, such as memory and a CPU, is reduced, and efficiency of data transmission is improved.
Abstract:
Embodiments of the present invention disclose a cooperative caching method and apparatus, relating to the field of network technologies, to improve the local hit ratio without increasing the local server costs. The technical solution provided in the present invention includes: obtaining, according to cache information, end-to-end delay between a local server and a neighbor server, and popularity in a cache list, a consolidated gain value of a cached video segment and a consolidated gain value of a candidate video segment in the local server; and replacing the cached video segment with the candidate video segment when the consolidated gain value of the cached video segment and the consolidated gain value of the candidate video segment in the local server meet a replacement condition.
Abstract:
Embodiments of the present invention disclose a data processing method including: sending global initial statistical information to each slave node; merging received local statistical information of each slave node, to obtain new global statistical information; if Gibbs sampling performed by a slave node has ended, calculating a probability distribution between a document and topic and a probability distribution between the topic and a word according to the new global statistical information; according to the probability distributions obtained through calculation, establishing a likelihood function of a text set, and maximizing the likelihood function, to obtain a new hLDA hyper-parameter; and if iteration of solving for an hLDA hyper-parameter has converged, and according to the new hLDA hyper-parameter, calculating and outputting the probability distribution between the document and topic and the probability distribution between the topic and word.
Abstract:
Embodiments of the present invention provide a data processing system and a data processing method. An MMIO address of a data request is obtained, where data stored in a PCIe storage device can be directly obtained from the MMIO address, and according to the MMIO address, a network adapter can directly read the data from the PCIe storage device of the data processing system, and transmit the data to a second data processing system, or directly write the data received from the second data processing system into the PCIe storage device. Therefore, the processing system can implement transmitting data directly from the PCIe storage device to the network adapter without using memory. During data transmission between the two data processing systems, a usage ratio of a resource, such as memory and a CPU, is reduced, and efficiency of data transmission is improved.
Abstract:
Embodiments of the present invention provide a data processing system and a data processing method. An MMIO address of a data request is obtained, where data stored in a PCIe storage device can be directly obtained from the MMIO address, and according to the MMIO address, a network adapter can directly read the data from the PCIe storage device of the data processing system, and transmit the data to a second data processing system, or directly write the data received from the second data processing system into the PCIe storage device. Therefore, the processing system can implement transmitting data directly from the PCIe storage device to the network adapter without using memory. During data transmission between the two data processing systems, a usage ratio of a resource, such as memory and a CPU, is reduced, and efficiency of data transmission is improved.
Abstract:
A parameter inference method to solve a problem that precision of a Latent Dirichlet Allocation model is poor is provided. The method includes: calculating a Latent Dirichlet Allocation model according to a preset initial first hyperparameter, a preset initial second hyperparameter, a preset initial number of topics, a preset initial count matrix of documents and topics, and a preset initial count matrix of topics and words to obtain probability distributions; obtaining the number of topics, a first hyperparameter, and a second hyperparameter that maximize log likelihood functions of the probability distributions; and determining whether the number of topics, the first hyperparameter, and the second hyperparameter converge, and if not, putting the number of topics, the first hyperparameter, and the second hyperparameter into the Latent Dirichlet Allocation model until the optimal number of topics, an optimal first hyperparameter, and an optimal second hyperparameter that maximize the log likelihood functions of the probability distributions.
Abstract:
An Ethernet multicast method and device are provided, which relate to the communications technology field, and improve the capability of distributing multicast data in the Ethernet. The method includes: receiving a request from a host/multicast source, wherein the request carries a multicast Media Access Control (MAC) address of a destination multicast group; selecting, according to the request, a switch satisfying a particular optimization condition as a multicast root node corresponding to the multicast MAC address of the destination multicast group when it is determined that the multicast MAC address of the destination multicast group does not have the corresponding multicast root node; transmitting an identification of the multicast root node to the host/multicast source. The embodiments of the present invention are mainly applied to the process of the multicast data distribution in the Ethernet.
Abstract:
Embodiments of the present invention disclose a data processing method including: sending global initial statistical information to each slave node; merging received local statistical information of each slave node, to obtain new global statistical information; if Gibbs sampling performed by a slave node has ended, calculating a probability distribution between a document and topic and a probability distribution between the topic and a word according to the new global statistical information; according to the probability distributions obtained through calculation, establishing a likelihood function of a text set, and maximizing the likelihood function, to obtain a new hLDA hyper-parameter; and if iteration of solving for an hLDA hyper-parameter has converged, and according to the new hLDA hyper-parameter, calculating and outputting the probability distribution between the document and topic and the probability distribution between the topic and word.