摘要:
Methods and apparatus to perform process placement for distributed applications are disclosed. An example method comprises determining a mapping between a communication graph representative of communications of a distributed application and a topology graph representative of communication costs associated with a computing network, and executing the distributed application with the processes of the distributed application assigned to the processing entities of the computing network based upon the mapping.
摘要:
Methods and apparatus to perform process placement for distributed applications are disclosed. An example method comprises determining a mapping between a communication graph representative of communications of a distributed application and a topology graph representative of communication costs associated with a computing network, and executing the distributed application with the processes of the distributed application assigned to the processing entities of the computing network based upon the mapping.
摘要:
Methods, apparatus, and articles of manufacture to perform runtime trace filtering associated with application performance analysis are disclosed. A disclosed example method involves generating a first performance value based on first performance data associated with a first function of a first application process. A difference value is generated based on the first performance value and a historical performance value associated with the first function. The difference value is compared to a threshold value, and first trace data associated with execution of the first application process is collected based on the comparison of the difference value to the threshold value.
摘要:
Methods, apparatus, and articles of manufacture to perform runtime trace filtering associated with application performance analysis are disclosed. A disclosed example method involves generating a first performance value based on first performance data associated with a first function of a first application process. A difference value is generated based on the first performance value and a historical performance value associated with the first function. The difference value is compared to a threshold value, and first trace data associated with execution of the first application process is collected based on the comparison of the difference value to the threshold value.
摘要:
An MEMS microphone is provided in the present disclosure. The MEMS microphone includes a protective structure comprising a housing and a PCB substrate covering the housing to form a receiving space, the housing is provided with a sound hole, an MEMS chip with a back cavity, received in the receiving space and fixed on the PCB substrate, the back cavity is communicated with the sound hole, and the MEMS chip comprises a first surface away from the PCB substrate and a second surface opposite to the first surface; and a waterproof part, bonded to the first surface of the MEMS chip.
摘要:
A MEMS microphone is disclosed. The MEMS microphone includes an encapsulation structure provided with an accommodation space; a MEMS chip for detecting sound signal accommodated in the accommodation space; an ASIC chip received in the accommodation space. The ASIC chip includes a signal processing module connected to MEMS chip for processing the sound signal detected by the MEMS chip and outputting the processed sound signal. The MEMS microphone further includes a temperature detection module for detecting temperature signal and outputting the temperature signal.
摘要:
A method and system for maintaining release consistency in shared memory programming on a computing device having multiple processing units includes, in response to a page fault, initiating a transfer, from one processing unit to another, of data associated with more than one but less than all of the pages of shared memory.
摘要:
A computer system may comprise a computer platform and input-output devices. The computer platform may include a plurality of heterogeneous processors comprising a central processing unit (CPU) and a graphics processing unit (GPU) and a shared virtual memory supported by a physical private memory space of at least one heterogeneous processor or a physical shared memory shared by the heterogeneous processor. The CPU (producer) may create shared multi-version data and store such shared multi-version data in the physical private memory space or the physical shared memory. The GPU (consumer) may acquire or access the shared multi-version data.
摘要:
A page table entry dirty bit system may be utilized to record dirty information for a software distributed shared memory system. In some embodiments, this may improve performance without substantially increasing overhead because the dirty bit recording system is already available in certain processors. By providing extra bits, coherence can be obtained with respect to all the other uses of the existing page table entry dirty bits.
摘要:
Embodiments of the invention provide a programming model for CPU-GPU platforms. In particular, embodiments of the invention provide a uniform programming model for both integrated and discrete devices. The model also works uniformly for multiple GPU cards and hybrid GPU systems (discrete and integrated). This allows software vendors to write a single application stack and target it to all the different platforms. Additionally, embodiments of the invention provide a shared memory model between the CPU and GPU. Instead of sharing the entire virtual address space, only a part of the virtual address space needs to be shared. This allows efficient implementation in both discrete and integrated settings.