Abstract:
Technologies for framework-level audio device virtualization include a computing device that executes multiple application framework instances. The computing device monitors for an application framework instance switch and, in response to an application framework instance switch, determines whether the current application framework instance is in the foreground. If in the foreground, the computing device selects a physical audio output device. The computing device may output audio data associated with the current application framework instance using a kernel audio driver associated with the physical audio output device. If not in the foreground, the computing device selects a null audio output device using a null audio hardware abstraction layer (HAL). The null audio HAL may sleep for the duration of audio data associated with the current application framework instance. The null audio HAL may be an operating-system- and device-independent shared library of the computing device. Other embodiments are described and claimed.
Abstract:
An electric vehicle computing sharing system (100) is adapted to receive a signal indicating the electric vehicle (110, 120, 130) is connected to a charging station (115, 125, 135). The computing sharing system (100) may be further adapted to receive information about the electric vehicle (110, 120, 130). The computing sharing system (100) may be further adapted to determine a predicted charging duration (535) for the electric vehicle (110, 120, 130). The computing sharing system (100) may be further adapted to identify a task for execution by a computing resource of the electric vehicle (110, 120, 130) based on the predicted charging duration (535). The computing sharing system (100) may be further adapted to transmit the task to the electric vehicle (110, 120, 130). The computing sharing system (100) may be further adapted to receive a result for the task from the electric vehicle (110, 120, 130).
Abstract:
Autonomous robots and methods of operating the same are disclosed. An autonomous robot includes a sensor and memory including machine readable instructions. The autonomous robot further includes at least one processor to execute the instructions to generate a velocity costmap associated with an environment in which the robot is located. The processor generates the velocity costmap based on a source image captured by the sensor. The velocity costmap includes velocity information indicative of movement of an obstacle detected in the environment.
Abstract:
Technologies for framework-level audio device virtualization include a computing device that executes multiple application framework instances. The computing device monitors for an application framework instance switch and, in response to an application framework instance switch, determines whether the current application framework instance is in the foreground. If in the foreground, the computing device selects a physical audio output device. The computing device may output audio data associated with the current application framework instance using a kernel audio driver associated with the physical audio output device. If not in the foreground, the computing device selects a null audio output device using a null audio hardware abstraction layer (HAL). The null audio HAL may sleep for the duration of audio data associated with the current application framework instance. The null audio HAL may be an operating-system- and device-independent shared library of the computing device. Other embodiments are described and claimed.
Abstract:
Embodiments of the invention provide a programming model for CPU-GPU platforms. In particular, embodiments of the invention provide a uniform programming model for both integrated and discrete devices. The model also works uniformly for multiple GPU cards and hybrid GPU systems (discrete and integrated). This allows software vendors to write a single application stack and target it to all the different platforms. Additionally, embodiments of the invention provide a shared memory model between the CPU and GPU. Instead of sharing the entire virtual address space, only a part of the virtual address space needs to be shared. This allows efficient implementation in both discrete and integrated settings.
Abstract:
A computing platform may include heterogeneous processors (e.g., CPU and a GPU) to support sharing of virtual functions between such processors. In one embodiment, a CPU side vtable pointer used to access a shared object from the CPU 110 may be used to determine a GPU vtable if a GPU-side table exists. In another embodiment, a shared non-coherent region, which may not maintain data consistency, may be created within the shared virtual memory. The CPU and the GPU side data stored within the shared non-coherent region may have a same address as seen from the CPU and the GPU side. However, the contents of the CPU-side data may be different from that of GPU-side data as shared virtual memory may not maintain coherency during the run-time. In one embodiment, the vptr may be modified to point to the CPU vtable and GPU vtable stored in the shared virtual memory.
Abstract:
Embodiments of the invention provide language support for CPU-GPU platforms. In one embodiment, code can be flexibly executed on both the CPU and GPU. CPU code can offload a kernel to the GPU. That kernel may in turn call preexisting libraries on the CPU, or make other calls into CPU functions. This allows an application to be built without requiring the entire call chain to be recompiled. Additionally, in one embodiment data may be shared seamlessly between CPU and GPU. This includes sharing objects that may have virtual functions. Embodiments thus ensure the right virtual function gets invoked on the CPU or the GPU if a virtual function is called by either the CPU or GPU.
Abstract:
Method, systems and apparatuses may provide for technology that divides an application into a plurality of portions that are each associated with one or more functions of the application and determine a plurality of transition probabilities between the plurality of portions. Some technology may also receive at least a first portion of the plurality of portions, and receive a relation file indicating the plurality of transition probabilities between the plurality of portions.
Abstract:
Disclosed is a method that may include hosting, by a virtual machine manager of a local machine, a virtual machine having a device driver. The method may include obtaining, by the virtual machine manager, from a stub driver on the remote machine, information about the I/O device on the remote machine. The I/O device on the remote machine may be bound to the stub driver on the remote machine. The method may include instantiating, by the virtual machine manager, a virtual I/O device on the local machine corresponding to the I/O device on the remote machine. The method may include collaborating, by the virtual machine manager, with the stub driver on the remote machine to effectuate a real access to the I/O device on the remote machine for an access to the virtual I/O device by the device driver on behalf of a program on the local machine. Other embodiments may be described and claimed.
Abstract:
An audio accelerator includes a decoder to decode first and second sets of data blocks, a processor to process the first and second sets of decoded data blocks, a storage area to store the first and second sets of processed data blocks, and a controller to generate interrupt signals for controlling operation of the decoder. The controller may control a rate at which data blocks are to be decoded by the decoder to reduce a time gap between outputting adjacent ones of the data blocks from the first and second sets in the storage area.