Abstract:
Computer-readable storage media, computing device and methods associated with dynamic modification of a rendering of a physical scene. In embodiments, one or more computer-readable storage media may have instructions stored thereon which, when executed by a computing device, may provide the computing device with a dynamic augmentation module. The dynamic augmentation module may, in some embodiments, cause the computing device to receive a manipulation of a physical scene. In response to receipt of the manipulation, the dynamic augmentation module may cause the computing device to dynamically modify a rendering of the physical scene. In some embodiments, this may be accomplished through real-time application of one or more virtual articles to the rendering of the physical scene or alteration of one or more virtual articles added to the rendering of the physical scene. Other embodiments may be described and/or claimed.
Abstract:
Systems and methods may provide for receiving a short range signal from a sensor that is collocated with a short range display and using the short range signal to detect a user interaction. Additionally, a display response may be controlled with respect to a long range display based on the user interaction. In one example, the user interaction includes one or more of an eye gaze, a hand gesture, a face gesture, a head position or a voice command, that indicates one or more of a switch between the short range display and the long range display, a drag and drop operation, a highlight operation, a click operation or a typing operation.
Abstract:
A mechanism is described for facilitating smart measurement of body dimensions despite loose clothing and/or other obscurities according to one embodiment. A method of embodiments, as described herein, includes capturing, by one or more capturing/sensing components of a computing device, a scan of a body of a user, and computing one or more primary measurements relating to one or more primary areas of the body, where the one or more primary measurements are computed based on depth data of the one or more primary areas of the body, where the depth data is obtained from the scan. The method may further include receiving at least one of secondary measurements and a three-dimensional (3D) avatar of the body based on the primary measurements, and preparing a report including body dimensions of the body based on at least one of the secondary measurements and the 3D avatar, and presenting the report at a display device.
Abstract:
An augmented reality (AR) device includes a 3D video camera to capture video images and corresponding depth information, a display device to display the video data, and an AR module to add a virtual 3D model to the displayed video data. A depth mapping module generates a 3D map based on the depth information, a dynamic scene recognition and tracking module processes the video images and the 3D map to detect and track a target object within a field of view of the 3D video camera, and an augmented video rendering module renders an augmented video of the virtual 3D model dynamically interacting with the target object. The augmented video is displayed on the display device in real time. The AR device may further include a context module to select the virtual 3D model based on context data comprising a current location of the augmented reality device.
Abstract:
Systems, apparatuses and methods may provide for technology that processes an inference workload in a first subset of layers of a neural network that prevents or inhibits data dependent branch operations, conducts an exit determination as to whether an output of the first subset of layers satisfies one or more exit criteria, and selectively bypasses processing of the output in a second subset of layers of the neural network based on the exit determination. The technology may also speculatively initiate the processing of the output in the second subset of layers while the exit determination is pending. Additionally, when the inference workloads include a plurality of batches, the technology may mask one or more of the plurality of batches from processing in the second subset of layers.
Abstract:
Systems, apparatuses, and/or methods to augment reality. An object identifier may identify an object in a field of view of a user that includes a reflection of the user from a reflective surface, such as a surface of a traditional mirror. In addition, a reality augmenter may generate an augmented reality object based on the identification of the object. In one example, eyeglasses including a relatively transparent display screen may be coupled with an image capture device on the user and the augmented reality object may be observable by the user on the transparent display screen when the user wears the eyeglasses. A localizer may position the augmented reality object on the transparent display screen relative to the reflection of the user that passes though the transparent display screen during natural visual perception of the reflection by the user.
Abstract:
An augmented reality (AR) device includes a 3D video camera to capture video images and corresponding depth information, a display device to display the video data, and an AR module to add a virtual 3D model to the displayed video data. A depth mapping module generates a 3D map based on the depth information, a dynamic scene recognition and tracking module processes the video images and the 3D map to detect and track a target object within a field of view of the 3D video camera, and an augmented video rendering module renders an augmented video of the virtual 3D model dynamically interacting with the target object. The augmented video is displayed on the display device in real time. The AR device may further include a context module to select the virtual 3D model based on context data comprising a current location of the augmented reality device.
Abstract:
A narrative presentation system may include at least one optical sensor capable of detecting objects added to the field-of-view of the at least one optical sensor. Using data contained in signals received from the at least one optical sensor, an adaptive narrative presentation circuit identifies an object added to the field-of-view and identifies an aspect of a narrative presentation logically associated with the identified object. The adaptive narrative presentation circuit modifies the aspect of the narrative presentation identified as logically associated with the identified object.
Abstract:
Disclosed in some examples are methods systems and machine readable mediums in which actions or states of a first user (e.g., natural interactions) having a first corresponding computing device are observed by a sensor on a second computing device corresponding to a second user. A notification describing the observed actions or states of a first user may be shared across a network with the first corresponding computing device. In this way, the first computing device may be provided with information concerning the state of the user without having to directly sense the user.
Abstract:
Computer-readable storage media, computing devices and methods are discussed herein. In embodiments, a computing device may include one or more display devices, a digital content module coupled with the one or more display devices, and an augmentation module coupled with the digital content module and the one or more display devices. The digital content module may be configured to cause a portion of textual content to be rendered on the one or more display devices. The textual content may be associated with a digital scene that may be utilized to augment the textual content. The augmentation module may be configured to dynamically adapt the digital scene, based at least in part on a real-time video feed, to be rendered on the one or more display devices to augment the textual content. Other embodiments may be described and/or claimed.