Abstract:
A media enhancement system includes an augmented reality (AR) device having a display, processing hardware, and a memory storing software code. The processing hardware executes the software code to monitor media content including a sequence of moving images displayed on a display screen separate from the AR device, receive playhead data indicating a playhead state of a media playout device playing out the media content, and detect, based on monitoring the media content, one or more image(s) in the sequence of moving images as one or more anchor image(s). The software code is further executed to obtain, using the anchor image(s), one or more AR effect(s) associated with the anchor image(s), and render, based on the playhead data, the AR effect(s) on the display of the AR device, wherein the AR effect(s) is/are spatially and temporally aligned with the sequence of moving images being displayed on the display screen.
Abstract:
Systems and methods generating a haptic output response are disclosed. Video content is displayed on a display. A location of a user touch on the display is detected while the video content is being displayed. A region of interest in the video content is determined based on the location of the user touch. And a haptic output response is generated to a user. A characteristic of the haptic output response is determined using one or more characteristics of the region of interest.
Abstract:
In one implementation, a virtual-world simulator includes a computing platform having a hardware processor and a memory storing a software code, a tracking system communicatively coupled to the computing platform, and a projection device communicatively coupled to the computing platform. The hardware processor is configured to execute the software code to obtain a map of a geometry of a real-world venue including the virtual-world simulator, to identify one or more virtual effects for display in the real-world venue, and to use the tracking system to track a moving perspective of one of a user in the real-world venue or a camera in the real-world venue. The hardware processor is further configured to execute the software code to control the projection device to simulate a virtual-world by conforming the identified one or more virtual effects to the geometry of the real-world venue from a present vantage point of the tracked moving perspective.
Abstract:
A system includes an augmented reality (AR) device having a first display, processing hardware, and a memory storing software code. The processing hardware is configured to execute the software code to monitor media content including a sequence of moving images displayed on a second display separate from the AR device and controlled by a media player device, detect, based on monitoring the media content, an image in the sequence of moving images for enhancement by one or more AR effects. The processing hardware further executes the software code to render the one or more AR effects on the first display, and transmit, contemporaneously with rendering the one or more AR effects on the first display, a signal configured to pause or loop the playing of the media content on the second display.
Abstract:
An image processing system includes a computing platform having processing hardware, a display, and a system memory storing a software code. The processing hardware executes the software code to receive a digital object, surround the digital object with virtual cameras oriented toward the digital object, render, using each one of the virtual cameras, a depth map identifying a distance of that one of the virtual cameras from the digital object, and generate, using the depth map, a volumetric perspective of the digital object from a perspective of that one of the virtual cameras, resulting in multiple volumetric perspectives of the digital object. The processing hardware further executes the software code to merge the multiple volumetric perspectives of the digital object to form a volumetric representation of the digital object, and to convert the volumetric representation of the digital object to a renderable form.
Abstract:
In one implementation, a virtual-world simulator includes a computing platform having a hardware processor and a memory storing a software code, a tracking system communicatively coupled to the computing platform, and a projection device communicatively coupled to the computing platform. The hardware processor is configured to execute the software code to obtain a map of a geometry of a real-world venue including the virtual-world simulator, to identify one or more virtual effects for display in the real-world venue, and to use the tracking system to track a moving perspective of one of a user in the real-world venue or a camera in the real-world venue. The hardware processor is further configured to execute the software code to control the projection device to simulate a virtual-world by conforming the identified one or more virtual effects to the geometry of the real-world venue from a present vantage point of the tracked moving perspective.
Abstract:
The present application includes a computer implemented method including at least two modes for analyzing a stereoscopic image corresponding to a two dimensional image. The method includes analyzing one or more layers of the two dimensional image to determine a depth pixel offset for every pixel in the two dimensional image and creating by the processing element a depth map, such as a gray scale map, by coloring every pixel a color shade based on the respective depth pixel offset for the pixel. The method further includes displaying on a display an output image corresponding to the stereoscopic image, receiving a first user selection corresponding a first depth pixel offset, determining a plurality of pixels of the output image corresponding to the first depth pixel offset, and applying a first identifier to the plurality of pixels on the output image corresponding to the first depth pixel offset. Additionally, in a first mode the output image displayed includes the first identifier and in a second mode the output image displayed includes the depth map and the first identifier.
Abstract:
In one implementation, a virtual-world simulator includes a computing platform having a hardware processor and a memory storing a software code, a tracking system communicatively coupled to the computing platform, and a projection device communicatively coupled to the computing platform. The hardware processor is configured to execute the software code to obtain a map of a geometry of a real-world venue including the virtual-world simulator, to identify one or more virtual effects for display in the real-world venue, and to use the tracking system to track a moving perspective of one of a user in the real-world venue or a camera in the real-world venue. The hardware processor is further configured to execute the software code to control the projection device to simulate a virtual-world by conforming the identified one or more virtual effects to the geometry of the real-world venue from a present vantage point of the tracked moving perspective.
Abstract:
Systems and methods generating a haptic output response are disclosed. Video content is displayed on a display. A location of a user touch on the display is detected while the video content is being displayed. A region of interest in the video content is determined based on the location of the user touch. And a haptic output response is generated to a user. A characteristic of the haptic output response is determined using one or more characteristics of the region of interest.
Abstract:
Techniques for using holographic imagery for eyeline reference for performers. A first computer generated object is identified for display to a first performer at a designated physical position on a set. A first holographic projection of the first computer generated object is generated using a first holographic display. The first holographic display is configured to make the first holographic projection appear, to the first performer, to be located at the designated physical position on the set. One or more images of the performer are captured using an image capture device with a field of view that encompasses both the first performer and the designated physical position on the set. The captured one or more images depict the first performer and do not depict the first holographic projection. The first computer generated object is added to the captured one or more images after the capturing.