Abstract:
Described embodiments relate to systems and methods for displaying information regarding tracked objects and receiving user input selecting the object(s) to be tracked. One embodiment takes the form of a method that includes: receiving, from a content server, a first representation of a video stream and an object-of-information identifier that indicates availability of a second representation of a portion of the video stream that depicts an object of interest; causing the display of both the first representation of the video stream and the object-of-interest identifier; responsive to a selection of the second representation of the portion of the video stream using the object-of-interest identifier, transmitting, to the content server, a request for the second representation of the portion of the video stream; and receiving and causing display of the second representation of the portion of the video stream.
Abstract:
Systems and methods are described for providing viewers of adaptive bit rate (ABR) streaming video with the option to view alternative streams in which an alternative tone mapping is applied to one or more regions of interest. The availability of streams with alternative tone mappings may be identified in a media presentation description (MPD) in an MPEG-DASH system. In some embodiments, the streaming video is divided into slices, and alternative tone mappings are applied to regions of interest within the slices. When a server receives a request from a client device for alternative tone mappings of different regions, slices with the appropriate mapping may be assembled on demand and delivered to the requestor as a single video stream. Tone mappings may be used, for example, to highlight particular players in a sporting event.
Abstract:
Described herein are methods and systems for enabling a client to effect a rapid change from displaying an unzoomed video to displaying a zoomed version of the video. In an exemplary embodiment, a video client receives (i) a primary video stream, (ii) metadata identifying a secondary video stream corresponding to a zoomed version of a region of the primary video stream, and (iii) data defining that region. In response to a user selection of the zoomed video stream, the video client requests the identified secondary video stream. Until the secondary video stream has been retrieved, the video client displays a locally-generated zoomed version of the video that is created by cropping areas outside the defined region and upscaling the defined region. Once enough of the secondary video stream has been retrieved, the video client displays the secondary video stream, which may be a higher-quality zoomed video.
Abstract:
Methods and systems are described for operating a video player, the method including receiving a manifest (such as a DASH MPD) identifying at least a first representation and a second representation of a video stream, retrieving and playing the first representation at a first playback frame rate, the first stream having a first captured frame rate, receiving a user selection of a slow-motion play mode, in response to the user selection, retrieving a second representation of the video content, the second representation having a second captured frame rate higher than the first captured frame rate, and playing the second representation at the first playback frame rate.
Abstract:
Systems and methods are described to enable video clients to zoom in to a region or object of interest without substantial loss of resolution. In an exemplary method, a server transmits a manifest, such as a DASH MPD, to a client device. The manifest identifies a plurality of sub-streams, where each sub-stream represents a respective spatial portion of a source video. The manifest also includes information associating an object of interest with a plurality of the spatial portions. To view high-quality zoomed video, the client requests the sub-streams that are associated with the object of interest and renders the requested sub-streams. In some embodiments, different sub-streams are available with different zoom ratios.
Abstract:
Systems and methods are described for providing clear areas related to objects of interest in a video display. In accordance with an embodiment, a method includes capturing, with a camera, a video frame of a scene; determining a camera orientation and camera location of the camera capturing the video; determining a location of an object of interest; mapping the location of the object of interest to a location on the video frame; determining an object-of-interest area based on the location of the object of interest on the video frame; determining a clear area on the video frame; transmitting a location of the clear area to a client device; and displaying the video frame and metadata associated with the object of interest in the clear area.
Abstract:
Systems and methods described herein relate to providing fast switching between different available video streams. In an exemplary embodiment, a user viewing a selected channel of video content receives a manifest file (such as a DASH MPD) that identifies various representations of the selected channel. The manifest file also identifies channel-change streams for one or more alternate channels. The channel-change streams may have a shorter segment size than regular streaming content. While displaying the selected content, a client also retrieves the channel-change streams of the alternate channels. If the client changes to one of the alternate channels, the client displays the appropriate channel-change stream while a regular representation of the alternate channel is being retrieved.
Abstract:
Systems and methods are described for video encoding for devices equipped with two video cameras, particularly where one of the video cameras is a zoom camera. Videos of a scene are simultaneously captured from both video cameras. Motion information (such as a motion field and/or motion vectors) collected from one video stream is used for the encoding of the other. For example, a motion vector from one video may be transformed into a grid of the other video. The transformed motion vector may be used to predict a block of pixels in the other video, or it may be used as a candidate or starting point in an algorithm for selecting a motion vector. The transformation of the motion vector may comprise aligning and scaling the vector, or other linear or nonlinear transformations may be used.
Abstract:
Systems and methods are described for enabling a consumer of streaming video to obtain different views of the video, such as zoomed views of one or more objects of interest. In an exemplary embodiment, a client device receives an original video stream along with data identifying objects of interest and their spatial locations within the original video. In one embodiment, in response to user selection of an object of interest, the client device switches to display of a cropped and scaled version of the original video to present a zoomed video of the object of interest. The zoomed video tracks the selected object even as the position of the selected object changes with respect to the original video. In some embodiments, the object of interest and the appropriate zoom factor are both selected with a single expanding-pinch gesture on a touch screen.
Abstract:
A normal broadcast video viewing experience may be augmented by providing access to enhanced views, such as zoomed or highlighted views of particular regions of interest, or partial or complete views of content with high resolution, high frame rate, high bit depth, or customized tone mapping. Such enhanced views, or zoom coded streams, may be made available over a source other than broadcast, such as a packet-switched network. Information, such as metadata, identifying the available zoom coded streams may be provided in-band in the broadcast video. A second video stream may be requested over the network using the received metadata. The second video stream may be received over the network and then displayed.