Abstract:
An apparatus for generating a 3D presentation is described wherein the 3D presentation (that could include a 3D movie, a VR (virtual reality) and/or an AR (augmented reality) content) from original 3D content including original audio and visual components, forming a plurality of scenes, each scene being associated with a respective environment or environments having respective audio characteristics. The original audio components in at least some of the scenes are modified by the apparatus to conform with or compensate for the acoustic characteristics of the respective scene and/or the environment of the location where the presentation is taking place.
Abstract:
An apparatus for generating a presentation from content having original audio and video components is described wherein an environment detector is configured to output an environment-type signal indicating a detected particular environment. An acoustics memory is configured to output selected acoustic characteristics indicative of the environment identified by the environment-type signal. An audio processor receives the audio components and the acoustic characteristics and operates to modify the original audio components to produce modified audio components based on the selected acoustic characteristics. The presentation including the modified audio components is output.
Abstract:
An apparatus generating audio cues for content indicative of the position of audio objects within the content comprising: an audio processor receiving raw audio tracks for said content and information indicative of the positions of at least some of said audio tracks within frames of said content, said audio processor generating corresponding audio parameters; an authoring tool receiving said audio parameters and generating encoding coefficients, said audio parameters including audio cue of the position of audio objects corresponding to said tracks in at least one spatial dimension; and a first audio/video encoder receiving an input and encoding said input into an audio visual content having visual objects and audio objects, said audio objects being disposed at location corresponding to said one spatial position, said encoder using said encoding coefficients for said encoding.
Abstract:
An apparatus generating audio cues for content indicative of the position of audio objects within the content comprising: an audio processor receiving raw audio tracks for said content and information indicative of the positions of at least some of said audio tracks within frames of said content, said audio processor generating corresponding audio parameters; an authoring tool receiving said audio parameters and generating encoding coefficients, said audio parameters including audio cue of the position of audio objects corresponding to said tracks in at least one spatial dimension; and a first audio/video encoder receiving an input and encoding said input into an audio visual content having visual objects and audio objects, said audio objects being disposed at location corresponding to said one spatial position, said encoder using said encoding coefficients for said encoding.
Abstract:
The present invention pertains to an apparatus and method for adding a graphic element, such as a subtitle to selected locations of the frames in a 3D movie. The authoring tool receives a depth map indicating the position of various objects in the frames of 3D content along a Z-axis. The authoring device then designates a position for at least one additional graphic element in at least some of the frames, these positions being determined in relation either to the positions of the objects or the position of the screen along said Z-axis. An encoder uses parameters from the authoring tool to reauthor the 3D movie by adding the graphic content to the positions designated by the parameters.
Abstract:
An apparatus for generating a presentation from content having original audio and video components is described wherein an environment detector is configured to output an environment-type signal indicating a detected particular environment. An acoustics memory is configured to output selected acoustic characteristics indicative of the environment identified by the environment-type signal. An audio processor receives the audio components and the acoustic characteristics and operates to modify the original audio components to produce modified audio components based on the selected acoustic characteristics. The presentation including the modified audio components is output.
Abstract:
An apparatus generating audio cues for content indicative of the position of audio objects within the content comprising:an audio processor receiving raw audio tracks for said content and information indicative of the positions of at least some of said audio tracks within frames of said content, said audio processor generating corresponding audio parameters;an authoring tool receiving said audio parameters and generating encoding coefficients, said audio parameters including audio cue of the position of audio objects corresponding to said tracks in at least one spatial dimension; anda first audio/video encoder receiving an input and encoding said input into an audio visual content having visual objects and audio objects, said audio objects being disposed at location corresponding to said one spatial position, said encoder using said encoding coefficients for said encoding.
Abstract:
An apparatus generating audio cues for content indicative of the position of audio objects within the content comprising: an audio processor receiving raw audio tracks for said content and information indicative of the positions of at least some of said audio tracks within frames of said content, said audio processor generating corresponding audio parameters; an authoring tool receiving said audio parameters and generating encoding coefficients, said audio parameters including audio cue of the position of audio objects corresponding to said tracks in at least one spatial dimension; and a first audio/video encoder receiving an input and encoding said input into an audio visual content having visual objects and audio objects, said audio objects being disposed at location corresponding to said one spatial position, said encoder using said encoding coefficients for said encoding.