Abstract:
To address needs of journalists and other editors, an audiovisual editing system associates segments of audiovisual materials with segments of related text. The editing system permits an editor to edit both the audiovisual materials of the segments and to edit the related text. The data structure representing the story includes a list of media segments and a list a text segments and a correspondence among the segments. Each media segment has a type, such as script, SOT and voice, and includes a list of video clips. These data structures enable the editing system to provide a simple and intuitive interface for editing a story, and can be transformed in a straightforward manner into a conventional, multitrack sequence used to represent an audiovisual composition. The editing system automatically defines parameters, such as designating the correct audio tracks, which minimize the chance for errors and speeds the workflow. Automatic ducking of these tracks also can be provided. For media segments with associated script in a corresponding text segment, an estimated time for reading the related script aloud by the newscaster may be determined. This estimated time for reading the text related a script segment may be compared to the actual duration of the audiovisual material of the script segment. The difference in the duration of the audiovisual material and the duration of the related text may be displayed to the editor, making it easier for the editor to match video to a script. Similarly, the duration of voice and SOT clips can be compared to the duration of video clips in a media segment. The difference in the duration of the video clips and the voice and SOT clips may be displayed to the editor, making it easier for the editor to match the video to the audio. Stories can be stored as text documents in a markup language. Data representing the audiovisual segments can be embedded in these stories. The editing system also may be interco ected through a computer network in a newsroom to a variety of other systems including asset management system, shared storage for media and a story database. By having the editing system store stories in the story database and sequences in the asset management system, both sequences and stories are made available to other editing systems.
Abstract:
A structured watermark may be embedded in data by applying an irregular mapping of variations defined by the structured watermark to frequency domain values representing the data. In particular, the frequency domain representation of the data comprises an ordered set of frequency domain values. The structured watermark is used to define an ordered set of variations to be applied to the frequency domain values. Each variation is a value defined by the structured watermark. An irregular mapping from positions in the ordered set of variations to positions in the ordered set of frequency domain values is defined. This irregular mapping is one-to-one and invertible. Application of the irregular mapping to the set of variations results in a set of values that may appear to be noise both in the frequency domain and in the signal domain of the data. The signal domain of the data may be n-dimensional, and may be a spatial, temporal or other domain from which data may be converted to the frequency domain. The signal domain of the data may be continuous or discrete. Each frequency domain value is modified by the variation mapped to the position of the frequency domain value by the irregular mapping. A frequency domain value may be modified using an additive or multiplicative operation. Using additive embedding, the modifications to the frequency domain values may be effected in the signal domain without computing the frequency domain values of the data by transforming, to the signal domain, the results of applying the irregular mapping to the set of variations. The watermark may be detected in target data by using the inverse of the irregular mapping on the frequency domain representation of the target data. Because the watermark is structured, it may be perceptible in the target data after the target data is processed by the inverse of the irregular mapping. A similarity metric, such as correlation, also can be used to detect the presence of the structured watermark in the processed target data.
Abstract:
Creation of a program with interactive content and time-based media would be improved by having several people working simultaneously on both the interactive content and the time-based media. The range of types of data that can be associated with the time-based media also can include a variety of types of content that enhance the time-based media, including data that is opaque to the editing system. To support editing of such programs, enhanced content may be edited into a program as either a clip or a locator, according to a user's selection. The components that represent enhanced content also may be edited in the same way as components that represent time-based media. Enhanced content may be represented by enhancement files that include data defining enhancements. An enhancement includes an identifier of the enhancement within the enhancement file, and a reference to a file defining enhanced content associated with the enhancement. Such enhancement files may be imported into a bin of an editing system to create components that describe the enhanced content referenced by the enhancements. The enhanced content may be modified using native applications for the enhanced content during editing of the program. The components describing the enhanced content in the editing system can be updated using identifiers of the enhanced content that are stored by the components. Enhanced content also may be automatically synchronized with time-based media.
Abstract:
A digital multimedia newsroom production system allows users of the system to create, browse and catalog multimedia assets. The system includes a multimedia capture and encoding system that captures multimedia data, and substantially simultaneously provides a first compressed version of the multimedia data having a first resolution, and a second compressed version of the multimedia data having a second resolution that is different from the first resolution; a multimedia storage system, coupled to the multimedia capture and encoding system, that stores multimedia information including the first and second compressed versions of the multimedia data; and a video editing and playback system coupled to the multimedia storage system. The video editing and playback system includes editing circuitry that generates a composition that uses a portion of the first compressed version, and playback circuitry that plays the composition using a portion of the second compressed version that corresponds to the portion of the first compressed version. The multimedia storage system stores multimedia information including the compressed versions of the multimedia data, and provides to a network the first compressed version of the multimedia data substantially simultaneously as the first compressed version is stored.
Abstract:
An apparatus (10) indicating the position of an adjustable lever (18), such as a control lever in a digital audio console, is described. A frame member (12, 13, 15, 17) constrains movement of the adjustable lever (18). The lever is operatively connected to an encoder apparatus (22), such that movement of the lever (18) causes movement of the encoder apparatus (22) to provide an indication of position of the lever (18) within the frame member (12, 13, 15, 17).
Abstract:
The method for adjusting quality during image capture includes computing a discrete cosine transform of a digital image to create DCT coefficients. A quantization table is generated that specifies frequency bands to be filtered and the DCT coefficients are digitized using the quantization table. It is preferred that the DCT coefficients be ordered in a zig-zag sequence to facilitate run-length encoding.
Abstract:
A data buffer that compensates the differences in data rates, between a storage device and an image compression processor. A method and apparatus for the real time indexing of frames in a video data sequence.
Abstract:
In a live audio presentation system, a stage rack and a front of house rack are interconnected by a transmission medium that transmits digital audio data and control information between them. Packets of digital audio data and control information are sent at the audio sampling rate. The packets of data are encoded using an encoding technique that provides unique codes that may be detected, thus enabling clock recovery of a clock signal embedded in the data packets. Data packets may include data for error checking. A serial digital transmission medium may be used as the transmission medium. Such a transmission medium uses low voltage signals and supports high bit rates. The system is synchronized by the FOH rack which transmits data to the stage rack according to a local audio sample clock or an external audio sample clock to which it synchronizes. This clock is embedded in data packets sent to the stage rack. The stack rack recovers the audio sample clock from data received from the FOH rack and synchronizes its receive and transmit operations to this recovered clock. The FOH rack recovers the audio sample clock from data received from the stage rack and synchronizes its receive operations to this recovered clock. The FOH rack may include an embedded computer that processes the digital audio data, and may include DSP processing that handles so-called plug-in software. A digital audio workstation or sequencer also may be connected to the FOH rack to either record the live presentation or to inject edited, recorded audio into the live presentation.
Abstract:
High quality intraframe-only compression of video can be achieved using rate distortion optimization and without resizing or bit depth modification. The compression process involves transforming portions of the image to generate frequency domain coefficients for each portion. A bit rate for each transformed portion using a plurality of scale factors is determined. Distortion for each portion is estimated according to the plurality of scale factors. A scale factor is selected for each portion to minimize the total distortion in the image to achieve a desired bit rate. A quantization matrix is selected according to the desired bit rate. The frequency domain coefficients for each portion are quantized using the selected plurality of quantizers as scaled by the selected scab factor for the portion. The quantized frequency domain coefficients are encoded using a variable length encoding to provide compressed data for each of the defined portions. The compressed data is output for each of the defined portions to provide a compressed bitstream at the desired bit rate. The bit stream format includes, for each image, a picture header followed by image scan data. The image scan data includes data corresponding to a plurality of macroblock rasterscans. The data for each macrobbck rasterscan includes data for a plurality of macroblocks for a band of lines in the image followed by padding. The padding ensures that data for each macroblock rasterscan terminates on a data boundary. The picture header references an image scan index that indicates a number of macroblock rasterscans in the image scan data and a number of lines per macroblock rasterscan, followed by entries of the index. Each entry in the index includes an offset of the macroblock rasterscan in image scan.
Abstract:
A display color calibration system uses a spectrophotometer or other sensor connect to a computer along with a computer program running on the computer to provide a mechanism to calibrate and standardize the color performance of display devices used by video applications. Such display devices include both computer displays connected to the computer and video displays connected to the computer. In particular, test colors are presented as an ordered set of samples controlled by the computer program and are presented via a video application to the display devices. The video application may be, for example, software for a digital nonlinear video editing system (NLE) or any other video playback engine on the computer. Once calibrated, these display devices can present a gamut of colors as a standard color space. This calibration improves the display accuracy and consistency for critical viewing and color evaluation of images.