摘要:
A method and apparatus for encoding and decoding video performs transformation of at least a portion of a high-resolution video frame into a low resolution image and a plurality of enhancement data sets, encodes the low resolution image as a primary coded picture in a bitstream format and encodes each of the plurality of enhancement data sets as a different redundant coded picture in the bitstream format. For decoding, a decoded low resolution image and a plurality of decoded enhancement data sets are generated and an inverse transform is performed to construct a decoded high-resolution image. The primary coded picture and a redundant coded picture may be formatted according to the ITU-T H.264 Advanced Coding specification. The transform may be a polyphase or a sub-band transform.
摘要:
A method and apparatus for encoding and decoding video performs transformation of at least a portion of a high-resolution video frame into a low resolution image and a plurality of enhancement data sets, encodes the low resolution image as a primary coded picture in a bitstream format and encodes each of the plurality of enhancement data sets as a different redundant coded picture in the bitstream format. For decoding, a decoded low resolution image and a plurality of decoded enhancement data sets are generated and an inverse transform is performed to construct a decoded high-resolution mage. The primary coded picture and a redundant coded picture may be formatted according to the ITU-T H.264 Advanced Coding specification. The transform may be a polyphase or a sub-band transform.
摘要:
A first video signal processor (103) receives a first encoded video signal from which a video unit (201) generates a second encoded video signal, where the second encoded video signal is a reduced data rate version of the first encoded video signal. An error encoder (203) generates error redundancy data for the second encoded video signal and a multiplexer (207) generates output video data comprising the first encoded video signal and the error correcting data but not comprising the second encoded video signal. A second video processor (105) receives the output video data and a video unit (303) regenerates the second video signal from the first video signal. An error unit (305) detects errors for at least a first segment of the second video signal in response to the error redundancy data. A combiner (307) then generates combined video data by combining corresponding segments of the first encoded video signal and the second encoded video signal.
摘要:
A no-reference estimation of video quality in streaming video is provided on a macroblock basis. Compressed video is being deployed in video in streaming and transmission applications. MB-level no-reference objective quality estimation is provided based on machine learning techniques. First the feature vectors are extracted from both the MPEG coded bitstream and the reconstructed video. Various feature extraction scenarios are proposed based on bitstream information, MB prediction error, prediction source and reconstruction intensity. The features are then modeled using both a reduced model polynomial network and a Bayes classifier. The classified features may be used as feature vector used by a client device assess the quality of received video without use of the original video as a reference.
摘要:
A first video signal processor (103) receives a first encoded video signal from which a video unit (201) generates a second encoded video signal, where the second encoded video signal is a reduced data rate version of the first encoded video signal. An error encoder (203) generates error redundancy data for the second encoded video signal and a multiplexer (207) generates output video data comprising the first encoded video signal and the error correcting data but not comprising the second encoded video signal. A second video processor (105) receives the output video data and a video unit (303) regenerates the second video signal from the first video signal. An error unit (305) detects errors for at least a first segment of the second video signal in response to the error redundancy data. A combiner (307) then generates combined video data by combining corresponding segments of the first encoded video signal and the second encoded video signal.
摘要:
A method (800) for improving a quality of a scalable video object plane enhancement layer transmission over an error-prone network. The enhancement layer transmission includes at least one re-synchronisation marker followed by a Video Packet Header and header extensions. A reference VOPs' identifier (e.g. ‘ref_select_code’) is replicated from the video object plane header into a number of enhancement layer header extensions (715). An error corrupting the reference VOPs' identifier is recovered (830, 840, 850, 860) by decoding a correct reference VOP's identifier from subsequent enhancement layer header extensions. Correct reference video object planes are identified (870, 880) to be used in a reconstruction of an enhancement layer video object plane in the scalable video transmission. This improves the error performance in an enhancement layer of video transmissions over wireless channels and the Internet where the errors can be severe.