US20010012324A1 - Method and apparatus for advanced encoder system - Google Patents
Method and apparatus for advanced encoder system Download PDFInfo
- Publication number
- US20010012324A1 US20010012324A1 US09/036,988 US3698898A US2001012324A1 US 20010012324 A1 US20010012324 A1 US 20010012324A1 US 3698898 A US3698898 A US 3698898A US 2001012324 A1 US2001012324 A1 US 2001012324A1
- Authority
- US
- United States
- Prior art keywords
- resources
- frame
- data blocks
- quantizing
- coupled
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 15
- 230000033001 locomotion Effects 0.000 claims abstract description 109
- 239000013598 vector Substances 0.000 claims description 72
- 230000008859 change Effects 0.000 claims description 16
- 238000010586 diagram Methods 0.000 description 9
- 238000005259 measurement Methods 0.000 description 9
- 238000013139 quantization Methods 0.000 description 7
- 238000001514 detection method Methods 0.000 description 6
- 238000004891 communication Methods 0.000 description 5
- 230000005540 biological transmission Effects 0.000 description 3
- 230000000875 corresponding effect Effects 0.000 description 3
- 238000012546 transfer Methods 0.000 description 3
- 230000008901 benefit Effects 0.000 description 2
- 239000000284 extract Substances 0.000 description 2
- 238000009499 grossing Methods 0.000 description 2
- 230000009467 reduction Effects 0.000 description 2
- 230000009466 transformation Effects 0.000 description 2
- 238000000844 transformation Methods 0.000 description 2
- 241000084490 Esenbeckia delta Species 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 230000003139 buffering effect Effects 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 230000006835 compression Effects 0.000 description 1
- 230000002596 correlated effect Effects 0.000 description 1
- 238000013144 data compression Methods 0.000 description 1
- 230000007423 decrease Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000004091 panning Methods 0.000 description 1
- 238000007670 refining Methods 0.000 description 1
- 238000012552 review Methods 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/189—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding
- H04N19/196—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding being specially adapted for the computation of encoding parameters, e.g. by averaging previously computed encoding parameters
- H04N19/198—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding being specially adapted for the computation of encoding parameters, e.g. by averaging previously computed encoding parameters including smoothing of a sequence of encoding parameters, e.g. by averaging, by choice of the maximum, minimum or median value
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/103—Selection of coding mode or of prediction mode
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/103—Selection of coding mode or of prediction mode
- H04N19/107—Selection of coding mode or of prediction mode between spatial and temporal predictive coding, e.g. picture refresh
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/103—Selection of coding mode or of prediction mode
- H04N19/114—Adapting the group of pictures [GOP] structure, e.g. number of B-frames between two anchor frames
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/124—Quantisation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/136—Incoming video signal characteristics or properties
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/136—Incoming video signal characteristics or properties
- H04N19/14—Coding unit complexity, e.g. amount of activity or edge presence estimation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/146—Data rate or code amount at the encoder output
- H04N19/152—Data rate or code amount at the encoder output by measuring the fullness of the transmission buffer
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/164—Feedback from the receiver or from the transmission channel
- H04N19/166—Feedback from the receiver or from the transmission channel concerning the amount of transmission errors, e.g. bit error rate [BER]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/172—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/189—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding
- H04N19/196—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding being specially adapted for the computation of encoding parameters, e.g. by averaging previously computed encoding parameters
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/20—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video object coding
- H04N19/23—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video object coding with coding of regions that are present throughout a whole video segment, e.g. sprites, background or mosaic
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/20—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video object coding
- H04N19/27—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video object coding involving both synthetic and natural picture components, e.g. synthetic natural hybrid coding [SNHC]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/507—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction using conditional replenishment
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/513—Processing of motion vectors
- H04N19/517—Processing of motion vectors by encoding
- H04N19/52—Processing of motion vectors by encoding by predictive encoding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/60—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
- H04N19/61—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/85—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
- H04N19/87—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving scene cut or scene change detection in combination with video compression
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/146—Data rate or code amount at the encoder output
- H04N19/149—Data rate or code amount at the encoder output by estimating the code amount by means of a model, e.g. mathematical model or statistical model
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/146—Data rate or code amount at the encoder output
- H04N19/15—Data rate or code amount at the encoder output by monitoring actual compressed data size at the memory before deciding storage at the transmission buffer
Definitions
- the present invention relates to encoding of video images and more particularly to implementing features to an encoder which improves the quality of the encoded video images.
- the modern encoders for encoding video images possess the intelligence which takes advantage of the many correlations between the pictures of a video sequence. Decoders on the other hand follow directions already encoded in the bitstream by the encoders and thus are relative simple compared to the encoder. During encoding, the encoders identify areas in motion, determine optimal motion vectors, control bitrate, control data buffering such that underflow and overflow do not occur, determine where to change quantization, determine when a given block can simply be repeated, determine when to code by intra and inter techniques, and vary all of these parameters and decisions dynamically so as to maximize quality for a given situation. However, even the modern encoders still do not provide the intelligence necessary to produce smooth video at low communication bitrates.
- the present invention provides an apparatus for advanced encoders and methods for operating the same which result in improved image quality of encoded images.
- the novel video encoder is based on identifying particular properties of input images and refining the coding techniques of the encoding engine to reflect the identified properties of the input images.
- the video encoder for encoding input images having a plurality of data blocks to provide compressed image data comprises DCT (discrete cosine transformer) resources configured to DCT the data blocks.
- Quantizing resources is coupled to the DCT resources configured to quantize the data blocks to provide quantized data blocks.
- Inverse quantizing resources is coupled to the quantizing resources to inverse quantize the quantized data blocks.
- Frame reconstruction resources is coupled to the inverse quantizing resources configured to reconstruct previous compressed frames.
- Motion estimation resources is coupled to the frame reconstruction resources configured to provide predicted data blocks.
- Subtraction resources is coupled to the DCT resources and the motion estimation resources to subtract the data blocks and the predicted data blocks.
- An output data buffer is coupled to the quantizing resources configured to provide a data rate signal to the quantizing resources for modifying quantizer values of the quantizing resources in order to maintain a particular target output data rate of the compressed image data.
- the video encoder further comprises image preclassifying resources coupled between the subtraction resources and the DCT resources configured to preclassify the data blocks as active and inactive regions wherein the quantizing resources responsive to preclassification of the data blocks limits quantizer values for the active regions.
- the combination of the preclassifying resources and the quantizing resources produces variable rate coding which affords constant image quality at variable data rates. Because the active regions are coded with a relatively fixed quantization and the inactive regions are coded with larger quantization values, the actives regions produce better image quality than the inactive regions, particularly in situations where data rates are reduced.
- the frame reconstruction resources includes an automatic scene change detector configured to determine whether to code inverse quantized data blocks as intra frames or predicted frames.
- the automatic scene change detector includes a scene similarity detector configured to compare an current frame with a previous frame to determine similarity between the previous frame and the current frame.
- a frame comparator is configured to provide a combination of distortion, differences in luminance, and color histogram information for comparing the previous frame with the current frame.
- the scene similarity detector directs the current frame to be encoded as an intra frame when the combination of distortion, differences in luminance, and color histogram information exceed an adaptively determined threshold.
- the frame reconstruction resources includes a reference picture controller configured to determine whether to code inverse quantized data blocks based upon a reference frame or a previous frame.
- the reference picture controller includes a frame comparator configured to compare a previous frame and a reference frame with a current frame to determine whether the previous frame or the reference frame is more similar to the current frame, and a frame encoder coupled to the frame comparator configured to encode the current frame based on a selected more similar frame from the frame comparator.
- the reference picture controller includes a reference picture store coupled to the frame encoder configured to receive updates of additional background information for the reference frame from the reference frame comparator. For instance, whenever the automatic scene change detector directs the current frame to be encoded as an intra frame, the reference picture controller updates the reference picture store to include the intra frame.
- the reference picture controller includes a synthetic background generator which generates a synthetic background as the reference frame for encoding the current frame.
- the synthetic background generator includes animation by a java applet.
- the frame reconstruct resource codes foreground regions of images and use the synthetic background as the reference image.
- the motion estimation resources constrains motion vectors to be smooth relative to each other.
- the motion estimation resources includes a motion vector search engine configured to receive a previous frame and a current frame to search an optimal motion vector, a motion vector biasor coupled to the motion vector search engine configured to bias the optimal motion vector to favor a direction consistent with that found in surrounding areas of the optimal motion vector and provide a modified distortion, and a signal to noise ratio (SNR) comparator configured to compare SNR of additional motion vector searches performed by the motion vector search engine in a direction consistent with the optimal motion vector with the modified distortion to select the motion vector associated with minimum distortion.
- SNR signal to noise ratio
- variable rate coding maintains a constant image quality at a variable data rate
- automatic scene change detection determines when input frame are coded as intra frames based on a combination of distortion, differences in luminance, and color histogram measurements from frame encoding to determine similarity between temporally adjacent frames.
- the reference picture determination choose the reference picture or the previous picture as the bases for encoding an input image.
- the motion vector smoothing preferentially biases the motion vectors so that the overall motion vector field is more smooth than it would otherwise improving the quality of motion estimation from one frame to another.
- FIG. 1 illustrates a system level block diagram of a video conferencing system.
- FIG. 2 illustrates a block diagram of a previous art video encoder.
- FIG. 3 illustrates a block diagram of a video encoder in accordance to the present invention.
- FIG. 4 illustrates a block diagram of the modified frame reconstruct block of the video encoder.
- FIG. 5 illustrates a block diagram of the modified motion estimation block of the video encoder.
- FIG. 1 generally shows a video conferencing system 10 .
- the video conferencing system 10 includes an Internet 100 , client 110 , client 120 , client 130 , and client 140 .
- the Internet 100 provides TCP/IP (Transmission Control Protocol over Internet Protocol).
- TCP/IP Transmission Control Protocol over Internet Protocol
- Other represented segments operating the TCP/IP including intranets, local area, and telephone networks are also suitable.
- Network connection 112 provides the client 110 access to the Internet 100 .
- Network connection 122 provides the client 120 access to the Internet 100 .
- Network connection 132 provides the client 130 access to the Internet 100 .
- Network connection 142 provides the client 140 access to the Internet 100 .
- Clients 110 , 120 , 130 , and 140 include logic circuits that are programmed to perform a series of specifically identified operations for video conferencing on the Internet 100 .
- Video camera 114 provides audio/video data from client 110 for transfer to another client on the Internet 100 .
- Client 140 for example, is configured to receive the audio/video data from client 110 and transfers the audio/video data from camera 144 to client 110 on the Internet 100 .
- client 120 includes camera 124
- client 130 includes camera 134 for video conferencing on the Internet 100 .
- clients 110 , 120 , 130 , and 140 include video conferencing links via the Internet 100 to provide video conferencing between the clients.
- the clients 110 , 120 , 130 , and 140 include logic circuits that are programmed to perform a series of specifically identified operations for encoding and decoding compressed bitstream video data.
- Video encoding techniques such as H.261 and H.324 standards have been developed for use in video teleconferencing and video telephony applications that provide a plurality of display frames wherein each display frame includes a plurality of display blocks.
- each picture is divided into groups of blocks (GOBs).
- a group of blocks (GOB) includes multiples of 16 lines depending on the picture format.
- Each GOB is divided into macroblocks.
- a macroblock relates to 16 pixels by 16 pixels of y, the luminance, and the spatially corresponding 8 pixels by 8 pixels of u and v, the two color components.
- a macroblock includes four luminance blocks and the two spatially corresponding color difference blocks.
- Each luminance or chrominance block relates to 8 pixels by 8 pixels of y, u or v.
- FIG. 2 shows a block diagram of a video encoder 20 as generally known in the video encoding arts.
- Red Green Blue (RGB) data on line 208 provides video input to RGB converter 210 .
- the RGB converter 210 codes the RGB data to luminance and two color difference components y, u, and v, respectively.
- the RGB converter 210 provides an output on line 212 to a subtractor 215 .
- the subtractor 215 subtracts the output from motion estimation block 270 and the yuv data on line 212 .
- the discrete cosine transform (DCT) block 220 provides the input to the quantizer 230 .
- the output of the quantizer 230 on line 233 provides inputs to the inverse quantizer 260 and a lossless coding stage 235 .
- DCT discrete cosine transform
- the output of the inverse quantizer 260 provides the input to a frame reconstruct block 265 .
- the motion estimation block 270 receives the output of the frame reconstruct block 265 .
- the DCT block 220 , quantizer 230 , inverse quantizer 260 , the frame reconstruct block 265 and the motion estimation block 270 provide a transform based, motion compensated, predictive engine for encoding that is used in the Motion Picture Experts Group (MPEG) encoding.
- MPEG Motion Picture Experts Group
- the MPEG encoding provides a plurality of display frames wherein each display frame includes a plurality of display blocks.
- Other video encoding techniques such as H.261, H.263, H.324, MPEG-2, and MPEG-4 standards are also encoded in a similar manner.
- FIG. 3 shows a block diagram of a video encoder 30 according to the present invention.
- the video encoder 30 includes a RGB converter 310 , subtractor 315 , an image preclassifier 318 , discrete cosine transform (DCT) block 320 , quantizer 330 , lossless coding stage 335 , data rate controller 340 , inverse quantizer 360 , modified frame reconstruct block 365 , and modified motion estimation block 370 .
- DCT discrete cosine transform
- the RGB converter 310 codes the RGB data to luminance and two color difference components y, u, and v, respectively.
- the RGB converter 310 directs converted input frames on line 312 to the subtractor 315 which subtracts the output from the modified motion estimation block 370 with a current input frame.
- the output of the subtractor 315 is directed to the image preclassifier 318 .
- the image preclassifier 318 includes circuitry which identifies areas of a received image into active and inactive regions. The identified active regions are flagged to later receive a relatively fixed quantization by the quantizer 330 .
- the quantizer 330 receives the DCT coefficients from the DCT block 320 and quantizes the DCT coefficients.
- the quantizer 330 determines a quantizer value for quantizing the DCT coefficients. The bigger the quantization value, the lower precision is the quantized DCT coefficient. Lower precision coefficients require fewer bits to represent.
- the quantizer 330 selectively limits the quantizer value within a relative fixed range for the active regions. No such constraints are imposed for the quantizer value for the inactive regions. Thus, by limiting the quantization value for the active regions of an image, the active regions are encoded with more precision than the inactive regions.
- Quantized data on line 333 are applied to the lossless coding stage 335 and the inverse quantizer 360 .
- the inverse quantizer 360 inverse quantizes the quantized data for frame reconstruction by the modified frame reconstruct block 365 .
- the lossless coding stage 335 codes the quantized data for transmission on line 338 .
- the data rate controller 338 monitors the rate of compressed data from the lossless coding stage 335 and provides the compressed data on line 350 .
- the quantizer 330 receives a data rate signal on line 345 from the data rate controller 340 .
- the data rate signal on line 345 increases quantizer values of the quantizer 330 if the data rate controller determines that data rate limits on line 350 exceed predetermined set limits.
- the data rate signal on line 345 decreases quantizer values of the quantizer 330 if the data rate controller 340 determines that data rate limits fall below predetermined set limits.
- the data rate controller 340 provides feed back to control the quantizer 330 to enable encoding having constant image quality at a variable data rate.
- the image preclassifier 318 assures that active regions of an image are more precisely encoded by limiting the quantizer 330 discretion in quantizing the active regions while enabling more discretion for the quantizer 330 in quantizing the inactive regions of an image.
- the modified frame reconstruct block 365 includes circuitry for automatic scene change detection and a reference picture controller.
- the automatic scene change detection determines whether input frames should be coded as intra frames instead of the normally predicted frames. Distortion measurements from frame encoding, differences in luminance, and color histograms determine similarity between temporally adjacent frames. When distortion measurements exceed an adaptively determined threshold, an intra frame is encoded instead of a predicted frame that is based on previously encoded frames.
- the adaptively determined threshold is derived by calculating an average of previous distortions, using either a sliding or autoregressive window. Distortion measurements for the current frame is compared to this average. If the current distortion measurements exceed the average significantly then the current frame is categorized as being an intra frame.
- the automatic scene detection and reference picture controller retains good image quality and improves error resilience of the bitstream because intra frames aid a downstream decoder (not shown) in recynchronizing when bitstream errors occur.
- the reference picture controller of the modified frame reconstruct block 365 determines whether the reference picture or an actual current picture is to be used during frame reconstruct.
- the reference picture controller includes a previous frame store and a reference picture store.
- the previous frame store contains the previously encoded frame.
- the reference picture store contains a static background of a scene as well as the background concealed by the moving objects in the scene. This reference picture is built up over time by updating additional background information to the scene as new background is revealed by motion in the scene. For example, when a particular piece of previously unrevealed background is uncovered then that piece of background is written to the reference picture store, until that time the reference picture in that area contains the average value of the known data in the surrounding areas. Noise reduction is also performed on the reference picture.
- the modified frame reconstruct block 365 encodes blocks of a current frame relative to the previously encoded frame and the reference frame and chooses the encoded blocks with the fewer encoded bits. For improved noise reduction, the modified frame reconstruct block 365 is biased to favor blocks encoded relative to the reference image. However, coding blocks of a current frame relative to both the previous input frame and the reference frame and choosing the encoded blocks with the least number of bits use excessive encoding resources.
- the reference picture controller selects the frame that is most similar to the current frame. Updates to the reference frame occur when the average background differs significantly from the corresponding area in the reference frame.
- a built in by-pass for the reference frame controller conserves bandwidth when data bitrates exceed a predetermined bitrate.
- the by-pass circuitry generates a synthetic reference background as the reference frame and encodes the foreground regions in efforts to reduce frame reconstruct data bitrates.
- the reference frame is synthetically generated and consists of animations controlled and generated by a java applet.
- a selected portion of the real video is coded, this selected portion is then superimposed or blended with the reference frame for use in encoding the current frame.
- the frame reconstruct data on line 366 from the modified frame reconstruct block 365 provides the input to the modified motion estimation block 370 .
- the modified motion estimation block 370 includes circuitry for motion vector biasing which smoothes a motion vector field estimation.
- the modified motion estimation block 370 exploits the premise that when motion over an image is relatively smooth, the susceptibility of block matching performed by motion estimation to noise is deductible.
- the modified motion estimation block 370 initializes a motion vector search of the surrounding areas with an initial condition equal to the average motion vector of surrounding areas to find an optimal vector.
- a motion vector biasor biases the optimal vector from the motion vector search to favor the direction consistent with that found in the surrounding areas.
- a modified distortion from the biased optimal vector is derived.
- the selected motion vector associated with the least distortion is directed to the subtractor 215 .
- Motion vectors selected using the modified motion estimation to find the optimal vector with the lowest distortion produces an overall smoother motion vector field.
- the overall smoother motion vector field produces a more efficient coding leading to reduced bit rate and higher quality representation of scenes which contain smooth motion particularly in instances such as camera pans.
- FIG. 4 illustrates a block diagram of the modified frame reconstruct block 370 .
- the modified frame reconstruct block 365 includes circuitry for an automatic scene change detector 367 , a reference picture controller 369 , a current frame store 431 , and previous frame store 433 .
- the automatic scene change detector 367 includes a frame comparator 405 , a scene similarity detector 410 and a frame encoder 420 .
- the frame comparator 405 receives a current image from the current frame store 431 and a previous frame from the previous frame store 433 and makes a number of measurements for similarity by determining the differences in color histograms, luminance, and signal to noise ratio of a reconstructed image.
- the scene similarity detector 410 receives the differences in color histograms, luminance, and signal to noise ratio of a reconstructed image from the frame comparator 405 . If the scene similarity detector 410 determines the measurements for the current image exceed an average value for the measured quantities, the scene similarity detector 410 directs the frame encoder 420 to encode the current image frame as an intra frame. Alternatively, if the scene similarity detector 410 determines the measurements for the current image are within an average value for the measured quantities, the scene similarity detector 410 directs the frame encoder 420 to encode the current image frame as a predicted frame based upon the previous frame.
- the average value for the measured quantities provides an adaptively determined threshold which is derived by calculating the average of the previous measurements for similarity, using either a sliding or autoregressive window.
- the reference picture controller 369 determines whether to use the reference picture or the previous picture as the picture upon which to reference the encoding of the current picture.
- the reference picture controller 369 includes a reference frame store 430 , reference frame comparator 440 , frame encoder 450 , and a synthetic background generator 460 which includes generation of animation applets.
- the reference frame store 430 provides a reference frame which is generated as a combination of previous input frames.
- the previous frame store 433 provides the previously reconstructed frame which had been generated.
- the reference frame comparator 440 receives a current frame from the current frame store 431 , the previously reconstructed frame from the previous frame store 433 , and a reference frame from the reference store 430 , and selects either the previous frame or the reference frame as the basis for encoding the current frame.
- the selected frame is the most similar frame to the current frame from the current frame store 431 .
- the reference frame store 430 receives the selected frame from the reference frame comparator 440 for storage as the reference frame.
- the frame encoder 450 receives the current frame and the selected frame and encodes the current frame based on the selected frame.
- the reference frame store 430 includes a plurality of reference stores which are selectable as the reference frame. Updates to the reference frame store 430 occur when any of a number of conditions occur. Firstly, if the automatic scene detector 367 determines that a frame should be coded as an intra frame, the reference frame is replaced by the coded intra frame. Secondly, if a decoder (not shown) signals the use of a different reference frame for decoding, the reference frame controller 369 switches the reference frame to that used by the decoder. This situation occurs when the decoder detects error in the bitstream and signals the encoder that errors have been detected.
- the reference frame store 430 updates the reference frame to one that is used for encoding the current frame.
- the synthetic background generator 460 provides synthetically generated backgrounds which may be provided by an attached computer system (not shown) to the reference frame store 433 .
- the synthetic background generated 460 includes generation of animation by programs, such as java running on the attached computer system.
- a built in by-pass for the reference picture controller 369 conserves bandwidth when data bitrates exceeds a predetermined bitrate.
- the by-pass circuitry directs the reference frame store 430 to provide synthetic backgrounds as the reference frames in which the frame encoder 450 encodes the foreground regions in efforts to reduce data bitrates.
- FIG. 5 illustrates a block diagram of the modified motion estimation block 370 .
- the modified motion estimation block 370 includes a motion vector search engine 510 , motion vector biasor 520 , and a SNR comparator 530 .
- the modified motion estimation block 370 receives a previous reconstructed frame and a current input frame. For each macroblock in the current frame, the modified motion estimation block 370 finds the best matching block in the previous reconstructed frame. The x and y offsets of this best matching block constitute the motion vectors.
- the motion vector search engine 510 finds an optimal motion vector
- the motion vector biasor 520 biases the optimal motion vector to favor the direction consistent with that found in the surrounding areas of the optimal motion vector and derives a modified distortion from the biased optimal vector.
- the motion vector search engine 510 initially finds an optimal vector (dx 1 dx 2 ) with a distortion dist init .
- the motion vector search engine 510 performs additional searches in the direction consistent with the initial optimal vector (dx 1 dx 2 ).
- the SNR comparator 530 compares the SNR of the additional motion vector searches with the initial modified distortion dist mod and selects the motion vector associated with the minimum distortion. The selected motion vector associated with the least distortion is directed to the subtractor 215 .
- Motion vectors selected using the modified motion estimation to find the optimal vector with the lowest distortion produces an overall smoother motion vector field. Fewer bits are generated to code the motion vectors in this case because the actual coding of the motion vectors is done relative to surrounding motion vectors. In other words the encoded bitstream contains the difference between motion vectors not the actual motion vectors i.e. delta coding. Since the modified motion estimation block 370 forces the motion vectors to point in roughly the same directions, the deltas between the motion vectors are smaller than would otherwise be the case.
- the overall smoother motion vector field extracts movement information from camera pans and zooms and provides movement parameters rather than encoding the movement information thereby reducing the number of encoded bits to represent the situations of camera panning or zooming.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Computing Systems (AREA)
- Theoretical Computer Science (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
Description
- 1. Field of the Invention
- The present invention relates to encoding of video images and more particularly to implementing features to an encoder which improves the quality of the encoded video images.
- 2. Description of the Related Arts
- As the Internet becomes more and more popular, more and more kinds data are being transferred using the Internet. The internet and other channels of communication have bandwidth limitations. Data compression is often used to maximize data transmission over such limited bandwidth channels. Most people access the Internet using fixed rate channels such as telephone lines. The fixed rate channels present problems for viewing video. Typically, the transfer of video images require high bandwidth channels. However, compression techniques have reduced the need for the high bandwidth channels but at the expense of choppy low quality video images.
- Thus, particularly in low bitrate communication, image quality and encoder performance are still in need of improvement to achieve the quality of broadcast or real-time video at the approximate 30 frames per second. Typically, in any video film clip there are many instances when sequential picture frames are very similar from one frame to the next. Digitizing each frame and comparing the two dimensional digital arrays result in samples which are highly correlated. In particular, adjacent samples within a picture are very likely to have similar intensities. Exploiting this correlation and others within each picture and from picture to picture enables encoders to compress picture sequences more effectively.
- The modern encoders for encoding video images possess the intelligence which takes advantage of the many correlations between the pictures of a video sequence. Decoders on the other hand follow directions already encoded in the bitstream by the encoders and thus are relative simple compared to the encoder. During encoding, the encoders identify areas in motion, determine optimal motion vectors, control bitrate, control data buffering such that underflow and overflow do not occur, determine where to change quantization, determine when a given block can simply be repeated, determine when to code by intra and inter techniques, and vary all of these parameters and decisions dynamically so as to maximize quality for a given situation. However, even the modern encoders still do not provide the intelligence necessary to produce smooth video at low communication bitrates.
- Therefore, it is desirable to provide an encoding apparatus and methods of operating the same which more intelligently manipulates correlations between individual pictures.
- The present invention provides an apparatus for advanced encoders and methods for operating the same which result in improved image quality of encoded images. The novel video encoder is based on identifying particular properties of input images and refining the coding techniques of the encoding engine to reflect the identified properties of the input images. Thus, according to one aspect of the invention, the video encoder for encoding input images having a plurality of data blocks to provide compressed image data comprises DCT (discrete cosine transformer) resources configured to DCT the data blocks. Quantizing resources is coupled to the DCT resources configured to quantize the data blocks to provide quantized data blocks. Inverse quantizing resources is coupled to the quantizing resources to inverse quantize the quantized data blocks. Frame reconstruction resources is coupled to the inverse quantizing resources configured to reconstruct previous compressed frames. Motion estimation resources is coupled to the frame reconstruction resources configured to provide predicted data blocks. Subtraction resources is coupled to the DCT resources and the motion estimation resources to subtract the data blocks and the predicted data blocks. An output data buffer is coupled to the quantizing resources configured to provide a data rate signal to the quantizing resources for modifying quantizer values of the quantizing resources in order to maintain a particular target output data rate of the compressed image data.
- According to another aspect of the invention, the video encoder further comprises image preclassifying resources coupled between the subtraction resources and the DCT resources configured to preclassify the data blocks as active and inactive regions wherein the quantizing resources responsive to preclassification of the data blocks limits quantizer values for the active regions. The combination of the preclassifying resources and the quantizing resources produces variable rate coding which affords constant image quality at variable data rates. Because the active regions are coded with a relatively fixed quantization and the inactive regions are coded with larger quantization values, the actives regions produce better image quality than the inactive regions, particularly in situations where data rates are reduced.
- According to another aspect of the invention, the frame reconstruction resources includes an automatic scene change detector configured to determine whether to code inverse quantized data blocks as intra frames or predicted frames. The automatic scene change detector includes a scene similarity detector configured to compare an current frame with a previous frame to determine similarity between the previous frame and the current frame. A frame comparator is configured to provide a combination of distortion, differences in luminance, and color histogram information for comparing the previous frame with the current frame. The scene similarity detector directs the current frame to be encoded as an intra frame when the combination of distortion, differences in luminance, and color histogram information exceed an adaptively determined threshold.
- According to yet another aspect of the invention, the frame reconstruction resources includes a reference picture controller configured to determine whether to code inverse quantized data blocks based upon a reference frame or a previous frame. The reference picture controller includes a frame comparator configured to compare a previous frame and a reference frame with a current frame to determine whether the previous frame or the reference frame is more similar to the current frame, and a frame encoder coupled to the frame comparator configured to encode the current frame based on a selected more similar frame from the frame comparator.
- According to another aspect of the invention, the reference picture controller includes a reference picture store coupled to the frame encoder configured to receive updates of additional background information for the reference frame from the reference frame comparator. For instance, whenever the automatic scene change detector directs the current frame to be encoded as an intra frame, the reference picture controller updates the reference picture store to include the intra frame.
- According to yet another aspect of the invention, the reference picture controller includes a synthetic background generator which generates a synthetic background as the reference frame for encoding the current frame. The synthetic background generator includes animation by a java applet. Moreover, to conserve bandwidth, the frame reconstruct resource codes foreground regions of images and use the synthetic background as the reference image.
- According to yet another aspect of the invention, the motion estimation resources constrains motion vectors to be smooth relative to each other. The motion estimation resources includes a motion vector search engine configured to receive a previous frame and a current frame to search an optimal motion vector, a motion vector biasor coupled to the motion vector search engine configured to bias the optimal motion vector to favor a direction consistent with that found in surrounding areas of the optimal motion vector and provide a modified distortion, and a signal to noise ratio (SNR) comparator configured to compare SNR of additional motion vector searches performed by the motion vector search engine in a direction consistent with the optimal motion vector with the modified distortion to select the motion vector associated with minimum distortion. By constraining the motion vectors to be smooth relative to each other, the motion estimation resources extracts zoom information from a zooming image instead of having to encode the entire zooming image. The overall amount of data generated by the encoder is reduced.
- An apparatus and method of operating an advanced encoder are provided whereby the encoder engine provides variable rate coding, automatic scene change detection, reference picture determination and update, and motion vector smoothing. The variable rate coding maintains a constant image quality at a variable data rate, while the automatic scene change detection determines when input frame are coded as intra frames based on a combination of distortion, differences in luminance, and color histogram measurements from frame encoding to determine similarity between temporally adjacent frames. The reference picture determination choose the reference picture or the previous picture as the bases for encoding an input image. The motion vector smoothing preferentially biases the motion vectors so that the overall motion vector field is more smooth than it would otherwise improving the quality of motion estimation from one frame to another.
- Other aspects and advantages of the present invention can be seen upon review of the figures, the detailed description, and the claims which follow.
- FIG. 1 illustrates a system level block diagram of a video conferencing system.
- FIG. 2 illustrates a block diagram of a previous art video encoder.
- FIG. 3 illustrates a block diagram of a video encoder in accordance to the present invention.
- FIG. 4 illustrates a block diagram of the modified frame reconstruct block of the video encoder.
- FIG. 5 illustrates a block diagram of the modified motion estimation block of the video encoder.
- The invention will be described with respect to the Figures in which FIG. 1 generally shows a
video conferencing system 10. Thevideo conferencing system 10 includes an Internet 100,client 110,client 120,client 130, andclient 140. TheInternet 100 provides TCP/IP (Transmission Control Protocol over Internet Protocol). Other represented segments operating the TCP/IP including intranets, local area, and telephone networks are also suitable. -
Network connection 112 provides theclient 110 access to theInternet 100.Network connection 122 provides theclient 120 access to theInternet 100.Network connection 132 provides theclient 130 access to theInternet 100.Network connection 142 provides theclient 140 access to theInternet 100.Clients Internet 100.Video camera 114 provides audio/video data fromclient 110 for transfer to another client on theInternet 100.Client 140 for example, is configured to receive the audio/video data fromclient 110 and transfers the audio/video data fromcamera 144 toclient 110 on theInternet 100. Similarly,client 120 includescamera 124, andclient 130 includescamera 134 for video conferencing on theInternet 100. Thus,clients Internet 100 to provide video conferencing between the clients. - The
clients - For example, each picture is divided into groups of blocks (GOBs). A group of blocks (GOB) includes multiples of 16 lines depending on the picture format. Each GOB is divided into macroblocks. A macroblock relates to 16 pixels by 16 pixels of y, the luminance, and the spatially corresponding 8 pixels by 8 pixels of u and v, the two color components. Further, a macroblock includes four luminance blocks and the two spatially corresponding color difference blocks. Each luminance or chrominance block relates to 8 pixels by 8 pixels of y, u or v. A more detailed description is contained in Document LBC-95-251 of the International Telecommunication Union Telecommunication Standardization Sector Study Group 15 entitled “Draft Recommendations H.263 (Video coding for low bitrate communication)”, contact Karel Rijkse, Tel: +31 70 332 8588; the Draft Recommendations H.263 (Video coding for low bitrate communication) is herein incorporated by reference in its entirety.
- FIG. 2 shows a block diagram of a
video encoder 20 as generally known in the video encoding arts. Red Green Blue (RGB) data online 208 provides video input toRGB converter 210. TheRGB converter 210 codes the RGB data to luminance and two color difference components y, u, and v, respectively. TheRGB converter 210 provides an output on line 212 to asubtractor 215. Thesubtractor 215 subtracts the output from motion estimation block 270 and the yuv data on line 212. The discrete cosine transform (DCT) block 220 provides the input to thequantizer 230. The output of thequantizer 230 online 233 provides inputs to theinverse quantizer 260 and alossless coding stage 235. The output of theinverse quantizer 260 provides the input to a frame reconstructblock 265. The motion estimation block 270 receives the output of the frame reconstructblock 265. TheDCT block 220,quantizer 230,inverse quantizer 260, the frame reconstructblock 265 and the motion estimation block 270 provide a transform based, motion compensated, predictive engine for encoding that is used in the Motion Picture Experts Group (MPEG) encoding. The MPEG encoding provides a plurality of display frames wherein each display frame includes a plurality of display blocks. Other video encoding techniques such as H.261, H.263, H.324, MPEG-2, and MPEG-4 standards are also encoded in a similar manner. - FIG. 3 shows a block diagram of a
video encoder 30 according to the present invention. Thevideo encoder 30 includes aRGB converter 310,subtractor 315, animage preclassifier 318, discrete cosine transform (DCT) block 320,quantizer 330,lossless coding stage 335,data rate controller 340,inverse quantizer 360, modified frame reconstructblock 365, and modifiedmotion estimation block 370. As RGB data online 308 is received, theRGB converter 310 codes the RGB data to luminance and two color difference components y, u, and v, respectively. TheRGB converter 310 directs converted input frames online 312 to thesubtractor 315 which subtracts the output from the modified motion estimation block 370 with a current input frame. The output of thesubtractor 315 is directed to theimage preclassifier 318. - The
image preclassifier 318 includes circuitry which identifies areas of a received image into active and inactive regions. The identified active regions are flagged to later receive a relatively fixed quantization by thequantizer 330. As theDCT block 320 receives the image with the preclassified regions, DCT coefficients are calculated. Thequantizer 330 receives the DCT coefficients from theDCT block 320 and quantizes the DCT coefficients. Thequantizer 330 determines a quantizer value for quantizing the DCT coefficients. The bigger the quantization value, the lower precision is the quantized DCT coefficient. Lower precision coefficients require fewer bits to represent. - The use of large quantization values allows the
encoder 30 to selectively discard regions of the image with less activity. Since theimage preclassifier 318 identified areas of the image into active and inactive regions, thequantizer 330 selectively limits the quantizer value within a relative fixed range for the active regions. No such constraints are imposed for the quantizer value for the inactive regions. Thus, by limiting the quantization value for the active regions of an image, the active regions are encoded with more precision than the inactive regions. - Quantized data on
line 333 are applied to thelossless coding stage 335 and theinverse quantizer 360. Theinverse quantizer 360 inverse quantizes the quantized data for frame reconstruction by the modified frame reconstructblock 365. Thelossless coding stage 335 codes the quantized data for transmission online 338. Thedata rate controller 338 monitors the rate of compressed data from thelossless coding stage 335 and provides the compressed data online 350. - The
quantizer 330 receives a data rate signal online 345 from thedata rate controller 340. The data rate signal online 345 increases quantizer values of thequantizer 330 if the data rate controller determines that data rate limits online 350 exceed predetermined set limits. Conversely, the data rate signal online 345 decreases quantizer values of thequantizer 330 if thedata rate controller 340 determines that data rate limits fall below predetermined set limits. Thus, thedata rate controller 340 provides feed back to control thequantizer 330 to enable encoding having constant image quality at a variable data rate. Moreover, theimage preclassifier 318 assures that active regions of an image are more precisely encoded by limiting thequantizer 330 discretion in quantizing the active regions while enabling more discretion for thequantizer 330 in quantizing the inactive regions of an image. - The modified frame reconstruct
block 365 includes circuitry for automatic scene change detection and a reference picture controller. The automatic scene change detection determines whether input frames should be coded as intra frames instead of the normally predicted frames. Distortion measurements from frame encoding, differences in luminance, and color histograms determine similarity between temporally adjacent frames. When distortion measurements exceed an adaptively determined threshold, an intra frame is encoded instead of a predicted frame that is based on previously encoded frames. The adaptively determined threshold is derived by calculating an average of previous distortions, using either a sliding or autoregressive window. Distortion measurements for the current frame is compared to this average. If the current distortion measurements exceed the average significantly then the current frame is categorized as being an intra frame. The automatic scene detection and reference picture controller retains good image quality and improves error resilience of the bitstream because intra frames aid a downstream decoder (not shown) in recynchronizing when bitstream errors occur. - The reference picture controller of the modified frame reconstruct
block 365 determines whether the reference picture or an actual current picture is to be used during frame reconstruct. The reference picture controller includes a previous frame store and a reference picture store. The previous frame store contains the previously encoded frame. The reference picture store contains a static background of a scene as well as the background concealed by the moving objects in the scene. This reference picture is built up over time by updating additional background information to the scene as new background is revealed by motion in the scene. For example, when a particular piece of previously unrevealed background is uncovered then that piece of background is written to the reference picture store, until that time the reference picture in that area contains the average value of the known data in the surrounding areas. Noise reduction is also performed on the reference picture. - The modified frame reconstruct
block 365 encodes blocks of a current frame relative to the previously encoded frame and the reference frame and chooses the encoded blocks with the fewer encoded bits. For improved noise reduction, the modified frame reconstructblock 365 is biased to favor blocks encoded relative to the reference image. However, coding blocks of a current frame relative to both the previous input frame and the reference frame and choosing the encoded blocks with the least number of bits use excessive encoding resources. - Alternatively, by comparing the current frame with the previous frame and the reference frame and choosing either the previous frame or the reference frame as the basis for encoding the current frame alleviates multiple encodes of the current frame while improving the encode of the current frame. Based upon the comparison between the previous frame and the reference frame, the reference picture controller selects the frame that is most similar to the current frame. Updates to the reference frame occur when the average background differs significantly from the corresponding area in the reference frame. A built in by-pass for the reference frame controller conserves bandwidth when data bitrates exceed a predetermined bitrate. The by-pass circuitry generates a synthetic reference background as the reference frame and encodes the foreground regions in efforts to reduce frame reconstruct data bitrates.
- In an alternative embodiment, the reference frame is synthetically generated and consists of animations controlled and generated by a java applet. In the case of the synthetic background or animations, a selected portion of the real video is coded, this selected portion is then superimposed or blended with the reference frame for use in encoding the current frame.
- The frame reconstruct data on
line 366 from the modified frame reconstructblock 365 provides the input to the modifiedmotion estimation block 370. The modifiedmotion estimation block 370 includes circuitry for motion vector biasing which smoothes a motion vector field estimation. The modified motion estimation block 370 exploits the premise that when motion over an image is relatively smooth, the susceptibility of block matching performed by motion estimation to noise is deductible. - The modified
motion estimation block 370 initializes a motion vector search of the surrounding areas with an initial condition equal to the average motion vector of surrounding areas to find an optimal vector. Of course, those skilled in the art will realize that other measures of smoothness other than the average can also be used. A motion vector biasor biases the optimal vector from the motion vector search to favor the direction consistent with that found in the surrounding areas. A modified distortion from the biased optimal vector is derived. The selected motion vector associated with the least distortion is directed to thesubtractor 215. Motion vectors selected using the modified motion estimation to find the optimal vector with the lowest distortion produces an overall smoother motion vector field. The overall smoother motion vector field produces a more efficient coding leading to reduced bit rate and higher quality representation of scenes which contain smooth motion particularly in instances such as camera pans. - Moreover a similar strategy is adaptable to cope with other types of smooth motion image transformations such as zooms. It will be obvious to those persons skilled in the art to apply the current teachings to the other types of smooth motion image transformations.
- FIG. 4 illustrates a block diagram of the modified frame reconstruct
block 370. The modified frame reconstructblock 365 includes circuitry for an automaticscene change detector 367, areference picture controller 369, acurrent frame store 431, and previous frame store 433. The automaticscene change detector 367 includes aframe comparator 405, ascene similarity detector 410 and aframe encoder 420. Theframe comparator 405 receives a current image from thecurrent frame store 431 and a previous frame from the previous frame store 433 and makes a number of measurements for similarity by determining the differences in color histograms, luminance, and signal to noise ratio of a reconstructed image. Thescene similarity detector 410 receives the differences in color histograms, luminance, and signal to noise ratio of a reconstructed image from theframe comparator 405. If thescene similarity detector 410 determines the measurements for the current image exceed an average value for the measured quantities, thescene similarity detector 410 directs theframe encoder 420 to encode the current image frame as an intra frame. Alternatively, if thescene similarity detector 410 determines the measurements for the current image are within an average value for the measured quantities, thescene similarity detector 410 directs theframe encoder 420 to encode the current image frame as a predicted frame based upon the previous frame. - The average value for the measured quantities provides an adaptively determined threshold which is derived by calculating the average of the previous measurements for similarity, using either a sliding or autoregressive window.
- The
reference picture controller 369 determines whether to use the reference picture or the previous picture as the picture upon which to reference the encoding of the current picture. Thereference picture controller 369 includes areference frame store 430,reference frame comparator 440,frame encoder 450, and asynthetic background generator 460 which includes generation of animation applets. Thereference frame store 430 provides a reference frame which is generated as a combination of previous input frames. The previous frame store 433 provides the previously reconstructed frame which had been generated. Thereference frame comparator 440 receives a current frame from thecurrent frame store 431, the previously reconstructed frame from the previous frame store 433, and a reference frame from thereference store 430, and selects either the previous frame or the reference frame as the basis for encoding the current frame. The selected frame is the most similar frame to the current frame from thecurrent frame store 431. Thereference frame store 430 receives the selected frame from thereference frame comparator 440 for storage as the reference frame. Theframe encoder 450 receives the current frame and the selected frame and encodes the current frame based on the selected frame. - In an alternative embodiment, the
reference frame store 430 includes a plurality of reference stores which are selectable as the reference frame. Updates to thereference frame store 430 occur when any of a number of conditions occur. Firstly, if theautomatic scene detector 367 determines that a frame should be coded as an intra frame, the reference frame is replaced by the coded intra frame. Secondly, if a decoder (not shown) signals the use of a different reference frame for decoding, thereference frame controller 369 switches the reference frame to that used by the decoder. This situation occurs when the decoder detects error in the bitstream and signals the encoder that errors have been detected. Switching the reference frame to that of the decoders enables error to be repaired without the need for the encoder to send a new intra frame. Thirdly, if the reference frame has not been used for encoding some number of frames, i.e. it no longer represents the material in the video, thereference frame store 430 updates the reference frame to one that is used for encoding the current frame. - The
synthetic background generator 460 provides synthetically generated backgrounds which may be provided by an attached computer system (not shown) to the reference frame store 433. The synthetic background generated 460 includes generation of animation by programs, such as java running on the attached computer system. Furthermore, a built in by-pass for thereference picture controller 369 conserves bandwidth when data bitrates exceeds a predetermined bitrate. The by-pass circuitry directs thereference frame store 430 to provide synthetic backgrounds as the reference frames in which theframe encoder 450 encodes the foreground regions in efforts to reduce data bitrates. - FIG. 5 illustrates a block diagram of the modified
motion estimation block 370. The modifiedmotion estimation block 370 includes a motionvector search engine 510,motion vector biasor 520, and aSNR comparator 530. The modifiedmotion estimation block 370 receives a previous reconstructed frame and a current input frame. For each macroblock in the current frame, the modifiedmotion estimation block 370 finds the best matching block in the previous reconstructed frame. The x and y offsets of this best matching block constitute the motion vectors. As the motionvector search engine 510 finds an optimal motion vector, themotion vector biasor 520 biases the optimal motion vector to favor the direction consistent with that found in the surrounding areas of the optimal motion vector and derives a modified distortion from the biased optimal vector. - For example, the motion
vector search engine 510 initially finds an optimal vector (dx1 dx2) with a distortion distinit. Themotion vector biasor 520 subtracts a bias from the distortion distinit, to provide a modified distortion distmod where distmod=distinit−bias. The motionvector search engine 510 performs additional searches in the direction consistent with the initial optimal vector (dx1 dx2). TheSNR comparator 530 compares the SNR of the additional motion vector searches with the initial modified distortion distmod and selects the motion vector associated with the minimum distortion. The selected motion vector associated with the least distortion is directed to thesubtractor 215. - Motion vectors selected using the modified motion estimation to find the optimal vector with the lowest distortion produces an overall smoother motion vector field. Fewer bits are generated to code the motion vectors in this case because the actual coding of the motion vectors is done relative to surrounding motion vectors. In other words the encoded bitstream contains the difference between motion vectors not the actual motion vectors i.e. delta coding. Since the modified motion estimation block370 forces the motion vectors to point in roughly the same directions, the deltas between the motion vectors are smaller than would otherwise be the case. Thus, in situations such as camera pans and zooms, the overall smoother motion vector field extracts movement information from camera pans and zooms and provides movement parameters rather than encoding the movement information thereby reducing the number of encoded bits to represent the situations of camera panning or zooming.
- While the foregoing detailed description has described several embodiments of the apparatus and methods for an intelligent encoder system in accordance with this invention, it is to be understood that the above description is illustrative only and not limiting of the disclosed invention. Obviously, many modifications and variations will be apparent to the practitioners skilled in this art. Accordingly, the apparatus and methods for an advanced intelligent encoder system have been provided. The advanced intelligent encoder system including variable rate coding, automatic scene change detection, reference picture controller, and refined motion estimation produces improved and smoother transitions between encoded images.
Claims (20)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US09/036,988 US6438165B2 (en) | 1998-03-09 | 1998-03-09 | Method and apparatus for advanced encoder system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US09/036,988 US6438165B2 (en) | 1998-03-09 | 1998-03-09 | Method and apparatus for advanced encoder system |
Publications (2)
Publication Number | Publication Date |
---|---|
US20010012324A1 true US20010012324A1 (en) | 2001-08-09 |
US6438165B2 US6438165B2 (en) | 2002-08-20 |
Family
ID=21891827
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US09/036,988 Expired - Fee Related US6438165B2 (en) | 1998-03-09 | 1998-03-09 | Method and apparatus for advanced encoder system |
Country Status (1)
Country | Link |
---|---|
US (1) | US6438165B2 (en) |
Cited By (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6385242B1 (en) * | 1998-05-04 | 2002-05-07 | General Instrument Corporation | Method and apparatus for inverse quantization of MPEG-4 video |
US20030030651A1 (en) * | 2001-08-10 | 2003-02-13 | Simon Gibbs | System and method for data assisted chroma-keying |
US20030030727A1 (en) * | 2001-08-10 | 2003-02-13 | Simon Gibbs | System and method for enhancing real-time data feeds |
US20030030658A1 (en) * | 2001-08-10 | 2003-02-13 | Simon Gibbs | System and method for mixed reality broadcast |
US20030030734A1 (en) * | 2001-08-10 | 2003-02-13 | Simon Gibbs | System and method for transitioning between real images and virtual images |
US20030128215A1 (en) * | 2001-11-27 | 2003-07-10 | Samsung Electronics Co., Ltd. | Method and apparatus for encoding and decoding position interpolator |
US20030232366A1 (en) * | 2000-10-12 | 2003-12-18 | Marical, L.L.C. | Polyvalent cation-sensing receptor in Atlantic Salmon |
WO2006007285A1 (en) * | 2004-06-18 | 2006-01-19 | Thomson Licensing | Method and apparatus for video encoding optimization |
US20070058724A1 (en) * | 2005-09-15 | 2007-03-15 | Walter Paul | Methods and systems for mixed spatial resolution video compression |
US20070183674A1 (en) * | 2002-10-18 | 2007-08-09 | Samsung Electronics Co., Ltd. | Method and apparatus for encoding and decoding key value data of coordinate interpolator |
US20070230565A1 (en) * | 2004-06-18 | 2007-10-04 | Tourapis Alexandros M | Method and Apparatus for Video Encoding Optimization |
WO2009067155A3 (en) * | 2007-11-16 | 2009-07-09 | Thomson Licensing | System and method for encoding video |
US20090234971A1 (en) * | 2008-03-14 | 2009-09-17 | Microsoft Corporation | Encoding/decoding while allowing varying message formats per message |
US20100166083A1 (en) * | 2004-06-16 | 2010-07-01 | Chupp Christopher E | Mark-based content modulation and detection |
US20120307884A1 (en) * | 2011-05-31 | 2012-12-06 | Broadcom Corporation | Selective intra and/or inter prediction video encoding |
CN104395696A (en) * | 2012-06-28 | 2015-03-04 | 皇家飞利浦有限公司 | A method of estimating the position of a device and an apparatus implementing the same |
US20160073109A1 (en) * | 2014-09-05 | 2016-03-10 | Canon Kabushiki Kaisha | Image processing apparatus, image processing method, and storage medium |
EP2661896A4 (en) * | 2011-01-03 | 2016-04-20 | Apple Inc | Video coding system using implied reference frames |
US20160112707A1 (en) * | 2014-10-15 | 2016-04-21 | Intel Corporation | Policy-based image encoding |
EP3013050A1 (en) * | 2014-10-22 | 2016-04-27 | Axis AB | Video compression with adaptive GOP length and adaptive quantization |
US9348829B2 (en) | 2002-03-29 | 2016-05-24 | Sony Corporation | Media management system and process |
US20170154236A1 (en) * | 2014-07-11 | 2017-06-01 | Sony Corporation | Image processing device, imaging device, image processing method, and program |
CN114666006A (en) * | 2020-12-22 | 2022-06-24 | 通用汽车环球科技运作有限责任公司 | Rate adaptive coding and decoding scheme for prioritized segmented data |
Families Citing this family (36)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4021998B2 (en) * | 1998-08-07 | 2007-12-12 | 松下電器産業株式会社 | Data transmission control system and data transmission control method |
US6414992B1 (en) * | 1999-01-27 | 2002-07-02 | Sun Microsystems, Inc. | Optimal encoding of motion compensated video |
GB2362532B (en) * | 2000-05-15 | 2004-05-05 | Nokia Mobile Phones Ltd | Video coding |
GB2362533A (en) * | 2000-05-15 | 2001-11-21 | Nokia Mobile Phones Ltd | Encoding a video signal with an indicator of the type of error concealment used |
US7751473B2 (en) * | 2000-05-15 | 2010-07-06 | Nokia Corporation | Video coding |
CA2428946C (en) * | 2000-11-14 | 2010-06-22 | Scientific-Atlanta, Inc. | Networked subscriber television distribution |
US8127326B2 (en) | 2000-11-14 | 2012-02-28 | Claussen Paul J | Proximity detection using wireless connectivity in a communications system |
AU2002245609A1 (en) * | 2001-03-05 | 2002-09-19 | Intervideo, Inc. | Systems and methods of error resilience in a video decoder |
JP3823804B2 (en) * | 2001-10-22 | 2006-09-20 | ソニー株式会社 | Signal processing method and apparatus, signal processing program, and recording medium |
JP3997749B2 (en) * | 2001-10-22 | 2007-10-24 | ソニー株式会社 | Signal processing method and apparatus, signal processing program, and recording medium |
JP3750583B2 (en) * | 2001-10-22 | 2006-03-01 | ソニー株式会社 | Signal processing method and apparatus, and signal processing program |
US20030192047A1 (en) * | 2002-03-22 | 2003-10-09 | Gaul Michael A. | Exporting data from a digital home communication terminal to a client device |
US7516470B2 (en) | 2002-08-02 | 2009-04-07 | Cisco Technology, Inc. | Locally-updated interactive program guide |
US7908625B2 (en) | 2002-10-02 | 2011-03-15 | Robertson Neil C | Networked multimedia system |
US8046806B2 (en) | 2002-10-04 | 2011-10-25 | Wall William E | Multiroom point of deployment module |
US7360235B2 (en) | 2002-10-04 | 2008-04-15 | Scientific-Atlanta, Inc. | Systems and methods for operating a peripheral record/playback device in a networked multimedia system |
US7903892B2 (en) * | 2002-10-29 | 2011-03-08 | Ati Technologies Ulc | Image analysis for image compression suitability and real-time selection |
US8094640B2 (en) | 2003-01-15 | 2012-01-10 | Robertson Neil C | Full duplex wideband communications system for a local coaxial network |
US7643679B2 (en) * | 2003-02-13 | 2010-01-05 | Ati Technologies Ulc | Method and apparatus for block based image compression with multiple non-uniform block encodings |
US8111928B2 (en) * | 2003-02-13 | 2012-02-07 | Ati Technologies Ulc | Method and apparatus for compression of multi-sampled anti-aliasing color data |
US7764833B2 (en) * | 2003-02-13 | 2010-07-27 | Ati Technologies Ulc | Method and apparatus for anti-aliasing using floating point subpixel color values and compression of same |
US7983835B2 (en) | 2004-11-03 | 2011-07-19 | Lagassey Paul J | Modular intelligent transportation system |
US7564874B2 (en) | 2004-09-17 | 2009-07-21 | Uni-Pixel Displays, Inc. | Enhanced bandwidth data encoding method |
US7606429B2 (en) * | 2005-03-25 | 2009-10-20 | Ati Technologies Ulc | Block-based image compression method and apparatus |
US7515565B2 (en) * | 2005-05-09 | 2009-04-07 | Kyocera Corporation | Multiple source wireless communication system and method |
US8316301B2 (en) * | 2005-08-04 | 2012-11-20 | Samsung Electronics Co., Ltd. | Apparatus, medium, and method segmenting video sequences based on topic |
JP4559935B2 (en) * | 2005-08-25 | 2010-10-13 | 株式会社東芝 | Image storage apparatus and method |
US7876998B2 (en) | 2005-10-05 | 2011-01-25 | Wall William E | DVD playback over multi-room by copying to HDD |
US7583845B2 (en) * | 2006-02-15 | 2009-09-01 | Panasonic Corporation | Associative vector storage system supporting fast similarity search based on self-similarity feature extractions across multiple transformed domains |
JP5196731B2 (en) * | 2006-04-20 | 2013-05-15 | キヤノン株式会社 | Image processing apparatus and image processing method |
US20070274385A1 (en) * | 2006-05-26 | 2007-11-29 | Zhongli He | Method of increasing coding efficiency and reducing power consumption by on-line scene change detection while encoding inter-frame |
KR100803611B1 (en) * | 2006-11-28 | 2008-02-15 | 삼성전자주식회사 | Image encoding and decoding method and apparatus |
JP4875007B2 (en) * | 2007-02-09 | 2012-02-15 | パナソニック株式会社 | Moving picture coding apparatus, moving picture coding method, and moving picture decoding apparatus |
US20080260023A1 (en) * | 2007-04-18 | 2008-10-23 | Chih-Ta Star Sung | Digital video encoding and decoding with refernecing frame buffer compression |
KR101611437B1 (en) * | 2009-10-28 | 2016-04-26 | 삼성전자주식회사 | Method and apparatus for encoding/decoding image by referencing to a plurality of frames |
KR101681589B1 (en) * | 2010-07-27 | 2016-12-01 | 엘지전자 주식회사 | Image processing apparatus and method thereof |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
AU657510B2 (en) * | 1991-05-24 | 1995-03-16 | Apple Inc. | Improved image encoding/decoding method and apparatus |
KR940020832A (en) * | 1993-02-25 | 1994-09-16 | 김주용 | Adaptive Quantization Method of High-definition Television and System Coder Using the Same |
JP3340532B2 (en) * | 1993-10-20 | 2002-11-05 | 株式会社日立製作所 | Video search method and apparatus |
US5774593A (en) * | 1995-07-24 | 1998-06-30 | University Of Washington | Automatic scene decomposition and optimization of MPEG compressed video |
JPH09130732A (en) * | 1995-11-01 | 1997-05-16 | Matsushita Electric Ind Co Ltd | Scene change detection method and dynamic image edit device |
US5796952A (en) * | 1997-03-21 | 1998-08-18 | Dot Com Development, Inc. | Method and apparatus for tracking client interaction with a network resource and creating client profiles and resource database |
-
1998
- 1998-03-09 US US09/036,988 patent/US6438165B2/en not_active Expired - Fee Related
Cited By (53)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6385242B1 (en) * | 1998-05-04 | 2002-05-07 | General Instrument Corporation | Method and apparatus for inverse quantization of MPEG-4 video |
US20030232366A1 (en) * | 2000-10-12 | 2003-12-18 | Marical, L.L.C. | Polyvalent cation-sensing receptor in Atlantic Salmon |
US7339609B2 (en) | 2001-08-10 | 2008-03-04 | Sony Corporation | System and method for enhancing real-time data feeds |
US20030030651A1 (en) * | 2001-08-10 | 2003-02-13 | Simon Gibbs | System and method for data assisted chroma-keying |
US20030030727A1 (en) * | 2001-08-10 | 2003-02-13 | Simon Gibbs | System and method for enhancing real-time data feeds |
US20030030658A1 (en) * | 2001-08-10 | 2003-02-13 | Simon Gibbs | System and method for mixed reality broadcast |
US20030030734A1 (en) * | 2001-08-10 | 2003-02-13 | Simon Gibbs | System and method for transitioning between real images and virtual images |
US8457350B2 (en) | 2001-08-10 | 2013-06-04 | Sony Corporation | System and method for data assisted chrom-keying |
US8022965B2 (en) * | 2001-08-10 | 2011-09-20 | Sony Corporation | System and method for data assisted chroma-keying |
US7091989B2 (en) * | 2001-08-10 | 2006-08-15 | Sony Corporation | System and method for data assisted chroma-keying |
US20060209088A1 (en) * | 2001-08-10 | 2006-09-21 | Simon Gibbs | System and method for data assisted chroma-keying |
US7173672B2 (en) | 2001-08-10 | 2007-02-06 | Sony Corporation | System and method for transitioning between real images and virtual images |
US20100322308A1 (en) * | 2001-11-27 | 2010-12-23 | Samsung Electronics Co., Ltd. | Apparatus for encoding and decoding key data and key value data of coordinate interpolator and recording medium containing bitstream into which coordinate interpolator is encoded |
US20030147470A1 (en) * | 2001-11-27 | 2003-08-07 | Samsung Electronics Co., Ltd. | Apparatus for encoding and decoding key data and key value data of coordinate interpolator and recording medium containing bitstream into which coordinate interpolator is encoded |
US8705610B2 (en) | 2001-11-27 | 2014-04-22 | Samsung Electronics Co., Ltd. | Apparatus for encoding and decoding key data and key value data of coordinate interpolator and recording medium containing bitstream into which coordinate interpolator is encoded |
US20070053600A1 (en) * | 2001-11-27 | 2007-03-08 | Samsung Electronics Co., Ltd. | Method and apparatus for encoding and decoding key value data of coordinate interpolator |
US20030128215A1 (en) * | 2001-11-27 | 2003-07-10 | Samsung Electronics Co., Ltd. | Method and apparatus for encoding and decoding position interpolator |
US7733345B2 (en) * | 2001-11-27 | 2010-06-08 | Samsung Electronics Co., Ltd. | Method and apparatus for encoding and decoding position interpolator |
US7809203B2 (en) | 2001-11-27 | 2010-10-05 | Samsung Electronics Co., Ltd. | Apparatus for encoding and decoding key data and key value data of coordinate interpolator and recording medium containing bitstream into which coordinate interpolator is encoded |
US8411975B2 (en) | 2001-11-27 | 2013-04-02 | Samsung Electronics Co., Ltd. | Method and apparatus for encoding and decoding key value data of coordinate interpolator |
US9348829B2 (en) | 2002-03-29 | 2016-05-24 | Sony Corporation | Media management system and process |
US20070183674A1 (en) * | 2002-10-18 | 2007-08-09 | Samsung Electronics Co., Ltd. | Method and apparatus for encoding and decoding key value data of coordinate interpolator |
US7809204B2 (en) | 2002-10-18 | 2010-10-05 | Samsung Electronics Co., Ltd. | Method and apparatus for encoding and decoding key value data of coordinate interpolator |
US20100166083A1 (en) * | 2004-06-16 | 2010-07-01 | Chupp Christopher E | Mark-based content modulation and detection |
US8842725B2 (en) * | 2004-06-16 | 2014-09-23 | Koplar Interactive Systems International L.L.C. | Mark-based content modulation and detection |
WO2006007285A1 (en) * | 2004-06-18 | 2006-01-19 | Thomson Licensing | Method and apparatus for video encoding optimization |
US20070230565A1 (en) * | 2004-06-18 | 2007-10-04 | Tourapis Alexandros M | Method and Apparatus for Video Encoding Optimization |
US20070058724A1 (en) * | 2005-09-15 | 2007-03-15 | Walter Paul | Methods and systems for mixed spatial resolution video compression |
US9060172B2 (en) * | 2005-09-15 | 2015-06-16 | Sri International | Methods and systems for mixed spatial resolution video compression |
US9098902B2 (en) | 2007-11-16 | 2015-08-04 | Thomson Licensing | System and method for encoding video |
US20100260271A1 (en) * | 2007-11-16 | 2010-10-14 | Thomson Licensing Llc. | Sysytem and method for encoding video |
WO2009067155A3 (en) * | 2007-11-16 | 2009-07-09 | Thomson Licensing | System and method for encoding video |
US8145794B2 (en) * | 2008-03-14 | 2012-03-27 | Microsoft Corporation | Encoding/decoding while allowing varying message formats per message |
US20090234971A1 (en) * | 2008-03-14 | 2009-09-17 | Microsoft Corporation | Encoding/decoding while allowing varying message formats per message |
US8812643B2 (en) | 2008-03-14 | 2014-08-19 | Microsoft Corporation | Conversion of hierarchical infoset type data to binary data |
EP2661896A4 (en) * | 2011-01-03 | 2016-04-20 | Apple Inc | Video coding system using implied reference frames |
US20120307884A1 (en) * | 2011-05-31 | 2012-12-06 | Broadcom Corporation | Selective intra and/or inter prediction video encoding |
US9295076B2 (en) * | 2011-05-31 | 2016-03-22 | Broadcom Corporation | Selective intra and/or inter prediction video encoding based on a channel rate |
US20150173037A1 (en) * | 2012-06-28 | 2015-06-18 | Koninklijke Philips N.V. | Method of estimating the position of a device and an apparatus implementing the same |
CN104395696A (en) * | 2012-06-28 | 2015-03-04 | 皇家飞利浦有限公司 | A method of estimating the position of a device and an apparatus implementing the same |
US9560619B2 (en) * | 2012-06-28 | 2017-01-31 | Koninklijke Philips N.V. | Method of estimating the position of a device and an apparatus implementing the same |
US10198658B2 (en) * | 2014-07-11 | 2019-02-05 | Sony Semiconductor Solutions Corporation | Image processing device, imaging device, image processing method, and program |
US20170154236A1 (en) * | 2014-07-11 | 2017-06-01 | Sony Corporation | Image processing device, imaging device, image processing method, and program |
US9860567B2 (en) * | 2014-09-05 | 2018-01-02 | Canon Kabushiki Kaisha | Image processing apparatus, image processing method, and storage medium |
US20160073109A1 (en) * | 2014-09-05 | 2016-03-10 | Canon Kabushiki Kaisha | Image processing apparatus, image processing method, and storage medium |
CN106717001A (en) * | 2014-10-15 | 2017-05-24 | 英特尔公司 | Policy-based image encoding |
WO2016060844A1 (en) * | 2014-10-15 | 2016-04-21 | Intel Corporation | Policy-based image encoding |
US20160112707A1 (en) * | 2014-10-15 | 2016-04-21 | Intel Corporation | Policy-based image encoding |
KR20160047395A (en) * | 2014-10-22 | 2016-05-02 | 엑시스 에이비 | Video encoding method and video encoder system |
EP3013050A1 (en) * | 2014-10-22 | 2016-04-27 | Axis AB | Video compression with adaptive GOP length and adaptive quantization |
US9992455B2 (en) | 2014-10-22 | 2018-06-05 | Axis Ab | Video encoding method and video encoder system |
KR101960006B1 (en) | 2014-10-22 | 2019-03-20 | 엑시스 에이비 | Video encoding method and video encoder system |
CN114666006A (en) * | 2020-12-22 | 2022-06-24 | 通用汽车环球科技运作有限责任公司 | Rate adaptive coding and decoding scheme for prioritized segmented data |
Also Published As
Publication number | Publication date |
---|---|
US6438165B2 (en) | 2002-08-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US6438165B2 (en) | Method and apparatus for advanced encoder system | |
US6430222B1 (en) | Moving picture coding apparatus | |
AU684901B2 (en) | Method and circuit for estimating motion between pictures composed of interlaced fields, and device for coding digital signals comprising such a circuit | |
US9197912B2 (en) | Content classification for multimedia processing | |
JP3133517B2 (en) | Image region detecting device, image encoding device using the image detecting device | |
US6414992B1 (en) | Optimal encoding of motion compensated video | |
US6765962B1 (en) | Adaptive selection of quantization scales for video encoding | |
US20040158719A1 (en) | Video encoder capable of differentially encoding image of speaker during visual call and method for compressing video signal using the same | |
US8780970B2 (en) | Motion wake identification and control mechanism | |
KR100643819B1 (en) | Parameterization for Fading Correction | |
US7095784B2 (en) | Method and apparatus for moving picture compression rate control using bit allocation with initial quantization step size estimation at picture level | |
JP2002525989A (en) | Intra-frame quantizer selection for video compression | |
US6028965A (en) | Method and apparatus for intelligent codec system | |
KR100211916B1 (en) | Method of determining encoding type and mode in object shape information encoding | |
JP2000350211A (en) | Method and device for encoding moving picture | |
US5812200A (en) | Method and apparatus for constant data rate output of encoders | |
US20040233984A1 (en) | Apparatus for variable bit rate control in video compression and target bit allocator thereof | |
US7133448B2 (en) | Method and apparatus for rate control in moving picture video compression | |
WO2005004496A1 (en) | Method for preventing noise when coding macroblocks | |
JP3202270B2 (en) | Video encoding device | |
EP1921866A2 (en) | Content classification for multimedia processing | |
Cho et al. | An adaptive quantization algorithm for video coding | |
JPH0646411A (en) | Picture coder | |
Choi et al. | Adaptive image quantization using total variation classification | |
GB2393060A (en) | Manipulation of video compression |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: NORMILE, JAMES OLIVER, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:NORMILE, JAMES OLIVER;REEL/FRAME:009025/0450 Effective date: 19980309 Owner name: LG ELECTRONICS, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:NORMILE, JAMES OLIVER;REEL/FRAME:009025/0450 Effective date: 19980309 |
|
REMI | Maintenance fee reminder mailed | ||
LAPS | Lapse for failure to pay maintenance fees | ||
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
FP | Expired due to failure to pay maintenance fee |
Effective date: 20060820 |