WO2006109990A1 - Procedes de codage et de decodage arithmetiques adaptatifs fondes sur le contexte et appareils a efficacite de codage amelioree et procedes de codage et decodage video et appareils mettant en oeuvre ceux-ci - Google Patents
Procedes de codage et de decodage arithmetiques adaptatifs fondes sur le contexte et appareils a efficacite de codage amelioree et procedes de codage et decodage video et appareils mettant en oeuvre ceux-ci Download PDFInfo
- Publication number
- WO2006109990A1 WO2006109990A1 PCT/KR2006/001343 KR2006001343W WO2006109990A1 WO 2006109990 A1 WO2006109990 A1 WO 2006109990A1 KR 2006001343 W KR2006001343 W KR 2006001343W WO 2006109990 A1 WO2006109990 A1 WO 2006109990A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- slice
- context model
- block
- given slice
- given
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 197
- 230000003044 adaptive effect Effects 0.000 title claims abstract description 84
- 230000002123 temporal effect Effects 0.000 claims description 25
- 239000000284 extract Substances 0.000 claims description 7
- 239000013598 vector Substances 0.000 description 8
- 238000013139 quantization Methods 0.000 description 7
- 230000005540 biological transmission Effects 0.000 description 4
- 238000010586 diagram Methods 0.000 description 4
- 230000015556 catabolic process Effects 0.000 description 2
- 230000006835 compression Effects 0.000 description 2
- 238000007906 compression Methods 0.000 description 2
- 238000006731 degradation reaction Methods 0.000 description 2
- 238000003491 array Methods 0.000 description 1
- 238000006073 displacement reaction Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/60—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
- H04N19/61—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
- H04N19/615—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding using motion compensated temporal filtering [MCTF]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/13—Adaptive entropy coding, e.g. adaptive variable length coding [AVLC] or context adaptive binary arithmetic coding [CABAC]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/136—Incoming video signal characteristics or properties
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/174—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a slice, e.g. a line of blocks or a group of blocks
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/176—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/60—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
- H04N19/61—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/60—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
- H04N19/63—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding using sub-band based transform, e.g. wavelets
Definitions
- Apparatuses and methods consistent with the present invention relate to context- based adaptive arithmetic coding and decoding with improved coding efficiency, and more particularly, to context-based adaptive arithmetic coding and decoding methods and apparatuses providing improved coding efficiency by initializing a context model for a given slice of an input video to a context model for a slice coded (decoded) temporally before the given slice for arithmetic coding and decoding.
- a video encoder performs entropy coding to convert data symbols representing video input elements into bitstreams suitably compressed for transmission or storage.
- the data symbols may include quantized transform coefficients, motion vectors, various headers, and the like.
- Examples of the entropy coding include predictive coding, variable length coding, arithmetic coding, and so on. Particularly, arithmetic coding offers the highest compression efficiency.
- context-based adaptive arithmetic coding utilizes local, spatial or temporal features.
- a Joint Video Team (JVT) scalable video model utilizes the context-based adaptive arithmetic coding in which probability models are adaptively updated using the symbols to be coded. Disclosure of Invention
- the context-based adaptive arithmetic coding method could provide adequate coding efficiency when the information has been accumulated by an increased number of blocks to be coded.
- the conventional context-based adaptive arithmetic coding method has a drawback in that, when a context model is intended to be initialized to a predefined probability model for each slice, unnecessary bits may be consumed to reach a predetermined coding efficiency after the initialization.
- the present invention provides video coding and decoding methods and apparatuses providing improved coding efficiency by using a context model for a slice having a similar statistical distribution to that of a given slice as an initial value of a context model for the given slice.
- the present invention also provides video coding and decoding methods and apparatuses providing improved coding efficiency by encoding and decoding a data symbol using different context models according to the type of a block in a given slice.
- the present invention also provides video coding and decoding methods and apparatuses providing improved coding efficiency by transmitting information about an optimum context model for a given slice to a decoder.
- a method for performing context-based adaptive arithmetic coding on a given slice in a high-pass frame in a temporal level in a temporally filtered hierarchical structure comprising resetting a context model for the given slice to a context model for a slice coded temporally before the given slice, arithmetically encoding a data symbol of the given slice using the reset context model, and updating the context model using a value of the arithmetically encoded data symbol.
- a method for performing context-based adaptive arithmetic decoding on a given slice in a high- pass frame in a temporal level in a temporally filtered hierarchical structure including resetting a context model for the given slice to a context model decoded temporally before the given slice, arithmetically decoding a bitstream corresponding to the given slice using the reset context model to generate a data symbol of the given slice, and updating the context model using a value of the data symbol.
- a video coding method including a method for performing context-based adaptive arithmetic coding on a given slice in a high-pass frame in a temporal level in a temporally filtered hierarchical structure, the video coding method including subtracting a predicted image for a block in the given slice from the block and generating a residual, performing spatial transform on the residual to create a transform coefficient, quantizing the transform coefficient, resetting a context model for the given slice to a context model for a slice coded temporally before the given slice, arithmetically encoding a data symbol containing the quantized transform coefficient using the reset context model to generate a bitstream, updating the context model using a value of the arithmetically encoded data symbol, and transmitting the bitstream.
- a video decoding method including a method for performing context-based adaptive arithmetic decoding on a given slice in a high-pass frame in a temporal level in a temporally filtered hierarchical structure, the video decoding method including parsing a bitstream and extracting data about a block in the given slice to be reconstructed, resetting a context model for the given slice to a context model for a slice decoded temporally before the given slice, arithmetically decoding a bitstream corresponding to the block using the reset context model to generate a data symbol of the given slice, updating the context model using a value of the data symbol, dequantizing the data symbol to generate a transform coefficient, performing inverse spatial transform on the transform coefficient to reconstruct a residual obtained by subtracting a predicted image from the block, and adding the predicted image reconstructed by motion compensation to the reconstructed residual and reconstructing the block.
- a method of context-based adaptive arithmetic coding of a video signal including resetting a context model for a given slice to a different context model varying according to a type of a block in the given slice, arithmetically encoding a data symbol of the block using the reset context model, and updating the context model reset according to the type of the block.
- a method of context-based adaptive arithmetic decoding of a video signal including resetting a context model for a given slice including a block to a different context model varying according to a type of the block in the given slice, arithmetically decoding a bitstream corresponding to the block type using a context model corresponding to the block type to generate a data symbol of the given slice, and updating the context model according to the block type using a value of the data symbol.
- a video coding method comprising subtracting a predicted image for a block from the block and generating a residual, performing spatial transform on the residual to create a transform coefficient, quantizing the transform coefficient, resetting a context model for a given slice comprising the block to a different context model varying model varying according to a type of the block in the given slice comprising the block, arithmetically encoding a data symbol of the block using a context model reset according to the type of the block to generate a bitstream, updating the context model reset according to the type of the block, and transmitting the bitstream.
- a video decoding method including parsing a bitstream and extracting data about a block to be reconstructed, resetting a context model for a given slice containing the block to a different context model varying according to a type of the block in the given slice containing the block, arithmetically decoding a bitstream corresponding to the block using a context model corresponding to the block type to generate a data symbol of the given slice, updating the context model according to the block type using a value of the data symbol, dequantizing the data symbol to generate a transform coefficient, performing inverse spatial transform on the transform coefficient to reconstruct a residual obtained by subtracting a predicted image from the block, and adding the predicted image reconstructed by motion compensation to the reconstructed residual and reconstructing the block.
- a video coding method including subtracting a predicted image for a block from the block and generating a residual, performing spatial transform on the residual to create a transform coefficient, quantizing the transform coefficient, resetting a context model for a given slice containing the block as a predetermined initial value, performing context-based adaptive arithmetic coding on a data symbol of the given slice containing the block using the context model and generating a final probability model, performing another context-based adaptive arithmetic coding on the data symbol of the given slice containing the block using information about the final probability model as an initial value to generate a bitstream, and transmitting the bitstream.
- a video decoding method including extracting an initial value of a context model in a given slice containing a block to be reconstructed from a bitstream, resetting a context model for the given slice using the initial value, arithmetically decoding a bitstream corresponding to the block using the reset context model to generate a data symbol of the given slice, updating the context model using a value of the data symbol, dequantizing the data symbol to generate a transform coefficient, performing inverse spatial transform on the transform coefficient to reconstruct a residual obtained by subtracting a predicted image from the block, and adding the predicted image reconstructed by motion compensation to the reconstructed residual and reconstructing the block.
- a video encoder for performing context-based adaptive arithmetic coding on a given slice in a high-pass frame in a temporal level in a temporally filtered hierarchical structure, the video encoder including a unit which subtracts a predicted image for a block in the given slice from the block and generates a residual, a unit which performs spatial transform on the residual to create a transform coefficient, a unit which quantizes the transform coefficient, a unit which resets a context model for the given slice to a context model for a slice coded temporally before the given slice, a unit which arithmetically encodes a data symbol containing the quantized transform coefficient using the reset context model to generate a bitstream, a unit which updates the context model using a value of the arithmetically encoded data symbol, and a unit which transmits the bitstream.
- a video decoder for performing context-based adaptive arithmetic decoding on a given slice in a high-pass frame in a temporal level in a temporally filtered hierarchical structure, the video decoder including a unit which parses a bitstream and extracts data about a block to be reconstructed in the given slice, a unit which resets a context model for the given slice to a context model for a slice decoded temporally before the given slice, a unit which arithmetically decodes a bitstream corresponding to the block using the reset context model to generate a data symbol of the given slice, a unit which updates the context model using a value of the data symbol, a unit which dequantizes the data symbol to generate a transform coefficient, a unit which performs inverse spatial transform on the transform coefficient to reconstruct a residual obtained by subtracting a predicted image from the block, and a unit which adds the predicted image reconstructed by motion compensation to the video decoder including a unit which parses a bitstream
- a video encoder including a unit which subtracts a predicted image for a block to be reconstructed from the block and generates a residual, a unit which performs spatial transform on the residual to create a transform coefficient, a unit which quantizes the transform coefficient, a unit which resets a context model for a slice containing the block to a different context model varying according to the type of the block, a unit which arithmetically encodes a data symbol of the block using a context model reset according to the type of the block to generate a bitstream, a unit which updates the context model reset according to the type of the block, and a unit which transmits the bitstream.
- a video decoder including a unit which parses a bitstream and extracts data about a block to be reconstructed, a unit which resets a context model for a given slice containing the block to a different context model varying according to the type of the block in the given slice containing the block, a unit which arithmetically decodes the bitstream corresponding to the block using a context model reset according to the type of the block to generate a data symbol of the given slice, a unit which updates the context model reset according to the type of the block using a value of the data symbol, a unit which dequantizes the data symbol to generate a transform coefficient, a unit which performs inverse spatial transform on the transform coefficient to reconstruct a residual obtained by subtracting a predicted image from the block, and a unit which adds the predicted image reconstructed by motion compensation to the reconstructed residual and reconstructing the block.
- a video encoder including a unit which subtracts a predicted image for a block from the block and generates a residual, a unit which performs spatial transform on the residual to create a transform coefficient, a unit which quantizes the transform coefficient, a unit which resets a context model for a given slice containing the block as a predetermined initial value, a unit which performs context-based adaptive arithmetic coding on a data symbol of the given slice using the context model and generates a final probability model, a unit which performs another context-based adaptive arithmetic coding on the data symbol of the given slice using information about the final probability model as an initial value to generate a bitstream, and a unit which transmits the bitstream including information about the final probability model.
- a video decoder including a unit which extracts an initial value of a context model in a slice containing a block to be reconstructed from a bitstream, a unit which resets a context model for the given slice as the initial value, a unit which arithmetically decodes a bitstream corresponding to the block to be reconstructed using the context model to generate a data symbol of the given slice, a unit which updates the context model using a value of the data symbol, a unit which dequantizes the data symbol to generate a transform coefficient, a unit which performs inverse spatial transform on the transform coefficient to reconstruct a residual obtained by subtracting a predicted image from the block, and a unit which adds the predicted image reconstructed by motion compensation to the reconstructed residual and reconstructs the block.
- FlG. 1 illustrates a context-based adaptive arithmetic coding method according to a first exemplary embodiment of the present invention
- FlG. 2 illustrates a context-based adaptive arithmetic coding method according to a second exemplary embodiment of the present invention
- FlG. 3 illustrates a context-based adaptive arithmetic coding method according to a third exemplary embodiment of the present invention
- FlG. 4 illustrates a context-based adaptive arithmetic coding method according to a fourth exemplary embodiment of the present invention
- FlG. 5 illustrates a context-based adaptive arithmetic coding method according to a fifth exemplary embodiment of the present invention
- FlG. 6 illustrates a context-based adaptive arithmetic coding method according to a sixth exemplary embodiment of the present invention
- FlG. 7 is a flowchart illustrating a video coding method comprising a context- based adaptive coding method according to an exemplary embodiment of the present invention
- FlG. 8 is a flowchart illustrating a video decoding method comprising a context- based adaptive decoding method according to an exemplary embodiment of the present invention
- FlG. 9 is a flowchart illustrating a video coding method comprising a context- based adaptive coding method according to an exemplary embodiment of the present invention.
- FlG. 10 is a flowchart illustrating a video decoding method comprising a context- based adaptive decoding method according to an exemplary embodiment of the present invention
- FlG. 11 is a flowchart illustrating a video coding method comprising a context- based adaptive arithmetic coding method according to an exemplary embodiment of the present invention, which includes transmitting data on optimum context model to a decoder;
- FlG. 12 is a flowchart illustrating a video decoding method comprising a context- based adaptive arithmetic decoding method according to an exemplary embodiment of the present invention, which includes receiving data about optimum context model;
- FlG. 13 is a block diagram of a video encoder according to an exemplary embodiment of the present invention.
- FlG. 14 is a block diagram of a video decoder according to an exemplary embodiment of the present invention.
- Context-based Adaptive Binary Arithmetic Coding achieves high compression performance by selecting a probability model for each symbol based on a symbol context, adapting probability estimates corresponding to the probability model based on local statistics and performing arithmetic coding on the symbol.
- the coding process of the data symbol consists of at most four elementary steps: 1. Binarization; 2. Context modeling; 3. Arithmetic coding; and 4. Probability updating.
- CABAC CABAC
- Non-binary valued symbols such as transform coefficients or motion vectors
- This process is similar to converting data symbols to variable length codes except that the binary codes are previously encoded by an arithmetic encoder prior to transmission.
- CABAC set forth but the invention is not limited thereto.
- a context model which is a probability model for one or more bins of binarized symbols and chosen based on the recently coded data symbol statistics, stores a probability for each bin to be '1 ' or 1 O.'
- An arithmetic encoder codes each bin based on the chosen probability model. Each bin has only two probability sub-ranges corresponding to values of T and 1 O,' respectively.
- the chosen probability model is updated using actually coded values. That is to say, if the bin value is '1,' the frequency count of l's is incremented by one.
- CABAC CABAC since context modeling is performed in units of slices, probability values of context models are initialized using fixed tables at the start of each slice.
- VLC variable length coding
- the CABAC technique requires that a predetermined amount of information accumulates such that context models are constantly updated using the statistics of the recently coded data symbols.
- initializing context models for each slice using predefined probability models may result in unnecessary consumption of bits until degraded performance is traded off as the number of coded block increases.
- FlG. 1 illustrates a context-based adaptive arithmetic coding method according to a first exemplary embodiment of the present invention.
- a context model for a slice coded immediately before the given slice may be used as an initial value of a context model for the current high-pass frame slice.
- the high-pass frames are encoded in the order from the lowest temporal level to the highest temporal level consecutively using a context model for a slice coded immediately before the given slice as an initial value of a context model for the given slice.
- a slice coded immediately before a given slice may indicate a corresponding slice of a neighboring high-pass frame in the same temporal level or a slice coded immediately before the given slice in the same high-pass frame.
- FIGS 1 to 6 it is assumed that each frame is made of one slice.
- slices in a high-pass frame are encoded in the order from the lowest temporal level to the highest temporal level while consecutively referring to a context model for a slice coded immediately before a given slice as an initial value of a context model for the given slice.
- Arrows shown in FIGS. 1 through 6 indicate directions in which context models are referred to.
- the context model for a slice coded immediately before a given slice is used as an initial value of a context model for the given slice.
- FlG. 2 illustrates a context-based adaptive arithmetic coding method according to a second exemplary embodiment of the present invention.
- a given slice in a high-pass frame in the same temporal level consecutively uses a context model for a slice coded temporally immediately before the given slice, thereby alleviating degradation of coding efficiency due to initialization of the context model, which is the same as described above in the first exemplary embodiment.
- slices 240, 230, and 220 can respectively use as initial models context models for their closest slices 230, 220, and 210 in high-pass frames in lower temporal levels coded immediately before slices 240, 230 and 220.
- the method of the second exemplary embodiment can reduce error propagation among the slices within the same temporal level compared to the method of the first exemplary embodiment.
- FlG. 3 illustrates a context-based adaptive arithmetic coding method according to a third exemplary embodiment of the present invention.
- the context-based adaptive arithmetic coding method according to the third exemplary embodiment of the present invention can provide for high coding efficiency by using statistical information on a slice in a lower level that is temporally closest to the given slice. Further, the method of the third exemplary embodiment can reduce error propagation compared to the methods of the first and second exemplary embodiments because an error occurring within a slice can propagate to only a slice at a higher level that uses the slice as a reference.
- FlG. 4 illustrates a context-based adaptive arithmetic coding method according to a fourth exemplary embodiment of the present invention.
- the context-based adaptive arithmetic coding method shown in FlG. 4 takes advantage of merits and effects of both the methods shown in FIGS. 2 and 3. That is, arithmetic coding for an odd-numbered high-pass frame is performed using a context model for a slice in a temporally closest high-pass frame in a lower level as an initial value for the odd-numbered high-pass frame, as indicated by arrows labeled 411 through 417, while arithmetic coding for an even-numbered high-pass frame is performed using a context model for a slice in the same temporal level coded immediately before the given slice, as indicated by arrows labeled 421 through 427.
- the method according to the fourth exemplary embodiment has advantages of reducing error propagation among slices and using similar statistical characteristics of a previous slice.
- FlG. 5 illustrates a context-based adaptive arithmetic coding method according to a fifth exemplary embodiment of the present invention, in which one of context models for slices, which are coded temporally before the given slice and have similar statistical characteristics to those of the given slice, is selected for arithmetic coding of the given slice. That is, the method according to the fifth exemplary embodiment includes selecting a context model that offers the highest coding efficiency among context models used in the first through fourth exemplary embodiments and performing arithmetic coding according to the selected model. Referring to FlG.
- a context model for a slice at the same level (521) coded immediately before the slice 510, and a context model for a slice in a lower level (522) temporally closest to the given slice is selected as an initial value of a context model for the slice 510.
- information about whether a predefined initial value has been used as an initial value of a context model for the given slice, and information about a referred context model for a certain slice when statistical information of the slice coded temporally before the given slice has been used in arithmetic coding are inserted into the bitstream for being transmitted to a decoder.
- FlG. 6 illustrates a context-based adaptive arithmetic coding method according to a sixth exemplary embodiment of the present invention.
- Context-based adaptive arithmetic coding is performed on a slice 630 in a first high-pass frame using statistical information about slices 610 and 620 in a low-pass frame(641 and 642).
- the context-based adaptive arithmetic coding is also performed on a slice 620 in a low-pass frame in a neighboring group of pictures (GOP) using statistical information about the slice 610 in the previously coded low- pass frame.
- GOP group of pictures
- an encoder can insert information about a context model for a slice used as a reference for the slice 630 in the first high-pass frame in a GOP and information about whether the slice 620 in the low-pass frame in the neighboring GOP is subjected to arithmetic coding with reference to the context model for the slice 610 in the previously coded low-pass frame into a bitstream for transmission to a decoder.
- FlG. 7 is a flowchart illustrating a video coding method including a context-based adaptive coding method according to an exemplary embodiment of the present invention.
- the video coding method includes subtracting a predicted image for a block from the block in a given slice to be compressed to generate a residual signal (step S710), performing spatial transform on the residual signal to create a transform coefficient (step S720), quantizing data symbols including a transform coefficient and a motion vector obtained during generation of the predicted image (step S730), entropy coding the quantized data symbols (steps S740 through S770), and transmitting an arithmetically encoded signal to a decoder (step S780).
- the entropy coding process includes binarization (step S740), resetting of a context model (step S755), arithmetic coding (step S760), and update of a context model (step S770).
- binarization step S740 may be skipped.
- a data symbol having a non-binary value is converted or binarized into a binary value.
- a context model for the slice is reset in the step S755.
- the entropy coding is performed in units of blocks and a context model is reset in units of slices to ensure independence of slices.
- the context model is reset for symbols of the first block in the slice.
- context models corresponding thereto are adaptively updated.
- a selected context model is reset by referring to a context model for a slice coded temporally before the given slice, which is as described above.
- FIGS. 1 through 6 Examples of a slice that will be used to reset a context model for a given slice are shown in FIGS. 1 through 6.
- a video coding method including the arithmetic coding method according to the sixth embodiment of the present invention, as shown in FIG. 6, may further include selecting one of context models available for reference. Criteria of selecting one of context models available for reference include coding efficiency, an error propagation probability, and so on. In other words, a context model having highest coding efficiency, or a context model having least error propagation probability, may be selected among context model candidates.
- the binarized symbol is subjected to arithmetic coding according to a probability model having a context model for the slice coded temporally before the given slice as an initial value.
- the context model is updated using the actual value of the binarized symbol. For example, if one bin of the data symbol has a value of 1 O,' the frequency count of O's is increased. Thus, the next time this model is selected, the probability of a '0' will be slightly higher.
- FIG. 8 is a flowchart illustrating a video decoding method including a context- based adaptive decoding method according to an exemplary embodiment of the present invention.
- a decoder parses a received bitstream and extracts data for a given block to be reconstructed.
- the data may include information about a selected context model, for example, slice information of the selected context model when a context model of a slice out of slices coded temporally before the given slice is selected for initialization of a context model of the given slice during arithmetic coding performed by an encoder.
- a context model for the given slice is reset to a context model for a slice decoded temporally before the given slice in step S825. Examples of a slice that will be used to reset a context model for the given slice are as shown in FIGS. 1 through 6. If the given slice was coded according to the fifth or sixth exemplary embodiment, the context model for the given slice can be reset according to the information about a referred slice extracted from the bitstream.
- step S830 a bitstream corresponding to the block is arithmetically decoded according to the context model.
- step S840 the context model is updated using the actual value of the decoded data symbol.
- the arithmetically decoded data symbol is converted or inversely binarized into a non-binary value in step S850.
- step S860 inverse quantization is performed on the inversely binarized data symbol to generate a transform coefficient and, in step S870, inverse spatial transform is performed on the transform coefficient to reconstruct a residual signal for the given block.
- step S880 a predicted image for the given block reconstructed by motion compensation is added to the reconstructed residual signal, thereby reconstructing the given block.
- a block in a slice is classified into an inter-prediction mode block, an intra-prediction mode block or an intra base layer (BL) prediction mode block according to a block prediction mode.
- a block will have different statistical characteristics depending on which mode is used in predicting the block, which will be described below by way of example. If a consideration is taken into statistical characteristics depending on the type of a block prediction mode in CABAC, coding efficiency can be further enhanced.
- JSVM joint scalable video model
- the model 'map' indicates whether a nonzero value exists in a given position or not.
- the model 'last' confirms that the current position is the last one in the map.
- the model 'sign' indicates a sign of a value in the current position.
- the model 'level' indicates the absolute value of a value of the current position.
- CBP Coded Block Pattern
- deltaQP indicates a difference between a predetermined quantization parameter and a quantization parameter selected from the given block.
- FlG. 9 is a flowchart illustrating a video coding method including a context-based adaptive coding method according to another exemplary embodiment of the present invention.
- the video coding method includes subtracting a predicted image for a block from the block in a given slice to generate a residual (step S910), performing spatial transform on the residual to create a transform coefficient (step S720), quantizing the transform coefficient (step S730), and entropy coding the quantized transform coefficient (steps S940 through S970), and transmitting an arithmetically encoded signal to a decoder (step S980).
- step S940 data symbols including a transform coefficient and a motion vector are binarized.
- the binarization step S940 may be skipped when a CABAC method is not used.
- step S950 a context model for a given slice containing a given block is initialized to a different context model varying according to the type of the block in the given slice.
- step S960 the data symbol of the given block is arithmetically encoded using a context model chosen according to the type of the block and, in step S970, the context model corresponding to the type of the block is updated. This means that data symbols of one slice can be encoded using a set of one or more context models.
- step S980 the arithmetically encoded signal is transmitted to a decoder.
- the different context model varying according to the type of the block in the slice is initialized into a context model for a slice coded temporally before the given slice, which has been described above in detail with reference to FIGS. 1 through 6.
- the context model for an inter-prediction mode block can use a context model for an inter-prediction mode block in the slice coded temporally before the given slice as an initialization value.
- FlG. 10 is a flowchart illustrating a video decoding method including a context- based adaptive decoding method according to another exemplary embodiment of the present invention.
- step SlOlO a received bitstream is parsed and data for a given block to be reconstructed is extracted.
- steps S 1020 through S 1050 entropy decoding is performed on the given block.
- step S 1060 the entropy decoded value is dequantized to generate a transform coefficient.
- Inverse spatial transform is performed on the transform coefficient in step S 1070 and the given block is reconstructed in step S 1080.
- the entropy decoding process sub-divided into steps S 1020 through S 1050 will be described as follows.
- step S 1020 a context model for a given slice containing the given block is reset to a different context model varying according to the type of the block in the given slice.
- the different context models varying according to the type of the block in the given slice can be reset to a context model for the same block type slice decoded temporally before the given slice.
- a method of referring to a context model for a different slice prior to initialization is the same as that described above with reference to FIGS. 1 through 6.
- the decoder may parse information about the reference slice in step SlOlO and reset a context model for the given slice to the context model for the reference slice in step S 1020.
- the decoder arithmetically decodes a bitstream corresponding to the given block using a context model for the inter-prediction mode block in step S 1030 and updates the context model in step S 1040.
- the arithmetically decoded value is converted or inversely binarized into a non-binary value in step S 1050.
- FlG. 11 is a flowchart illustrating a video coding method including a context-based adaptive arithmetic coding method according to an exemplary embodiment of the present invention, in which simplified information about a final probability model is transmitted and context-based adaptive arithmetic coding is performed using the simplified information as an initial value.
- the video coding method includes subtracting a predicted image for a given block from the given block to generate a residual (step Sl 110), performing spatial transform on the residual to create a transform coefficient (step Sl 120), quantizing the transform coefficient (step Sl 130), and performing context-based adaptive arithmetic coding on a data symbol of the given block including the quantized transform coefficient (steps Sl 140 through Sl 170), and transmitting information about an initial value of a context model to a decoder (step Sl 180).
- An encoder initializes a context model for a given slice containing the given block to a predetermined initial value in step Sl 140 and arithmetically encodes data symbols of the given slice and updates the context model using a typical context-based adaptive arithmetic coding method in step Sl 150.
- the predetermined initial value may be an optimal probability model experimentally obtained by encoding a plurality of general images or a context model for a slice having similar statistical characteristics to those of the given slice.
- a final context model obtained by the update step (S 1150) may be considered as the optimum context model for the given slice, thus information about the final context model is transmitted to the decoder in step Sl 180.
- Information about the final context model may be the context model itself or a difference between either the initial value or a context model for a base layer slice corresponding to the given slice and the context model for the given slice (step Sl 160).
- step Sl 170 the data symbols of the given slice is subjected to context-based adaptive arithmetic coding using the final context model or the probability model obtained by simplifying the final context model as an initial value before being transmitted to the decoder.
- FlG. 12 is a flowchart illustrating a video decoding method including a context- based adaptive arithmetic decoding method using information about a final probability model, which corresponds to the video encoding method shown in FlG. 11.
- the video decoding method includes extracting an initial value of a context model from a bitstream (step S 1210), resetting a context model for a given slice as the initial value (step S 1220), performing context-based adaptive arithmetic decoding on a bitstream corresponding to a given block in the given slice (step S 1230), updating the context model using the arithmetically decoded value (step S 1240), dequantizing the arithmetically decoded value (step S 1250), performing inverse spatial transform on a transform coefficient to reconstruct a residual signal (step S 1260), and adding a reconstructed predicted image to the residual signal and the given block is reconstructed (step S 1270).
- the initial value of the context model is information about the final probability model obtained using the process described above with reference to FlG. 11.
- FlG. 13 is a block diagram of a video encoder 1300 according to an embodiment of the present invention.
- the video encoder 1300 includes a spatial transformer 1340, a quantizer 1350, an entropy coding unit 1360, a motion estimator 1310, and a motion compensator 1320.
- the motion estimator 1310 performs motion estimation on a given frame among input video frames using a reference frame to obtain motion vectors.
- a block matching algorithm is widely used for the motion estimation.
- a given motion block is moved in units of pixels within a particular search area in the reference frame, and displacement giving a minimum error is estimated as a motion vector.
- hierarchical variable size block matching HVSBM
- simple fixed block size motion estimation is used.
- the motion estimator 1310 transmits motion data such as motion vectors obtained as a result of motion estimation, a motion block size and a reference frame number to the entropy coding unit 1360.
- the motion compensator 1320 reduces temporal redundancy within the input video frame.
- the motion compensator 1320 performs motion compensation on t he reference frame using the motion vector calculated by the motion estimator 1310 and generates temporally estimated frame for the given frame.
- a subtracter 1330 calculates a difference between the given frame and a temporally predicted frame and temporal redundancy is removed from the input video frame.
- the spatial transformer 1340 uses spatial transform technique supporting spatial scalability to remove spatial redundancy within the frame in which temporal redundancy has been removed by the subtractor 1330.
- the spatial transform method may include a Discrete Cosine Transform (DCT), or wavelet transform. Spatially- transformed values are referred to as transform coefficients.
- the quantizer 1350 applies quantization to the transform coefficient obtained by the spatial transformer 1340. Quantization means the process of expressing the transform coefficients formed in arbitrary real values by discrete values, and matching the discrete values with indices according to the predetermined quantization table. The quantized result value is referred to as a quantized coefficient.
- the entropy coding unit 1360 losslessly encodes data symbols including the quantized transform coefficient obtained by the quantizer 1350 and the motion data received from the motion estimator 1310.
- the entropy coding unit 1360 includes a binarizer 1361, a context model selector 1362, an arithmetic encoder 1363, and a context model updater 1364.
- the binarizer 1361 converts the data symbols into a binary value that is then sent to the context model selector 1362.
- the binarizer 1361 may be omitted when CABAC is not used.
- the context model selector 1362 selects either an initial value predefined as an initial value of a context model for a given slice or a context model for a slice coded temporally before the given slice. Information about the selected initial value of the context model is sent to a bitstream generator 1370 and inserted into a bitstream for transmission. Meanwhile, when a method of referring to slices coded temporally before the given slice in order to initialize a context model for a given slice is predefined between an encoder part and a decoder part, the context model selector 1362 may not be provided.
- the arithmetic encoder 1363 performs context-based adaptive arithmetic coding on data symbols of the given block using the selected context model.
- arithmetic coding is performed on a data symbol of the slice using different context models varying according to the type of the block in the given slice.
- a final probability model is generated by performing context-based adaptive arithmetic coding and context-based adaptive arithmetic coding is performed on the data symbol of the slice using the final probability model as an initial value .
- information about the final probability model is transmitted to the bitstream generator 1370 to then be transmitted to a decoder.
- the context model updater 1364 updates the context model using a value of the arithmetically encoded data symbol.
- the context model corresponding to the block type of the given block is updated.
- the video encoder 1300 may further include a dequantizer and an inverse spatial transformer.
- FlG. 14 is a block diagram of a video decoder 1400 according to an exemplary embodiment of the present invention.
- the video decoder 1400 includes a bitstream parser 1410, an entropy decoding unit
- the bitstream parser 1410 parses a bitstream received from an encoder to extract information needed for the entropy decoding unit 1420 to decode the bitstream.
- the entropy decoding unit 1420 performs lossless decoding that is the inverse operation of entropy coding to extract motion data that are then fed to the motion compensator 1450 and texture data that are then fed to the dequantizer 1430.
- the entropy decoding unit 1420 includes a context model setter 1421, an arithmetic decoder 1422, a context model updater 1423, and an inverse binarizer 1424.
- the context model setter 1421 initializes a context model for a slice to be decoded according to the information extracted by the bitstream parser 1410.
- the information extracted by the bitstream parser 1410 may contain information about a slice having a context model to be used as an initial value of a context model for a given slice and information about a probability model to be used as the initial value of the context model for the given slice.
- context models independent of the type of a block in a slice may be initialized.
- the arithmetic decoder 1422 performs context-based adaptive arithmetic decoding on a bitstream corresponding to data symbols of the given slice according to the context model set by the context model setter 1421.
- arithmetic decoding may be performed using different context models varying according to the type of a block to be decoded.
- the context model updater 1423 updates the current context model using a value of the arithmetically decoded data symbol. Alternatively, the context model updater 1423 may update a context model selected according to the type of the decoded block.
- the inverse binarizer 1424 converts the decoded binary values obtained by the arithmetic decoder 1422 into non-binary values.
- the dequantizer 1430 dequantizes texture information received from the entropy decoding unit 1420.
- the dequantization is a process of obtaining quantized coefficients from matched quantization indices received from the encoder.
- the inverse spatial transformer 1440 performs inverse spatial transform on coefficients obtained after the dequantization to reconstruct a residual image in a spatial domain.
- the motion compensator 1450 performs motion compensation on the previously reconstructed video frame using the motion data from the entropy decoding unit 1420 and generates a motion-compensated frame. The motion compensation is re- strictively applied to a case where the given frame is coded through temporal estimation in the encoder part.
- an adder 1460 adds a motion-compensated image received from the motion compensator 1450 to the residual image and a video frame is reconstructed.
- various components mean, but are not limited to, software or hardware components, such as a Field Programmable Gate Arrays (FPGAs) or Application Specific Integrated Circuits (ASICs), which perform certain tasks.
- the components may advantageously be configured to reside on the addressable storage media and configured to execute on one or more processors.
- the functionality provided for in the components and modules may be combined into fewer components and modules or further separated into additional components and modules.
- context-based adaptive arithmetic coding and decoding methods and apparatuses of the present invention provides at least the following advantages.
- the video coding and decoding methods and apparatuses of the present invention can improve the overall coding efficiency by consecutively using a context model for a slice having similar statistical characteristics to those of a given slice.
- the video coding and decoding methods can improve overall coding efficiency by encoding a data symbol using different context models varying according to the type of the block in a slice.
- the video coding method and apparatus also provide for improved coding efficiency by transmitting information about an optimum context model for the given slice to a decoder.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
L'invention concerne des procédés de codage et de décodage arithmétiques adaptatifs fondés sur le contexte et des appareils à efficacité de codage améliorée mettant en oeuvre ceux-ci. Le procédé de codage arithmétique adaptatif fondé sur le contexte consiste à remettre à zéro un modèle de contexte pour la tranche donnée pour un modèle de contexte destiné à une tranche codée de manière temporaire avant la tranche donnée, à coder sur le plan arithmétique un symbole de données de la tranche donnée au moyen du modèle de contexte remis à zéro et à mettre à jour le modèle de contexte au moyen d'une valeur du symbole de données. Le procédé de décodage arithmétique adaptatif fondé sur le contexte consiste à remettre à zéro un modèle de contexte pour la tranche donnée pour un modèle de contexte codé temporairement avant la tranche donnée, à décoder sur le plan arithmétique un train de bits correspondant à la tranche donnée au moyen du modèle de contexte remis à zéro, aux fins de génération d'un symbole de données de la tranche donnée, et à mettre à jour le modèle de contexte au moyen d'une valeur du symbole de données.
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US67070305P | 2005-04-13 | 2005-04-13 | |
US60/670,703 | 2005-04-13 | ||
KR1020050050944A KR100664936B1 (ko) | 2005-04-13 | 2005-06-14 | 코딩 효율이 향상된 컨텍스트 기반 적응적 산술 코딩 및디코딩 방법과 이를 위한 장치, 이를 포함하는 비디오 코딩및 디코딩 방법과 이를 위한 장치 |
KR10-2005-0050944 | 2005-06-14 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2006109990A1 true WO2006109990A1 (fr) | 2006-10-19 |
Family
ID=37087230
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/KR2006/001343 WO2006109990A1 (fr) | 2005-04-13 | 2006-04-12 | Procedes de codage et de decodage arithmetiques adaptatifs fondes sur le contexte et appareils a efficacite de codage amelioree et procedes de codage et decodage video et appareils mettant en oeuvre ceux-ci |
Country Status (1)
Country | Link |
---|---|
WO (1) | WO2006109990A1 (fr) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2008134623A1 (fr) * | 2007-04-26 | 2008-11-06 | Texas Instruments Incorporated | Procédé de décodage de relevé d'importance du codage cabac utilisable sur des processeurs de données vliw |
CN116366868A (zh) * | 2023-05-31 | 2023-06-30 | 合肥综合性国家科学中心人工智能研究院(安徽省人工智能实验室) | 一种并发视频包过滤方法、系统及储存介质 |
CN116437102A (zh) * | 2023-06-14 | 2023-07-14 | 中国科学技术大学 | 可学习通用视频编码方法、系统、设备及存储介质 |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6646578B1 (en) * | 2002-11-22 | 2003-11-11 | Ub Video Inc. | Context adaptive variable length decoding system and method |
US6690307B2 (en) * | 2002-01-22 | 2004-02-10 | Nokia Corporation | Adaptive variable length coding of digital video |
US6856701B2 (en) * | 2001-09-14 | 2005-02-15 | Nokia Corporation | Method and system for context-based adaptive binary arithmetic coding |
-
2006
- 2006-04-12 WO PCT/KR2006/001343 patent/WO2006109990A1/fr active Application Filing
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6856701B2 (en) * | 2001-09-14 | 2005-02-15 | Nokia Corporation | Method and system for context-based adaptive binary arithmetic coding |
US6690307B2 (en) * | 2002-01-22 | 2004-02-10 | Nokia Corporation | Adaptive variable length coding of digital video |
US6646578B1 (en) * | 2002-11-22 | 2003-11-11 | Ub Video Inc. | Context adaptive variable length decoding system and method |
Non-Patent Citations (1)
Title |
---|
MARPE D. ET AL.: "Context-based adaptive binary arithmetic coding in the H.264/AVC video compression standard", IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, vol. 13, no. 7, July 2003 (2003-07-01), pages 620 - 636 * |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2008134623A1 (fr) * | 2007-04-26 | 2008-11-06 | Texas Instruments Incorporated | Procédé de décodage de relevé d'importance du codage cabac utilisable sur des processeurs de données vliw |
US7813567B2 (en) | 2007-04-26 | 2010-10-12 | Texas Instruments Incorporated | Method of CABAC significance MAP decoding suitable for use on VLIW data processors |
CN116366868A (zh) * | 2023-05-31 | 2023-06-30 | 合肥综合性国家科学中心人工智能研究院(安徽省人工智能实验室) | 一种并发视频包过滤方法、系统及储存介质 |
CN116366868B (zh) * | 2023-05-31 | 2023-08-25 | 合肥综合性国家科学中心人工智能研究院(安徽省人工智能实验室) | 一种并发视频包过滤方法、系统及储存介质 |
CN116437102A (zh) * | 2023-06-14 | 2023-07-14 | 中国科学技术大学 | 可学习通用视频编码方法、系统、设备及存储介质 |
CN116437102B (zh) * | 2023-06-14 | 2023-10-20 | 中国科学技术大学 | 可学习通用视频编码方法、系统、设备及存储介质 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7292165B2 (en) | Context-based adaptive arithmetic coding and decoding methods and apparatuses with improved coding efficiency and video coding and decoding methods and apparatuses using the same | |
US20060233240A1 (en) | Context-based adaptive arithmetic coding and decoding methods and apparatuses with improved coding efficiency and video coding and decoding methods and apparatuses using the same | |
KR101962183B1 (ko) | 인트라 예측 모드 부호화/복호화 방법 및 장치 | |
JP6491691B2 (ja) | 改善されたエントロピー符号化および復号を行う方法および装置 | |
US8351502B2 (en) | Method and apparatus for adaptively selecting context model for entropy coding | |
CN101682773B (zh) | 用于对视频数据进行熵编码和熵解码的方法和设备 | |
US8345752B2 (en) | Method and apparatus for entropy encoding/decoding | |
KR100736104B1 (ko) | 독립적 파싱을 지원하는 비디오 코딩 방법 및 장치 | |
US6925119B2 (en) | Process and system for the compression of digital video signals, a system and a computer program product therefor | |
US20060232452A1 (en) | Method for entropy coding and decoding having improved coding efficiency and apparatus for providing the same | |
US8340181B2 (en) | Video coding and decoding methods with hierarchical temporal filtering structure, and apparatus for the same | |
CA2543947A1 (fr) | Methode et appareil de selection adaptative de modele contextuel pour le codage entropique | |
US20070177664A1 (en) | Entropy encoding/decoding method and apparatus | |
KR101973571B1 (ko) | 영상 변환 부호화/복호화 방법 및 장치 | |
KR100834757B1 (ko) | 엔트로피 부호화 효율을 향상시키는 방법 및 그 방법을이용한 비디오 인코더 및 비디오 디코더 | |
US20050157794A1 (en) | Scalable video encoding method and apparatus supporting closed-loop optimization | |
US7796825B2 (en) | Losslessly improving compression of compressed image data | |
WO2006109990A1 (fr) | Procedes de codage et de decodage arithmetiques adaptatifs fondes sur le contexte et appareils a efficacite de codage amelioree et procedes de codage et decodage video et appareils mettant en oeuvre ceux-ci | |
WO2006112643A1 (fr) | Procedes et appareils de codage et decodage arithmetique adaptatif a base de contexte, a meilleur rendement de codage, et procedes et appareils de codage et decodage video les utilisant | |
WO2006085725A1 (fr) | Procedes de codage et de decodage video comprenant une structure de filtrage temporel hierarchique et appareil correspondant |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
NENP | Non-entry into the national phase |
Ref country code: DE |
|
NENP | Non-entry into the national phase |
Ref country code: RU |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 06747351 Country of ref document: EP Kind code of ref document: A1 |