WO2007026763A1 - ステレオ符号化装置、ステレオ復号装置、及びステレオ符号化方法 - Google Patents
ステレオ符号化装置、ステレオ復号装置、及びステレオ符号化方法 Download PDFInfo
- Publication number
- WO2007026763A1 WO2007026763A1 PCT/JP2006/317104 JP2006317104W WO2007026763A1 WO 2007026763 A1 WO2007026763 A1 WO 2007026763A1 JP 2006317104 W JP2006317104 W JP 2006317104W WO 2007026763 A1 WO2007026763 A1 WO 2007026763A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- signal
- evaluation
- stereo
- time domain
- frequency domain
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/04—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
- G10L19/16—Vocoder architecture
- G10L19/18—Vocoders using multiple modes
- G10L19/24—Variable rate codecs, e.g. for generating different qualities using a scalable representation such as hierarchical encoding or layered encoding
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/008—Multichannel audio signal coding or decoding using interchannel correlation to reduce redundancy, e.g. joint-stereo, intensity-coding or matrixing
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/04—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S3/00—Systems employing more than two channels, e.g. quadraphonic
Definitions
- the present invention relates to a stereo encoding device, a stereo decoding device, and a stereo encoding method.
- the present invention is used when encoding Z decoding of a stereo audio signal or a stereo audio signal in a mobile communication system or a packet communication system using the Internet Protocol (IP).
- IP Internet Protocol
- the present invention relates to a stereo encoding device, a stereo decoding device, and a stereo encoding method.
- a current mobile phone can already be equipped with a multimedia player having a stereo function and an FM radio function. Therefore, it is natural to add functions such as recording and playback of stereo audio signals using only stereo audio signals to 4th generation mobile phones and IP phones.
- MPEG-2 AAC Motion Picture Experts Group-2 Advanced Audio Coding
- MDCT Modified Discrete Cosine Transform
- Non-Patent Document l ISO / IEC 13818-7: 1997- MPEG-2 Advanced Audio Coding (AAC) Invention Disclosure
- MPEG-2 AAC is more suitable for audio signals and has a problem that it is not suitable for audio signals.
- MPEG-2 AAC reduces the bit rate while maintaining good sound quality while reducing the number of quantization bits for spectral information that is not important for audio signal communication.
- the MPEG-2 AAC which provides a very good audio quality in the audio signal, was applied to the audio signal. In some cases, satisfactory sound quality may not be obtained.
- MPEG-2 AAC Another problem with MPEG-2 AAC is the delay due to the algorithm.
- the frame size used for MP EG-2 AAC is 1024 sample Z frames. For example, if the sampling frequency exceeds 32 kHz, the frame delay will be less than 32 milliseconds, which is an acceptable delay in a real-time voice communication system.
- MPEG-2 AAC requires an MDCT process that performs overlap-and-add (overlay addition) of two adjacent frames to decode the encoded signal. This is not suitable for real-time communication systems because processing delays due to the system always occur.
- AMR-WB Adaptive Multi-Rate Wide Band
- An object of the present invention is to encode a stereo signal with a low bit rate with high accuracy, and to suppress a delay in voice communication or the like, a stereo encoding device, a stereo decoding device, and a stereo It is to provide a sign key method.
- the stereo coding apparatus of the present invention performs evaluation in the time domain on the first channel signal of the stereo signal, and encodes the evaluation result, the time domain evaluation means, A frequency domain evaluation unit that divides the frequency band of one channel signal into a plurality of parts, evaluates the first channel signal of each band in the frequency domain, and encodes the evaluation result. take.
- a stereo signal can be encoded with a low bit rate with high accuracy, and a delay in voice communication or the like can be suppressed.
- FIG. 1 is a block diagram showing the main configuration of a stereo coding apparatus according to an embodiment of the present invention.
- FIG. 2 is a block diagram showing a main configuration of a time domain evaluation unit according to an embodiment of the present invention.
- FIG. 3 is a block diagram showing a main configuration of a frequency domain evaluation unit according to an embodiment of the present invention.
- FIG. 4 is a flowchart for explaining the operation of the bit allocation control unit according to the embodiment of the present invention.
- FIG. 5 is a block diagram showing the main configuration of the stereo decoding apparatus according to the embodiment of the present invention.
- FIG. 1 is a block diagram showing the main configuration of stereo coding apparatus 100 according to an embodiment of the present invention.
- Stereo encoding apparatus 100 employs a hierarchical configuration mainly composed of first layer 110 and second layer 120.
- a monaural signal M is generated from the left channel signal L and the right channel signal R constituting the stereo audio signal, and the monaural signal is encoded to generate the code information P and the monaural drive.
- a sound source signal e is generated.
- the first layer 110 is the monaural synthesis unit
- the monaural synthesis unit 101 synthesizes the monaural signal M with the left channel signal L, the right channel signal R, and the force.
- the monaural code unit 102 employs the configuration of an AMR-WB encoding device.
- the monaural code unit 102 encodes the monaural signal M output from the monaural synthesizing unit 101 using the AMR-WB method, obtains encoded information P, and outputs the encoded information P to the multiplexing unit 108. Also mono
- the sign key unit 102 outputs the monaural driving sound source signal e obtained in the sign key process to the second level.
- the second layer 120 evaluation and prediction (prediction and estimation) in the time domain and the frequency domain are performed on the stereo audio signal, and various types of encoded information are generated.
- spatial information included in the left channel signal L constituting the stereo audio signal is detected and calculated. Due to this spatial information, the stereo audio signal gives a sense of presence (a sense of spread).
- an evaluation signal similar to the left channel signal L is generated by applying this spatial information to the monaural signal. Then, information regarding each process is output as sign key information.
- the second layer 120 includes a filtering unit 103, a time domain evaluation unit 104, a frequency domain evaluation unit 105, a residual encoding unit 106, and a bit allocation control unit 107, and each unit performs the following operations.
- Filtering section 103 generates an LPC (Linear Predictive Coding) coefficient from left channel signal L by LPC analysis, and outputs it as multiplexing information P to multiplexing section 108.
- LPC Linear Predictive Coding
- the filtering unit 103 generates a left channel drive sound source signal e using the left channel signal L and the LPC coefficient, and outputs it to the time domain evaluation unit 104.
- the time domain evaluation unit 104 includes a monaural driving sound source signal e generated in the monaural coding unit 102 of the first layer 110 and a left channel generated in the filtering unit 103.
- a valence signal e is generated and output to the frequency domain evaluation unit 105. That is, the time domain evaluation unit 1 estl
- the frequency domain evaluation unit 105 includes a left channel driving sound source signal e generated by the filtering unit 103 and a time domain evaluation signal e generated by the time domain evaluation unit 104.
- Target information is detected and calculated.
- Residual sign key unit 106 includes frequency domain evaluation signal e generated in frequency domain evaluation unit 105 and left channel drive excitation signal est2 generated in filtering unit 103.
- a residual signal is obtained from e, and this signal is encoded and encoded information P is generated and multiplexed.
- the bit allocation control unit 107 includes a monaural driving sound source signal e generated in the monaural code key unit 102 and a left channel driving sound source signal generated in the filtering unit 103.
- the time domain evaluation unit 104 the frequency domain evaluation unit 105, and the residual
- Encoding bits are allocated to the encoding unit 106.
- the bit allocation control unit 107 encodes information on the number of bits allocated to each unit, and the obtained code information P
- the multiplexing unit 108 multiplexes the code information up to P power up to P, and the bit stream after multiplexing is multiplexed.
- the stereo decoding apparatus corresponding to the stereo encoding apparatus 100 includes the code signal information P of the monaural signal generated in the first layer 110 and the left channel signal generated in the second layer 120.
- the signal can be decoded.
- a right channel signal can also be generated from the decoded monaural signal and left channel signal.
- FIG. 2 is a block diagram showing a main configuration of the time domain evaluation unit 104.
- the time domain evaluation unit 104 receives the monaural driving sound source signal e as a target signal and drives the left channel driving sound source signal.
- No. e is input as a reference signal.
- the spatial information in the time domain consists of amplitude information (and delay information ⁇ ).
- the energy calculation unit 141 1 receives a monaural driving sound source signal e and receives the signal.
- the energy in the interspace is calculated.
- the energy calculating unit 141 2 receives the driving sound source signal e of the left channel and receives the energy.
- the ratio calculation unit 142 receives the energy values calculated by the energy calculation units 141-1 and 141-2, respectively, and outputs the monaural driving sound source signal e and the left channel driving sound source signal.
- the energy ratio to e is calculated, and the monaural driving sound source signal e and the left channel driving sound source signal are calculated.
- Correlation value calculation section 143 receives monaural driving sound source signal e and left channel driving sound source signal e.
- the delay detection unit 144 receives the cross-correlation value calculated by the correlation value calculation unit 143, detects a time delay between the left channel driving sound source signal e and the monaural driving sound source signal e, and
- the evaluation signal generation unit 145 generates the left channel from the monaural driving sound source signal e based on the amplitude information ⁇ calculated by the ratio calculation unit 142 and the delay information ⁇ calculated by the delay detection unit 144.
- the time domain evaluation unit 104 performs spatial analysis in the time domain between the monaural driving sound source signal e and the left channel driving sound source signal e once per frame of the audio signal processing.
- C Spatial information is composed of amplitude information (X and delay information.
- the time domain evaluation unit 104 gives this spatial information to the monaural driving sound source signal e and is similar to the left channel driving sound source signal e.
- a time domain evaluation signal e is generated.
- FIG. 3 is a block diagram showing the main configuration of the frequency domain evaluation unit 105.
- the frequency domain evaluation unit 105 uses the time domain evaluation signal e generated by the time domain evaluation unit 104 as a target signal.
- the left channel drive sound source signal e is input as a reference signal in the frequency domain.
- the spatial information in the frequency domain is composed of spectrum amplitude information
- FFT 3 ⁇ 4151 The left channel driving sound source signal e , which is a time domain signal, is converted into a frequency domain signal (spectrum) by fast Fourier transform (FFT).
- the dividing unit 152-1 divides the frequency domain signal band generated by the FFT unit 151-1 into a plurality of bands (subbands). Each subband may follow a Bark Scale corresponding to the human auditory system, or may be equally divided within the bandwidth.
- the energy calculation unit 153-1 calculates the spectral energy of the driving sound source signal e of the left channel.
- the calculation is performed for each subband output from the dividing unit 152-1.
- the FFT unit 151-2 converts the time domain evaluation signal e estl into a frequency domain signal by the same processing as the FFT unit 151-1.
- Divider 152-2 divides the frequency domain signal band generated by FFT unit 151-2 into a plurality of subbands by the same processing as divider 152-1.
- the energy calculation unit 153-2 performs the same processing as the energy calculation unit 153-1 on the spectral energy of the time domain evaluation signal e by each sub-buffer estl output from the division unit 152-2.
- the ratio calculation unit 154 uses the spectral energy of each subband calculated by the energy calculation unit 153-1 and the energy calculation unit 153-2 to drive the left channel driving sound source signal e
- the phase calculation unit 155-1 includes each of the sub-bands of the driving sound source signal e of the left channel.
- Phase selection section 156 selects one phase suitable for encoding from the phase of the spectrum in each subband in order to reduce the amount of encoded information.
- Phase calculation section 155-2 calculates the phase of each spectrum in each subband of time domain evaluation signal e by the same processing as phase calculation section 155-1.
- Phase difference calculation section 157 calculates the phase difference between drive excitation signal e for the left channel and time domain evaluation signal e for the phase in each subband selected by phase selection section 156.
- the evaluation signal generator 158 includes a left channel driving sound source signal e, a time domain evaluation signal e,
- Amplitude information ⁇ 8 during L estl, left channel drive source signal e and time domain evaluation signal e To the frequency domain estl from the time domain evaluation signal e based on both of the phase difference information ⁇
- An evaluation signal e is generated.
- the frequency domain evaluation unit 105 uses the left-channel driving sound source signal e and the time domain evaluation signal e generated by the time domain evaluation unit 104 as a plurality of subbands estl.
- the frequency domain evaluation unit 105 calculates a time domain evaluation signal e that is similar to the left-channel driving sound source signal e obtained by the time domain evaluation, by frequency domain evaluation.
- a frequency domain evaluation signal e similar to the signal is generated. Further, the frequency domain evaluation unit 105 converts this spatial information into the time domain est2
- a frequency domain evaluation signal e similar to L is generated.
- bit allocation control unit 107 For each frame of the audio signal, the number of bits allocated to the code ⁇ is predetermined. The bit allocation control unit 107 determines whether or not the left channel driving sound source signal e and the monaural driving sound source signal e are similar to each other in order to achieve optimum sound quality at the predetermined bit rate.
- the number of bits allocated to each processing unit is adaptively determined.
- FIG. 4 is a flowchart for explaining the operation of the bit distribution control unit 107.
- bit allocation control unit 107 performs the monaural driving sound source signal e.
- the bit allocation control unit 107 calculates a mean square error between the monaural driving sound source signal e and the left channel driving sound source signal e, and uses this as a predetermined threshold value.
- the difference between the two signals in the time domain is less than
- the number of bits required to make it smaller may be smaller.
- the time domain evaluation unit 104 has a smaller number of other parts (frequency domain evaluation unit 105, residual code unit 106), in particular non-uniformity that allocates more bits to the frequency domain evaluation unit 105. If bit allocation is performed efficiently, code efficiency is improved because of efficient bit allocation. Therefore, if the bit allocation control unit 107 determines that they are similar in ST1072, it allocates a smaller number of bits to the time domain evaluation in ST1073 and distributes the remaining bits equally to other processing in ST1074. To do.
- the monaural driving sound source signal e and the left channel driving sound source signal e are not similar.
- the bit allocation control unit 107 in ST1072, outputs the monaural driving sound source signal e and the left channel driving sound source signal.
- FIG. 5 is a block diagram showing the main configuration of stereo decoding apparatus 200 according to the present embodiment.
- the stereo decoding apparatus 200 has a hierarchical configuration mainly including the first layer 210 and the second layer 220 as in the stereo encoding apparatus 100. Also, each process of stereo decoding apparatus 200 is basically an inverse process of each process corresponding to stereo encoding apparatus 100. That is, the stereo decoding apparatus 200 predicts and generates a monaural signal power left channel signal using the encoded information sent from the stereo encoding apparatus 100, and further uses the monaural signal and the left channel signal, Generate a right channel signal.
- Separating section 201 separates the input bitstream into encoded information with P power up to P.
- First layer 210 includes monaural decoding section 202.
- the monaural decoding unit 202 Decoding encoded information P to generate monaural signal M ′ and monaural driving sound source signal e
- the second layer 220 includes a bit allocation information decoding unit 203, a time domain evaluation unit 204, a frequency domain evaluation unit 205, and a residual decoding unit 206, and each unit performs the following operations.
- the bit allocation information decoding unit 203 decodes the encoded information P, and the time domain evaluation unit 204
- the number of bits used by wave number domain evaluation section 205 and residual decoding section 206 is output.
- the time domain evaluation unit 204 includes the monaural driving sound source signal e 'generated by the monaural decoding unit 202, the code key information P output from the separation unit 201, and the bit allocation information decoding unit.
- time domain evaluation signal e ′ is generated.
- the frequency domain evaluation unit 205 the time domain evaluation signal e 'generated in the time domain evaluation unit 204, the code key information P output from the separation unit 201, and the bit allocation information decoding estl D
- the frequency domain evaluation unit 205 uses the number of bits passed from the unit 203, evaluation and prediction in the frequency domain are performed, and a frequency domain evaluation signal e ′ is generated.
- the frequency domain evaluation unit 205 performs est2 in the frequency domain.
- an FFT unit that performs frequency conversion is provided prior to evaluation and prediction.
- Residual decoding section 206 includes sign key information P and bit allocation information output from separation section 201.
- the residual signal is decoded using the number of bits passed from the decoding unit 203.
- the residual decoding unit 206 converts this est2 into the frequency domain evaluation signal e ′ generated by the frequency domain evaluation unit 205.
- the generated residual signal is given, and the left channel drive sound source signal e ′ is generated.
- Synthesis filtering section 207 decodes the encoded information P-power LPC coefficient, and outputs this LPC coefficient.
- Stereo conversion section 208 generates right channel signal R using monaural signal M decoded by monaural decoding section 202 and left channel signal L generated by synthesis filter 207.
- stereo audio signals are first evaluated and predicted in the time domain, then further detailed evaluation and prediction are performed in the frequency domain, and information on these two-level evaluation and prediction is output as code information. . Therefore, complementary evaluation and prediction can be performed in the frequency domain for powerful information that cannot be sufficiently expressed by evaluation and prediction in the time domain, and a stereo audio signal can be accurately encoded at a low bit rate. can do.
- the time domain evaluation in time domain evaluation section 104 corresponds to evaluating the average level of the spatial information of the signal over the entire frequency band.
- the time domain evaluation unit 104 processes the signal to be encoded of one frame as it is as one signal, and the entire or average energy of this signal is processed. The ratio and time delay are obtained.
- the frequency domain evaluation in the frequency domain evaluation unit 105 divides the frequency band of the signal to be encoded into a plurality of subbands, and evaluates the subdivided individual signals.
- the evaluation signal is finely adjusted by performing further evaluation in the frequency domain. Therefore, since the signal that is subject to encoding is treated as a single signal, it is subdivided into multiple signals for further evaluation, so that the accuracy of encoding of the stereo audio signal is improved. be able to.
- a predetermined bit rate is within a range according to the degree of similarity between the monaural signal and the left channel signal (or right channel signal), that is, according to the situation of the stereo audio signal.
- bits are allocated appropriately for each processing such as time domain evaluation and frequency domain evaluation. As a result, it is possible to perform coding efficiently and accurately, and to realize bit rate scalability.
- the bit rate can be reduced because the time domain evaluation is performed with few parameters such as the energy ratio and the time delay.
- the bit rate can be reduced because the time domain evaluation is performed with few parameters such as the energy ratio and the time delay.
- the present embodiment since a hierarchical configuration having two layer forces is adopted, it is possible to scale from a mono level to a stereo level. Therefore, even if information related to frequency domain evaluation cannot be decoded for some reason, decoding only information related to time domain evaluation can decode a stereo audio signal of a predetermined quality, although the quality is somewhat degraded. , Can improve scalability.
- the bit rate can be kept low.
- stereo coding apparatus stereo decoding apparatus, and stereo coding method according to the present embodiment can be implemented with various modifications.
- monaural signal and left channel signal are to be encoded by stereo encoding apparatus 100, and stereo decoding apparatus 200 decodes these signals by decoding the monaural signal and left channel signal.
- stereo decoding apparatus 200 decodes these signals by decoding the monaural signal and left channel signal.
- the case where the right channel signal is decoded by synthesizing the decoded signal has been described as an example.
- the signal to be encoded by the stereo encoding apparatus 100 is not limited to this, and the stereo encoding apparatus 100 is monaural.
- the left channel signal may be generated by combining the right channel signal and the monaural signal decoded by the stereo decoding apparatus 200 with the signal and the right channel signal as encoding targets.
- filtering section 103 uses encoded information for the LPC coefficients by converting the LPC coefficients into other equivalent parameters (for example, LSP parameters). Also good.
- the number of bits used in each unit is determined in advance without performing the bit allocation control process in which a predetermined number of bits are allocated to each process by the bit allocation control unit 107.
- Fixed bit allocation may be performed.
- the bit allocation control unit 107 is not required in the stereo encoding device 100.
- this fixed bit allocation ratio is common to the stereo encoding device 100 and the stereo decoding device 200, the bit allocation information decoding unit 203 is not required in the stereo decoding device 200 as well.
- the bit allocation control unit 107 adaptively allocates bits according to the status of the stereo audio signal and adaptively allocates bits according to the status of the network. May be.
- the residual code key unit 106 performs sign keying using a predetermined number of bits distributed by the bit allocation control unit 107, thereby enabling a lossy system.
- An example of encoding using a predetermined number of bits is vector quantization.
- the residual code part is an encoding system with different characteristics, such as a lossy system or a lossless system, depending on the difference in the sign method.
- the lossless system has the feature that the signal can be decoded more accurately by the decoding device than the lossy system, but the bit rate increases because the compression ratio is low.
- the residual signal is coded by the residual code key unit 106 by a noiseless code method such as Huffin coding or Rice code key, a lossless system is obtained.
- the ratio calculation unit 142 performs the monaural driving sound source signal e and the left channel.
- the energy ratio with the drive sound source signal e is calculated as amplitude information ⁇ .
- the amplitude information ⁇ can be used instead of calculating the energy difference.
- ratio calculation section 154 calculates a spectral energy ratio between left-channel drive sound source signal e and time-domain evaluation signal e in each subband, and performs a vibration calculation.
- the width information j8 is used, but instead of the energy ratio, an energy difference may be calculated and used as the amplitude information j8.
- Spatial information in the time domain is composed of amplitude information oc and delay information ⁇ .
- the spatial information may further include other information, or may include other information that is completely different from the amplitude information ⁇ , the delay information, and the like.
- Spatial information in the frequency domain between L estl consists of amplitude information and phase difference information ⁇ , but this spatial information may further contain other information, amplitude information j8, phase difference Information may consist of other information completely different from ⁇ .
- time domain evaluation section 104 uses monaural driving sound source signal e and left channel.
- the detection and calculation of the spatial information with the driving sound source signal e of the channel is performed for each frame.
- This process can be performed multiple times within one frame.
- phase selection section 156 has one spectrum in each subband. The force that selects the phase of the spectrum. Multiple spectral phases may be selected.
- the phase difference calculation unit 157 calculates the average of the phase differences ⁇ between the left channel driving sound source signal e and the time domain evaluation signal e in the plurality of phases, and outputs the average to the phase difference calculation unit 157.
- residual code unit 106 performs time-domain coding on the residual signal, but may perform frequency-domain coding.
- the stereo encoding device, the stereo decoding device, and the stereo encoding method according to the present invention can It can also be applied to audio signals.
- the stereo encoding device and the stereo decoding device according to the present invention can be mounted on a communication terminal device and a base station device in a mobile communication system, and thereby have communication effects similar to those described above.
- a terminal device, a base station device, and a mobile communication system can be provided.
- the power described by taking the case where the present invention is configured as nodeware as an example can be realized by software.
- the stereo coding method and the stereo decoding method algorithm according to the present invention are described in a programming language, and the program is stored in a memory and executed by an information processing means, whereby the stereo coding method according to the present invention is performed. It is possible to realize the same functions as those of the ⁇ and stereo decoding device.
- Each functional block used in the description of each of the above embodiments is typically realized as an LSI which is an integrated circuit. These may be individually made into one chip, or may be made into one chip so as to include some or all of them.
- the method of circuit integration is not limited to LSI's, and implementation using dedicated circuitry or general purpose processors is also possible. It is also possible to use a field programmable gate array (FPGA) that can be programmed after LSI manufacturing, or a reconfigurable processor that can reconfigure the connection or setting of circuit cells inside the LSI. Furthermore, if integrated circuit technology that replaces LSI emerges as a result of progress in semiconductor technology or other derived technology, it is naturally also possible to perform functional block integration using that technology. There is a possibility of adaptation of biotechnology.
- FPGA field programmable gate array
- the stereo encoding device, stereo decoding device, and stereo encoding method according to the present invention are suitable for mobile phones, IP phones, video conferences, and the like.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Signal Processing (AREA)
- Acoustics & Sound (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Health & Medical Sciences (AREA)
- Computational Linguistics (AREA)
- Human Computer Interaction (AREA)
- Multimedia (AREA)
- Mathematical Physics (AREA)
- Quality & Reliability (AREA)
- Compression, Expansion, Code Conversion, And Decoders (AREA)
- Stereophonic System (AREA)
Abstract
Description
Claims
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP06797077A EP1912206B1 (en) | 2005-08-31 | 2006-08-30 | Stereo encoding device, stereo decoding device, and stereo encoding method |
CN2006800319487A CN101253557B (zh) | 2005-08-31 | 2006-08-30 | 立体声编码装置及立体声编码方法 |
KR1020087005096A KR101340233B1 (ko) | 2005-08-31 | 2006-08-30 | 스테레오 부호화 장치, 스테레오 복호 장치 및 스테레오부호화 방법 |
US12/064,995 US8457319B2 (en) | 2005-08-31 | 2006-08-30 | Stereo encoding device, stereo decoding device, and stereo encoding method |
JP2007533292A JP5171256B2 (ja) | 2005-08-31 | 2006-08-30 | ステレオ符号化装置、ステレオ復号装置、及びステレオ符号化方法 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2005252778 | 2005-08-31 | ||
JP2005-252778 | 2005-08-31 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2007026763A1 true WO2007026763A1 (ja) | 2007-03-08 |
Family
ID=37808848
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2006/317104 WO2007026763A1 (ja) | 2005-08-31 | 2006-08-30 | ステレオ符号化装置、ステレオ復号装置、及びステレオ符号化方法 |
Country Status (6)
Country | Link |
---|---|
US (1) | US8457319B2 (ja) |
EP (1) | EP1912206B1 (ja) |
JP (1) | JP5171256B2 (ja) |
KR (1) | KR101340233B1 (ja) |
CN (1) | CN101253557B (ja) |
WO (1) | WO2007026763A1 (ja) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2011527763A (ja) * | 2008-07-11 | 2011-11-04 | サムスン エレクトロニクス カンパニー リミテッド | マルチチャネルの符号化及び復号化方法並びに装置 |
KR101274802B1 (ko) | 2008-12-29 | 2013-06-13 | 모토로라 모빌리티 엘엘씨 | 오디오 신호를 인코딩하기 위한 장치 및 방법 |
KR101274827B1 (ko) | 2008-12-29 | 2013-06-13 | 모토로라 모빌리티 엘엘씨 | 다수 채널 오디오 신호를 디코딩하기 위한 장치 및 방법, 및 다수 채널 오디오 신호를 코딩하기 위한 방법 |
KR101275892B1 (ko) | 2008-12-29 | 2013-06-17 | 모토로라 모빌리티 엘엘씨 | 오디오 신호를 인코딩하고 디코딩하기 위한 방법 및 장치 |
JP5413839B2 (ja) * | 2007-10-31 | 2014-02-12 | パナソニック株式会社 | 符号化装置および復号装置 |
WO2024111300A1 (ja) * | 2022-11-22 | 2024-05-30 | 富士フイルム株式会社 | 音データ作成方法及び音データ作成装置 |
Families Citing this family (27)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7461106B2 (en) | 2006-09-12 | 2008-12-02 | Motorola, Inc. | Apparatus and method for low complexity combinatorial coding of signals |
US8576096B2 (en) | 2007-10-11 | 2013-11-05 | Motorola Mobility Llc | Apparatus and method for low complexity combinatorial coding of signals |
US8209190B2 (en) | 2007-10-25 | 2012-06-26 | Motorola Mobility, Inc. | Method and apparatus for generating an enhancement layer within an audio coding system |
JP5153791B2 (ja) * | 2007-12-28 | 2013-02-27 | パナソニック株式会社 | ステレオ音声復号装置、ステレオ音声符号化装置、および消失フレーム補償方法 |
US7889103B2 (en) | 2008-03-13 | 2011-02-15 | Motorola Mobility, Inc. | Method and apparatus for low complexity combinatorial coding of signals |
JP5340261B2 (ja) * | 2008-03-19 | 2013-11-13 | パナソニック株式会社 | ステレオ信号符号化装置、ステレオ信号復号装置およびこれらの方法 |
US8639519B2 (en) | 2008-04-09 | 2014-01-28 | Motorola Mobility Llc | Method and apparatus for selective signal coding based on core encoder performance |
US8175888B2 (en) | 2008-12-29 | 2012-05-08 | Motorola Mobility, Inc. | Enhanced layered gain factor balancing within a multiple-channel audio coding system |
WO2010091555A1 (zh) * | 2009-02-13 | 2010-08-19 | 华为技术有限公司 | 一种立体声编码方法和装置 |
US8848925B2 (en) | 2009-09-11 | 2014-09-30 | Nokia Corporation | Method, apparatus and computer program product for audio coding |
KR101710113B1 (ko) * | 2009-10-23 | 2017-02-27 | 삼성전자주식회사 | 위상 정보와 잔여 신호를 이용한 부호화/복호화 장치 및 방법 |
CN102081927B (zh) * | 2009-11-27 | 2012-07-18 | 中兴通讯股份有限公司 | 一种可分层音频编码、解码方法及系统 |
US8423355B2 (en) | 2010-03-05 | 2013-04-16 | Motorola Mobility Llc | Encoder for audio signal including generic audio and speech frames |
SG10202101745XA (en) | 2010-04-09 | 2021-04-29 | Dolby Int Ab | Audio Upmixer Operable in Prediction or Non-Prediction Mode |
ES2911893T3 (es) * | 2010-04-13 | 2022-05-23 | Fraunhofer Ges Forschung | Codificador de audio, decodificador de audio y métodos relacionados para procesar señales de audio estéreo usando una dirección de predicción variable |
KR101276049B1 (ko) * | 2012-01-25 | 2013-06-20 | 세종대학교산학협력단 | 조건부 스플릿 벡터 양자화를 이용한 음성 압축 장치 및 그 방법 |
JP6063555B2 (ja) | 2012-04-05 | 2017-01-18 | 華為技術有限公司Huawei Technologies Co.,Ltd. | マルチチャネルオーディオエンコーダ及びマルチチャネルオーディオ信号を符号化する方法 |
CN104170007B (zh) * | 2012-06-19 | 2017-09-26 | 深圳广晟信源技术有限公司 | 对单声道或立体声进行编码的方法 |
KR102204136B1 (ko) * | 2012-08-22 | 2021-01-18 | 한국전자통신연구원 | 오디오 부호화 장치 및 방법, 오디오 복호화 장치 및 방법 |
US9129600B2 (en) | 2012-09-26 | 2015-09-08 | Google Technology Holdings LLC | Method and apparatus for encoding an audio signal |
CN105247613B (zh) * | 2013-04-05 | 2019-01-18 | 杜比国际公司 | 音频处理系统 |
EP3067887A1 (en) | 2015-03-09 | 2016-09-14 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Audio encoder for encoding a multichannel signal and audio decoder for decoding an encoded audio signal |
US12125492B2 (en) | 2015-09-25 | 2024-10-22 | Voiceage Coproration | Method and system for decoding left and right channels of a stereo sound signal |
MY188370A (en) * | 2015-09-25 | 2021-12-06 | Voiceage Corp | Method and system for decoding left and right channels of a stereo sound signal |
USD793458S1 (en) | 2015-12-24 | 2017-08-01 | Samsung Electronics Co., Ltd. | Ice machine for refrigerator |
USD794093S1 (en) | 2015-12-24 | 2017-08-08 | Samsung Electronics Co., Ltd. | Ice machine handle for refrigerator |
CN110660400B (zh) * | 2018-06-29 | 2022-07-12 | 华为技术有限公司 | 立体声信号的编码、解码方法、编码装置和解码装置 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH10105193A (ja) * | 1996-09-26 | 1998-04-24 | Yamaha Corp | 音声符号化伝送方式 |
JPH11317672A (ja) * | 1997-11-20 | 1999-11-16 | Samsung Electronics Co Ltd | ビット率の調節可能なステレオオーディオ符号化/復号化方法及び装置 |
JP2004289196A (ja) * | 2002-03-08 | 2004-10-14 | Nippon Telegr & Teleph Corp <Ntt> | ディジタル信号符号化方法、復号化方法、符号化装置、復号化装置及びディジタル信号符号化プログラム、復号化プログラム |
JP2004302259A (ja) * | 2003-03-31 | 2004-10-28 | Matsushita Electric Ind Co Ltd | 音響信号の階層符号化方法および階層復号化方法 |
JP2005517987A (ja) * | 2002-02-18 | 2005-06-16 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | パラメトリックオーディオ符号化 |
JP2005252778A (ja) | 2004-03-05 | 2005-09-15 | Oki Electric Ind Co Ltd | 時間分割波長ホップ光符号による通信方法及び通信装置 |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1021044A1 (en) * | 1999-01-12 | 2000-07-19 | Deutsche Thomson-Brandt Gmbh | Method and apparatus for encoding or decoding audio or video frame data |
US7292901B2 (en) | 2002-06-24 | 2007-11-06 | Agere Systems Inc. | Hybrid multi-channel/cue coding/decoding of audio signals |
AU2003213439A1 (en) * | 2002-03-08 | 2003-09-22 | Nippon Telegraph And Telephone Corporation | Digital signal encoding method, decoding method, encoding device, decoding device, digital signal encoding program, and decoding program |
BRPI0304540B1 (pt) * | 2002-04-22 | 2017-12-12 | Koninklijke Philips N. V | Methods for coding an audio signal, and to decode an coded audio sign, encoder to codify an audio signal, codified audio sign, storage media, and, decoder to decode a coded audio sign |
KR100528325B1 (ko) | 2002-12-18 | 2005-11-15 | 삼성전자주식회사 | 비트율 조절이 가능한 스테레오 오디오 부호화 및복호화방법 및 그 장치 |
CN1748247B (zh) | 2003-02-11 | 2011-06-15 | 皇家飞利浦电子股份有限公司 | 音频编码 |
KR20050116828A (ko) * | 2003-03-24 | 2005-12-13 | 코닌클리케 필립스 일렉트로닉스 엔.브이. | 다채널 신호를 나타내는 주 및 부 신호의 코딩 |
WO2005027095A1 (ja) * | 2003-09-16 | 2005-03-24 | Matsushita Electric Industrial Co., Ltd. | 符号化装置および復号化装置 |
-
2006
- 2006-08-30 US US12/064,995 patent/US8457319B2/en active Active
- 2006-08-30 KR KR1020087005096A patent/KR101340233B1/ko not_active Expired - Fee Related
- 2006-08-30 CN CN2006800319487A patent/CN101253557B/zh not_active Expired - Fee Related
- 2006-08-30 JP JP2007533292A patent/JP5171256B2/ja not_active Expired - Fee Related
- 2006-08-30 EP EP06797077A patent/EP1912206B1/en not_active Not-in-force
- 2006-08-30 WO PCT/JP2006/317104 patent/WO2007026763A1/ja active Application Filing
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH10105193A (ja) * | 1996-09-26 | 1998-04-24 | Yamaha Corp | 音声符号化伝送方式 |
JPH11317672A (ja) * | 1997-11-20 | 1999-11-16 | Samsung Electronics Co Ltd | ビット率の調節可能なステレオオーディオ符号化/復号化方法及び装置 |
JP2005517987A (ja) * | 2002-02-18 | 2005-06-16 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | パラメトリックオーディオ符号化 |
JP2004289196A (ja) * | 2002-03-08 | 2004-10-14 | Nippon Telegr & Teleph Corp <Ntt> | ディジタル信号符号化方法、復号化方法、符号化装置、復号化装置及びディジタル信号符号化プログラム、復号化プログラム |
JP2004302259A (ja) * | 2003-03-31 | 2004-10-28 | Matsushita Electric Ind Co Ltd | 音響信号の階層符号化方法および階層復号化方法 |
JP2005252778A (ja) | 2004-03-05 | 2005-09-15 | Oki Electric Ind Co Ltd | 時間分割波長ホップ光符号による通信方法及び通信装置 |
Non-Patent Citations (4)
Title |
---|
GOTO M. ET AL.: "Onsei Tushinyo Scalable Stereo Onsei Fugoka Hoho no Kento", DAI 4 KAI FORUM ON INFORMATION TECHNOLOGY KOEN RONBUNSHU, 22 August 2005 (2005-08-22), pages 299 - 300, XP003009891 * |
OSHIKIRI M. ET AL.: "Pitch Filtering ni Motozuku Spectrum Fugoka o Mochiita Cho Kotaiiki Scalale Onsei Fugoka no Kaizen", JOURNAL OF THE ACOUSTICAL SOCIETY OF JAPAN 2004 NEN SHUKI KENKYU HAPPYOKAI KOEN RONBUNSHU-I-, 21 September 2004 (2004-09-21), pages 297 - 298, XP002994276 * |
See also references of EP1912206A4 |
YOSHIDA K. ET AL.: "Scalable Stereo Onsei Fugoka no Channel-kan Yosoku ni Kansuru Yobi Kento", PROCEEDINGS OF THE 2005 IEICE GENERAL CONFERENCE, 7 March 2005 (2005-03-07), pages 118, XP003009892 * |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5413839B2 (ja) * | 2007-10-31 | 2014-02-12 | パナソニック株式会社 | 符号化装置および復号装置 |
JP2011527763A (ja) * | 2008-07-11 | 2011-11-04 | サムスン エレクトロニクス カンパニー リミテッド | マルチチャネルの符号化及び復号化方法並びに装置 |
KR101274802B1 (ko) | 2008-12-29 | 2013-06-13 | 모토로라 모빌리티 엘엘씨 | 오디오 신호를 인코딩하기 위한 장치 및 방법 |
KR101274827B1 (ko) | 2008-12-29 | 2013-06-13 | 모토로라 모빌리티 엘엘씨 | 다수 채널 오디오 신호를 디코딩하기 위한 장치 및 방법, 및 다수 채널 오디오 신호를 코딩하기 위한 방법 |
KR101275892B1 (ko) | 2008-12-29 | 2013-06-17 | 모토로라 모빌리티 엘엘씨 | 오디오 신호를 인코딩하고 디코딩하기 위한 방법 및 장치 |
WO2024111300A1 (ja) * | 2022-11-22 | 2024-05-30 | 富士フイルム株式会社 | 音データ作成方法及び音データ作成装置 |
Also Published As
Publication number | Publication date |
---|---|
JP5171256B2 (ja) | 2013-03-27 |
US8457319B2 (en) | 2013-06-04 |
US20090262945A1 (en) | 2009-10-22 |
CN101253557A (zh) | 2008-08-27 |
EP1912206A1 (en) | 2008-04-16 |
JPWO2007026763A1 (ja) | 2009-03-26 |
KR101340233B1 (ko) | 2013-12-10 |
CN101253557B (zh) | 2012-06-20 |
KR20080039462A (ko) | 2008-05-07 |
EP1912206A4 (en) | 2011-03-23 |
EP1912206B1 (en) | 2013-01-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP5171256B2 (ja) | ステレオ符号化装置、ステレオ復号装置、及びステレオ符号化方法 | |
EP1798724B1 (en) | Encoder, decoder, encoding method, and decoding method | |
CN101432802B (zh) | 使用有损编码的数据流和无损扩展数据流对源信号进行无损编码的方法以及设备 | |
JP5383676B2 (ja) | 符号化装置、復号装置およびこれらの方法 | |
JP5413839B2 (ja) | 符号化装置および復号装置 | |
JP5404412B2 (ja) | 符号化装置、復号装置およびこれらの方法 | |
EP2133872B1 (en) | Encoding device and encoding method | |
JPWO2006022308A1 (ja) | マルチチャネル信号符号化装置およびマルチチャネル信号復号装置 | |
WO2010016270A1 (ja) | 量子化装置、符号化装置、量子化方法及び符号化方法 | |
JP4606418B2 (ja) | スケーラブル符号化装置、スケーラブル復号装置及びスケーラブル符号化方法 | |
WO2006046547A1 (ja) | 音声符号化装置および音声符号化方法 | |
WO2006129615A1 (ja) | スケーラブル符号化装置およびスケーラブル符号化方法 | |
JPWO2008132826A1 (ja) | ステレオ音声符号化装置およびステレオ音声符号化方法 | |
Bang et al. | Audio Transcoding Algorithm for Mobile Multimedia Application |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 200680031948.7 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
ENP | Entry into the national phase |
Ref document number: 2007533292 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2006797077 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 12064995 Country of ref document: US |
|
WWE | Wipo information: entry into national phase |
Ref document number: 1020087005096 Country of ref document: KR |
|
NENP | Non-entry into the national phase |
Ref country code: DE |