WO1999066497A1 - Voice/music signal encoder and decoder - Google Patents
Voice/music signal encoder and decoder Download PDFInfo
- Publication number
- WO1999066497A1 WO1999066497A1 PCT/JP1999/003185 JP9903185W WO9966497A1 WO 1999066497 A1 WO1999066497 A1 WO 1999066497A1 JP 9903185 W JP9903185 W JP 9903185W WO 9966497 A1 WO9966497 A1 WO 9966497A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- signal
- linear prediction
- music
- audio
- excitation
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/04—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
- G10L19/08—Determination or coding of the excitation function; Determination or coding of the long-term prediction parameters
- G10L19/12—Determination or coding of the excitation function; Determination or coding of the long-term prediction parameters the excitation function being a code excitation, e.g. in code excited linear prediction [CELP] vocoders
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/02—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using spectral analysis, e.g. transform vocoders or subband vocoders
- G10L19/0204—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using spectral analysis, e.g. transform vocoders or subband vocoders using subband decomposition
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/02—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using spectral analysis, e.g. transform vocoders or subband vocoders
- G10L19/0212—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using spectral analysis, e.g. transform vocoders or subband vocoders using orthogonal transformation
Definitions
- the present invention relates to an encoding device and a decoding device for transmitting an audio / music signal at a low bit rate.
- CELP Code Excited Linear Prediction
- CELF Code Excited Linear Prediction
- a synthesized speech signal is obtained by adding a linear prediction filter in which linear prediction coefficients obtained by linearly predicting and decoding input speech are set to the sum of a signal representing the pitch period of speech and a noise-like signal. It is generated by driving with the sound source signal represented by.
- the CELP is described in M. Schroeder et al., “Code excited linear prediction: High quality speech at very low bit ratesj (Proc. ICASSP, pp. 937-940, 1985) (Reference 1).
- the encoding performance for music signals can be improved by using a band division configuration, in which the reproduced signal is an excitation signal obtained by adding the excitation signal corresponding to each band, and drives the linear prediction synthesis filter. Generated by
- FIG. 1 is a block diagram showing an example of a conventional audio / music signal encoding device.
- the number of bands is assumed to be 2 for simplicity.
- An input signal (input vector) generated by sampling a voice or music signal and combining the plurality of samples into one vector as one frame is input from an input terminal 10.
- the input vector from the input terminal 10 is input to the linear prediction coefficient calculation circuit 170.
- the linear prediction coefficient calculation circuit 170 performs a linear prediction analysis on the input vector. (Input vector) is input from input terminal 10.
- the input vector from the input terminal 10 is input to the linear prediction coefficient calculation circuit 170.
- the linear prediction coefficient calculation circuit 170 performs a linear prediction analysis on the input vector to obtain a linear prediction coefficient.
- the linear prediction coefficient calculation circuit 170 further quantizes the linear prediction coefficient to obtain a quantized linear prediction coefficient.
- the linear prediction coefficients are output to the weighting filter 140 and the weighting filter 141.
- the index corresponding to the quantized linear prediction coefficient is output to a linear prediction synthesis filter 130, a linear prediction synthesis filter 131, and a sign output circuit 190.
- the index output from the first minimizing circuit 150 is input to the first sound source generating circuit 110.
- the first sound source generation circuit 110 reads a first sound source vector corresponding to the index from a table in which a plurality of sound source vectors are stored, and reads the first sound source vector into the first sound source vector. Output to the gain circuit 16 0.
- the index output from the second minimization circuit 151 is input to the second sound source generation circuit 111.
- the second sound source vector corresponding to the index is read from a table in which a plurality of sound source vectors are stored, and is output to the second gain circuit 161.
- the index output from the first minimizing circuit 150 and the first sound source vector output from the first sound source generating circuit 110 are input to the first gain circuit 160. It is.
- the first gain circuit 160 reads a first gain corresponding to the index from a table in which a plurality of gain values are stored. Thereafter, the first gain circuit 160 multiplies the first gain by the first sound source vector to generate a third sound source vector, and the third sound source vector Output to the first bandpass filter 120.
- the index output from the second minimizing circuit 151 and the second sound source vector output from the second sound source generating circuit 111 are input to the second gain circuit 161.
- the second gain circuit 161 reads out a second gain corresponding to the index from a table in which a plurality of gain values are stored. Thereafter, the second gain circuit 161 multiplies the second gain by the second sound source vector to generate a fourth sound source vector, and generates the fourth sound source vector.
- a second bandpass filter 1 2 Output to 1.
- the third sound source vector output from the first gain circuit 160 is input.
- the third sound source vector is band-limited to a first band by this filter, and generates a first excitation vector.
- the first bandpass filter 120 outputs the first excitation vector to the linear prediction synthesis filter 130.
- the second band-pass filter 122 receives a fourth sound source vector output from the second gain circuit 161.
- the fourth sound source vector is band-limited to the second band by this filter, and generates a second excitation vector.
- the second bandpass filter 12 21 outputs the second excitation vector to the linear prediction synthesis filter 13 1.
- the linear prediction synthesis filter 130 corresponds to the first excitation vector output from the first bandpass filter 120 and the quantized linear prediction coefficient output from the linear prediction coefficient calculation circuit 170.
- the input data to be input is input.
- the linear prediction synthesis filter 130 reads out a quantized linear prediction coefficient corresponding to the index from a table in which a plurality of quantized linear prediction coefficients are stored.
- a first reproduction signal (reproduction vector) is generated by driving the filter in which the quantized linear prediction coefficient is set by the first excitation vector.
- the first regeneration vector is output to a first differentiator 180.
- the linear prediction synthesis filter 13 1 corresponds to the second excitation vector output from the second band-pass filter 12 1 and the quantized linear prediction coefficient output from the linear prediction coefficient calculation circuit 17 0 Is input.
- the linear prediction synthesis filter 13 1 reads out a quantized linear prediction coefficient corresponding to the index from a table in which a plurality of quantized linear prediction coefficients are stored. By driving the filter in which the quantized linear prediction coefficient is set by the second excitation vector, a second reproduction vector is generated. The second reproduction vector is output to a second differentiator 18 1.
- the input vector is input to the first differentiator 180 through the input terminal 10, and the first reproduction vector output from the linear prediction synthesis filter 130 is input. No.
- the one differentiator 180 calculates the difference between the input vector and the first reproduction vector. This difference is output to the weighting filter 140 and the second differentiator 181, as a first difference vector.
- the second differencer 18 1 receives the first difference vector from the first differencer 180 and the second regeneration vector output from the linear prediction synthesis filter 13 1. Is done.
- the second differentiator 18 1 calculates a difference between the first difference vector and the second reproduction vector. This difference is output to the weighting filter 141 as a second difference vector.
- the first difference vector output from the first differentiator 180 and the linear prediction coefficient output from the linear prediction coefficient calculation circuit 170 are input to the weighting filter 140.
- the weighting filter 140 generates a weighting filter corresponding to human auditory characteristics using the linear prediction coefficient, and drives the weighting filter with the first difference vector.
- the above operation of the weighting filter 140 generates a first weighted difference vector.
- the first weighted difference vector is output to a first minimizing circuit 150.
- the second difference vector output from the second differentiator 18 1 and the linear prediction coefficient output from the linear prediction coefficient calculation circuit 170 are input to the weighting filter 14 1.
- the weighting filter 1441 generates a weighting filter corresponding to human auditory characteristics using the linear prediction coefficient, and drives the weighting filter with the second difference vector.
- the above operation of the weighting filter 141 generates a second weighting difference vector.
- the second weighted difference vector is output to a second minimizing circuit 151.
- the first minimizing circuit 150 sequentially outputs indices corresponding to all the first sound source vectors stored in the first sound source generating circuit 110 to the first sound source generating circuit 110. Then, the indices corresponding to all the first gains stored in the first gain circuit 160 are sequentially output to the first gain circuit 160. Further, the first weighting difference vector output from the weighting filter 140 is sequentially input to the first minimizing circuit 150. The first minimization circuit 150 calculates its norm. The first minimizing circuit 150 is configured so that the first norm is minimized. And the first gain is selected, and an index corresponding to these is output to the code output circuit 190.
- the second minimizing circuit 15 1 sequentially outputs indices corresponding to all of the second sound source vectors stored in the second sound source generating circuit 1 11 to the second sound source generating circuit 1 1 1. Then, the indices corresponding to all of the second gains stored in the second gain circuit 16 1 are sequentially output to the second gain circuit 16 1. In addition, the second weighting difference vector output from the weighting filter 141 is sequentially input to the second minimizing circuit 151.
- the second minimizing circuit 15 1 calculates its norm.
- the second gain circuit 161 selects the second sound source vector and the second gain so that the norm is minimized, and outputs an index corresponding to these to a code output circuit 19 ⁇ . Output to
- An index corresponding to the quantized linear prediction coefficient output from the linear prediction coefficient calculation circuit 170 is input to the code output circuit 190, and the index is output from the first minimization circuit 150 and the first An index corresponding to each of the sound source vector and the first gain is input, output from the second minimization circuit 151, and corresponding to each of the second sound source vector and the second gain.
- the index is entered.
- the code output circuit 190 converts each index into a code of a bit sequence, and outputs the converted index via the output terminal 20.
- FIG. 2 is a block diagram showing an example of a conventional audio / music signal decoding device.
- the code of the bit sequence from the input terminal 30 is input to the code input circuit 310.
- the code input circuit 310 converts the code of the bit sequence input from the input terminal 30 into an index.
- the index corresponding to the first sound source vector is output to first sound source generation circuit 110.
- the index corresponding to the second sound source vector is output to the second sound source generation circuit 111.
- the index corresponding to the first gain is output to first gain circuit 160.
- the index corresponding to the second gain is output to the second gain circuit 161.
- the index corresponding to the quantized linear prediction coefficient is output to the linear prediction synthesis filter 130 and the linear prediction synthesis filter 131.
- the first sound source generation circuit 110 has an index output from the code input circuit 310. Box is entered.
- the first sound source generation circuit 110 reads a first sound source vector corresponding to the index from a table storing a plurality of sound source vectors, and a first gain circuit 16 Output to 0.
- the index output from the code input circuit 310 is input to the second sound source generation circuit 111.
- the second sound source generation circuit 111 reads a second sound source vector corresponding to the index from a table in which a plurality of sound source vectors are stored, and sends the second sound source vector to a second gain circuit 161. Output.
- the index output from the code input circuit 310 and the first sound source vector output from the first sound source generation circuit 110 are input to the first gain circuit 160.
- the first gain circuit 160 reads a first gain corresponding to the index from a table storing a plurality of gain values.
- the first gain circuit 160 multiplies the first gain by the first sound source vector to generate a third sound source vector.
- the third sound source vector is output to the first band pass filter 120.
- the second gain circuit 161 To the second gain circuit 161, the index output from the code input circuit 31 ⁇ and the second sound source vector output from the second sound source generation circuit 111 are input.
- the second gain circuit 161 reads a second gain corresponding to the index from a table in which a plurality of gain values are stored. Thereafter, the second gain circuit 161 multiplies the second gain by the second sound source vector to generate a fourth sound source vector.
- the fourth sound source vector is output to the second band pass filter 122.
- the third sound source vector output from the first gain circuit 160 is input.
- the third sound source vector is band-limited to the first band by the filter, and generates a first excitation vector.
- the first bandpass filter 120 outputs the first excitation vector to the linear prediction synthesis filter 130.
- the second bandpass filter 1 2 1 receives the fourth sound source vector output from the second gain circuit 16 1. Since the fourth sound source vector is band-limited to the second band by this filter, the second band-pass filter 1 2 1 Generate the excitation vector. The second band-pass filter 12 1 outputs the second excitation vector to the linear prediction synthesis filter 13 1.
- the linear prediction synthesis filter 130 corresponds to the first excitation vector output from the first bandpass filter 120 and the quantized linear prediction coefficient output from the code input circuit 310.
- the index is entered.
- the quantized linear prediction coefficient corresponding to the above-mentioned index is read from a table in which a plurality of quantized linear prediction coefficients are stored. Thereafter, the linear prediction synthesis filter 130 generates a first reproduction vector by driving the filter in which the quantized linear prediction coefficient is set by the first excitation vector.
- the first reproduction vector is output to the adder 182.
- the linear prediction synthesis filter 1 3 1 has a second excitation vector output from the second band-pass filter 1 2 1 and an index corresponding to the quantized linear prediction coefficient output from the code input circuit 3 10. Is entered.
- the quantized linear prediction coefficient corresponding to the above-mentioned index is read from a table in which a plurality of quantized linear prediction coefficients are stored.
- the linear prediction synthesis filter 1331 generates a second reproduction vector by driving the filter in which the quantized linear prediction coefficient is set by the second excitation vector.
- the second reproduction vector is output to the adder 182.
- the first reproduction vector output from the linear prediction synthesis filter 130 and the second reproduction vector output from the linear prediction synthesis filter 131 are input to the adder 182. You. Calculate the sum of these.
- the adder 182 outputs the sum of the first reproduction vector and the second reproduction vector as a third reproduction vector via an output terminal 40.
- the signal is obtained by adding an excitation signal having a band characteristic corresponding to a low band of an input signal and an excitation signal having a band characteristic corresponding to a high band of the input signal. Since the reproduced signal is generated by driving the linear prediction synthesis filter obtained from the input signal using the excitation signal, the CELP-based coding is performed in the band belonging to the high frequency band. The coding performance of the audio music signal in the entire band is degraded due to the decrease in the coding performance in the band belonging to.
- CELP which models the speech generation process, cannot generate signals in the high frequency band with high accuracy.
- An object of the present invention is to solve the above-mentioned problems and to provide an audio / music signal encoding device capable of encoding an audio / music signal satisfactorily over the entire band. Disclosure of the invention
- the audio / music signal encoding apparatus drives a linear prediction synthesis filter obtained from an input signal by an excitation signal corresponding to a first band to convert a first reproduced signal. And generating a residual signal by driving an inverse filter of the linear predictive synthesis filter based on a difference signal between the input signal and the first reproduced signal, the residual signal corresponding to a second band in the residual signal. Is encoded after orthogonal transform.
- the device of the present invention 1 includes a means for generating a first reproduced signal by driving the linear prediction synthesis filter with an excitation signal corresponding to a first band (110 in FIG. 3).
- the audio / music signal encoding apparatus of the present invention drives the linear predictive synthesis filter obtained from the input signal by using the excitation signals corresponding to the first and second bands, thereby obtaining the first signal. And a second reproduction signal, and a residual signal is generated by driving an inverse filter of the linear prediction synthesis filter with a difference signal between the signal obtained by adding the first and second reproduction signals and the input signal. Then, a component corresponding to a third band in the residual signal is encoded after orthogonal transform.
- the device includes a step of generating the first and second reproduced signals by driving the linear prediction synthesis filter with the excitation signals corresponding to the first and second bands. 1001 in FIG. 10, 1002), and driving an inverse filter of the linear prediction synthesis filter by a difference signal between the signal obtained by adding the first and second reproduced signals and the input signal. Generates a residual signal, and corresponds to a third band in the residual signal. (1003 in FIG. 10) for encoding the component after orthogonal transformation.
- the audio / music signal encoding apparatus of the present invention drives a linear predictive synthesis filter obtained from an input signal with an excitation signal corresponding to the first to (N ⁇ 1) th bands.
- the first to (N ⁇ 1) th reproduced signals are generated, and the linear prediction synthesis filter is generated by a difference signal between the signal obtained by adding the first to (N ⁇ 1) th reproduced signals and the input signal.
- a residual signal is generated by driving the inverse filter of, and a component corresponding to the N-th band in the residual signal is coded after orthogonal transform.
- the device of the present invention 3 performs the first to (N ⁇ 1) th reproduction by driving the linear prediction synthesis filter with an excitation signal corresponding to the first to (N ⁇ 1) th bands.
- the audio / music signal encoding apparatus of the present invention uses a linear signal obtained from an input signal by a difference signal between the first encoded signal and the input signal in the second encoding.
- a residual signal is generated by driving an inverse filter of the prediction synthesis filter, and a component corresponding to an arbitrary band in the residual signal is encoded after orthogonal transform.
- the device of the present invention 4 comprises means for calculating the difference between the first encoded signal and the input signal (180 in FIG. 13), and a linear prediction synthesis filter obtained from the input signal.
- the audio / music signal encoding apparatus includes a third encoding unit that converts a signal obtained by adding the first and second encoded decoded signals into a difference signal between the input signal and the input signal. Accordingly, a residual signal is generated by driving an inverse filter of the linear prediction synthesis filter obtained from the input signal, and a component corresponding to an arbitrary band in the residual signal is encoded after orthogonal transform.
- the device of the present invention 5 is a signal obtained by adding the first and second encoded / decoded signals.
- the audio / music signal encoding apparatus of the present invention (the apparatus of the present invention 6) is configured such that, in the N-th encoding, the difference between the signal obtained by adding the first to the --- 1 encoded and decoded signals and the input signal is obtained.
- a residual signal is generated by driving an inverse filter of the linear prediction synthesis filter obtained from the input signal using the signal, and a component corresponding to an arbitrary band in the residual signal is encoded after orthogonal transform.
- the device of the present invention 6 comprises means for calculating a difference signal between the signal obtained by adding the first to (N ⁇ 1) th encoded and decoded signals and the input signal (18 in FIG. 15). 0, 1 8 0 2), and a residual signal is generated by driving an inverse filter of the linear prediction synthesis filter obtained from the input signal with the differential signal, and corresponds to an arbitrary band in the residual signal.
- the audio / music signal encoding apparatus of the present invention uses a pitch prediction filter when generating an excitation signal corresponding to the first band of the input signal.
- the device of the present invention 7 has pitch predicting means (112, 162, 184, 510 in FIG. 16).
- the audio / music signal encoding apparatus of the present invention performs down-sampling of a first input signal sampled at a first sampling frequency to a second sampling frequency to generate a second input signal.
- An input signal is generated, and a synthesis filter in which a first linear prediction coefficient obtained from the second input signal is set is driven by an excitation signal to generate a first reproduction signal, and the first reproduction is performed.
- a second reproduced signal is generated by up-sampling the signal to the first sampling frequency, and the linear prediction coefficient obtained from the first input signal and the first linear prediction coefficient are converted to a first sample.
- a third linear prediction coefficient is calculated from a difference between the second linear prediction coefficient and the second linear prediction coefficient obtained by converting the sampling frequency to the ring frequency, and a third linear prediction coefficient is calculated from the sum of the second linear prediction coefficient and the third linear prediction coefficient. And calculating a fourth linear prediction coefficient according to a difference signal between the first input signal and the second reproduced signal.
- a residual signal is generated by driving the data, and a component corresponding to an arbitrary band in the residual signal is encoded after orthogonal transform.
- the device includes means for generating a second input signal by down-sampling a first input signal sampled at a first sampling frequency to a second sampling frequency. (780 in FIG. 7) and means for generating a first reproduced signal by driving a synthesis filter in which a first linear prediction coefficient determined from the second input signal is set by an excitation signal (a second signal). Means for generating a second reproduced signal by up-sampling the first reproduced signal to the first sampling frequency (770, 132 in FIG. 17) (781 in FIG.
- the third linear prediction coefficient is obtained from the difference between the linear prediction coefficient obtained from the input signal of the first and second linear prediction coefficients obtained by converting the first linear prediction coefficient and the sampling frequency to the first sampling frequency.
- the audio / music signal decoding apparatus of the present invention (the apparatus of the ninth invention) generates an excitation signal corresponding to the second band by orthogonally and inversely transforming the decoded orthogonal transform coefficient.
- a second reproduction signal is generated by driving a linear prediction synthesis filter, and a first reproduction signal is generated by driving the linear prediction filter with an excitation signal corresponding to the decoded first band.
- a decoded voice music is generated by adding the first reproduced signal and the second reproduced signal.
- the device includes means (440 and 460 in FIG. 18) for generating an excitation signal corresponding to the second band by performing orthogonal inverse transform on the decoded signal and the orthogonal transform coefficient.
- Means for generating a second reproduced signal by driving a linear prediction synthesis filter with the excitation signal (131 in FIG. 18); and driving the linear prediction filter with an excitation signal corresponding to the first band.
- the audio / music signal decoding device of the present invention (the device of the present invention 10) generates an excitation signal corresponding to a third band by orthogonally and inversely transforming the decoded orthogonal transformation coefficient.
- a second reproduced signal are generated, and a decoded voice music is generated by adding the first to third reproduced signals.
- the device of the present invention 10 generates an excitation signal corresponding to the third band by performing orthogonal inverse transform on the decoded orthogonal transform coefficient, and performs a linear prediction synthesis filter with the excitation signal.
- Means for generating a third reproduced signal by driving (1053 in FIG. 24); and driving the linear prediction filter with an excitation signal corresponding to the first and second bands to obtain the first signal.
- the audio / music signal decoding device (the device of the present invention 11) of the present invention generates an excitation signal corresponding to the N-th band by orthogonally and inversely transforming the decoded orthogonal transform coefficient.
- To generate a decoded speech music by adding the first to N-th reproduced signals from the above.
- the device of the present invention 11 generates an excitation signal corresponding to the N-th band by orthogonally inversely transforming the decoded orthogonal transform coefficient, and performs a linear prediction synthesis filter with the excitation signal.
- the audio / music signal decoding apparatus of the present invention performs the second decoding. Then, an excitation signal is generated by orthogonally inverse-transforming the decoded orthogonal transform coefficient, and a reproduced signal is generated by driving a linear prediction synthesis filter with the excitation signal, and the reproduced signal and the first decoded signal are Is added to generate decoded voice music.
- the device of the twelfth aspect of the present invention generates an excitation signal by orthogonally and inversely transforming the decoded orthogonal transform coefficient, and generates a reproduced signal by driving a linear prediction synthesis filter with the excitation signal. Means (1052 in FIG. 26) and means (182 in FIG. 26) for generating decoded voice music by adding the reproduction signal and the first decoded signal.
- the audio / music signal decoding device of the present invention performs an orthogonal inverse transform on the decoded orthogonal transform coefficient in the third decoding to generate an excitation signal, and generates a linear prediction synthesis filter. Is driven by the excitation signal to generate a reproduced signal, and the reproduced signal is added to the first and second decoded signals to generate decoded voice music.
- the device of the thirteenth invention generates an excitation signal by performing orthogonal inverse transform on the decoded orthogonal transform coefficient, and generates a reproduction signal by driving a linear prediction synthesis filter with the excitation signal.
- the audio / music signal decoding apparatus (the apparatus according to the present invention 14) performs an orthogonal inverse transform of the decoded orthogonal transform coefficients in the N-th decoding to generate an excitation signal, and generates a linear prediction synthesis filter. Is driven by the excitation signal to generate a reproduced signal, and the reproduced signal is added to the first to (N ⁇ 1) th decoded signals to generate decoded voice music.
- the apparatus of the present invention 14 generates an excitation signal by performing orthogonal inverse transform on the decoded orthogonal transform coefficient, and generates a reproduction signal by driving a linear prediction synthesis filter with the excitation signal.
- the audio / music signal decoding device of the present invention (the device of the 15th present invention) A pitch prediction filter is used in generating a corresponding excitation signal.
- the device of the present invention 15 further includes a pitch prediction means (112, 162, 184, 510 in FIG. 29).
- the audio / music signal decoding device of the present invention converts a signal obtained by driving a first linear predictive synthesis filter with a first excitation signal for a first band into a first signal.
- the second excitation signal corresponding to the second band is generated by up-sampling to a sampling frequency of 1, generating a first reproduced signal, and performing orthogonal inverse transform on the decoded orthogonal transform coefficient.
- the second reproduced signal is generated by driving the second linear prediction synthesis filter with the excitation signal of the above, and the decoded voice music is generated by adding the first reproduced signal and the second reproduced signal.
- the device of the present invention 16 converts a signal obtained by driving a first linear prediction synthesis filter with a first excitation signal corresponding to a first band to a first sampling frequency.
- the first reproduced signal 13, 78 1 in FIG. 30
- orthogonal inverse transform of the decoded orthogonal transform coefficients the second band corresponding to the second band is obtained.
- Means for generating a second reproduction signal by generating a second excitation signal and driving a second linear predictive synthesis filter with the second excitation signal (440, 831, FIG. 30)
- means for generating decoded voice music by adding the first reproduced signal and the second reproduced signal (182 in FIG. 30).
- the audio / music signal encoding / decoding device of the present invention decodes the code output from the device of the present invention 1 with the device of the present invention 9.
- the device of the present invention 17 includes audio / music signal encoding means (FIG. 3) and audio / music signal decoding means (FIG. 18).
- the audio / music signal encoding / decoding device of the present invention decodes the code output from the device of the present invention 2 by the device of the present invention 10.
- the device of the present invention 18 has a voice / music signal encoding means (FIG. 10) and a voice / music signal decoding means (FIG. 24).
- the audio / music signal encoding / decoding device of the present invention decodes the code output from the device of the present invention 3 with the device of the present invention 11.
- the apparatus of the present invention 19 comprises a sound / music signal encoding means (FIG. 11) and a sound / music signal decoding means. (FIG. 25).
- the audio / music signal encoding / decoding device of the present invention decodes the code output from the device of the present invention 4 with the device of the present invention 12.
- the device of the present invention 20 includes audio / music signal encoding means (FIG. 13) and audio / music signal decoding means (FIG. 26).
- the audio / music signal encoding / decoding device of the present invention decodes the code output from the device of the present invention 5 by the device of the present invention 13.
- the device of the present invention 21 includes audio / music signal encoding means (FIG. 14) and audio / music signal decoding means (FIG. 27).
- the audio / music signal encoding / decoding device of the present invention decodes the code output from the device of the present invention 6 with the device of the present invention 14.
- the device of the present invention 22 includes audio / music signal encoding means (FIG. 15) and audio / music signal decoding means (FIG. 28).
- the audio / music signal encoding / decoding device of the present invention decodes the code output from the device of the present invention 7 with the device of the present invention 15.
- the device of the present invention 23 has a voice / music signal encoding means (FIG. 16) and a voice / music signal decoding means (FIG. 29).
- the audio / music signal encoding / decoding device of the present invention decodes the code output from the device of the present invention 8 by the device of the present invention 16.
- the device of the present invention 24 includes audio / music signal encoding means (FIG. 17) and audio / music signal decoding means (FIG. 30).
- a first reproduction signal is generated by driving a linear prediction synthesis film obtained from an input signal by an excitation signal having a band characteristic corresponding to a low band of the input signal, and the input signal and the first signal are generated.
- a residual signal is generated by driving an inverse filter of the linear prediction synthesis filter with a difference signal from a reproduced signal, and a high-frequency component of the residual signal is encoded using an encoding method based on orthogonal transform.
- encoding based on orthogonal transform is performed instead of CELP. Coding based on the orthogonal transform has higher coding performance for signals having properties different from speech than CELP. For this reason, The coding performance for the high frequency component of the input signal is improved. As a result, the audio / music signal can be satisfactorily encoded over the entire band.
- FIG. 1 is a block diagram showing an embodiment of a conventional audio / music signal encoding apparatus.
- FIG. 2 is a block diagram showing an embodiment of a conventional audio / music signal decoding apparatus.
- FIG. 3 is a block diagram showing the configuration of the audio and music signal encoding device according to the first embodiment of the present invention.
- FIG. 4 is a block diagram showing a configuration of the first sound source generation circuit 110.
- FIG. 5 is a diagram for explaining a method of generating a sub-vector in the band selection circuit 250.
- FIG. 6 is a block diagram showing a configuration of the orthogonal transform coefficient quantization circuit 260.
- FIG. 7 is a block diagram equivalent to FIG. 3, showing the configuration of the audio / music signal encoding apparatus according to the first embodiment of the present invention.
- FIG. 8 is a block diagram showing the configuration of the first encoding circuit 1 • 01 1 in FIG.
- FIG. 9 is a block diagram showing a configuration of the second encoding circuit 1002 in FIG.
- FIG. 10 is a block diagram showing a configuration of a speech and music signal encoding device according to a second embodiment of the present invention.
- FIG. 11 is a block diagram showing a configuration of a speech and music signal encoding device according to a third embodiment of the present invention.
- FIG. 12 is a block diagram showing the configuration of the first encoding circuit 101 in FIG.
- FIG. 13 is a block diagram showing a configuration of a voice and music signal encoding device according to a fourth embodiment of the present invention.
- FIG. 14 shows the configuration of an audio / music signal encoding apparatus according to a fifth embodiment of the present invention. It is a block diagram.
- FIG. 15 is a block diagram showing a configuration of a speech and music signal encoding device according to a sixth embodiment of the present invention.
- FIG. 16 is a block diagram showing a configuration of a speech and music signal encoding device according to a seventh embodiment of the present invention.
- FIG. 17 is a block diagram showing a configuration of an audio / music signal encoding apparatus according to an eighth embodiment of the present invention.
- FIG. 18 is a block diagram showing the configuration of the audio and music signal decoding device according to the ninth embodiment of the present invention.
- FIG. 19 is a diagram for explaining a method of generating a second excitation vector in the orthogonal transform coefficient inverse quantization circuit 460.
- FIG. 20 is a block diagram showing the configuration of the orthogonal transform coefficient inverse quantization circuit 460.
- FIG. 21 is a block diagram equivalent to FIG. 36, showing the configuration of the audio / music signal decoding apparatus according to the ninth embodiment of the present invention.
- FIG. 22 is a block diagram showing a configuration of the first decoding circuit 1051 in FIG.
- FIG. 23 is a block diagram showing a configuration of the second decoding circuit 1052 in FIG.
- FIG. 24 is a block diagram showing the configuration of the audio / music signal decoding apparatus according to the tenth embodiment of the present invention.
- 25th It is a block diagram showing the configuration of the audio and music signal decoding device according to the eleventh embodiment of the present invention.
- FIG. 26 is a block diagram showing the configuration of the audio / music signal decoding apparatus according to the 12th embodiment of the present invention.
- FIG. 27 is a block diagram showing the configuration of the audio and music signal decoding device according to the thirteenth embodiment of the present invention.
- FIG. 28 is a block diagram showing the configuration of the audio / music signal decoding apparatus according to the 14th embodiment of the present invention.
- FIG. 29 shows the configuration of the audio / music signal decoding apparatus according to the fifteenth embodiment of the present invention. It is a block diagram.
- FIG. 30 is a block diagram showing the configuration of the audio and music signal decoding device according to the sixteenth embodiment of the present invention.
- FIG. 31 is a diagram for explaining the correspondence between the index and the code of the bit sequence in the code output circuit 290.
- FIG. 32 is a diagram for explaining a method of generating the first pitch vector in the pitch signal generation circuit 112. BEST MODE FOR CARRYING OUT THE INVENTION
- FIG. 3 is a block diagram showing the configuration of the audio / music signal encoding device according to the first embodiment of the present invention.
- An input signal (input vector), which is obtained by sampling a voice or music signal and combining the plurality of samples into one frame to generate one vector, is input from the input terminal 10.
- L is the vector length.
- LSP line spectrum pair
- the first sound source generation circuit 110 receives the index output from the first minimization circuit 150.
- the first sound source vector corresponding to the index is read from a table in which a plurality of sound source signals (sound source vectors) are stored, and is output to the first gain circuit 160.
- the table 110 included in the first sound source generation circuit 110 stores Ne sound source vectors. For example, Ne is 2 56.
- the index i output from the first minimizing circuit 150 is input to the switch 1102 via the input terminal 1103.
- the switch 1 1 2 selects a sound source vector corresponding to the index from the table, and uses this sound source vector as a first sound source vector, via an output terminal 1104, a first gain. Output to circuit 160.
- a method of efficiently expressing the excitation signal using a multi-pulse signal composed of a plurality of pulses and defined by the pulse position and the pulse amplitude can be used.
- Ozawa et al. “MF-CELP speech coding based on multi-pulse vector quantization sound source and high-speed search,” regarding encoding of source signals using multi-pulse signals (IEICE Transactions A, pp. 1655). -1663, 199 6) (Reference 5). This concludes the description of the first sound source generation circuit 110.
- the first gain circuit 160 has a table in which gain values are stored.
- the index output from the first minimizing circuit 150 and the first sound source vector output from the first sound source generating circuit 110 are input to the first gain circuit 160.
- a first gain corresponding to the index is read from the table, and the first gain is multiplied by the first sound source vector to generate a second sound source vector.
- the generated second sound source vector is output to the first band pass filter 120.
- the first band pass filter 120 outputs to the first gain circuit 160
- the second sound vector is input.
- the second sound source vector is band-limited to a first band by this filter to obtain a first excitation vector.
- the first band-pass filter 120 outputs the first excitation vector to the linear prediction synthesis filter 13 °.
- the first band is from Fsl [Hz] to Fel [Hz :.
- the first band-pass filter 120 has a characteristic of limiting the band to the first band, and has a linear prediction order of about 100th order. (z) can also be realized.
- the linear prediction synthesis filter 130 includes a table in which the quantized linear prediction coefficients are stored.
- the linear prediction synthesis filter 130 includes a first excitation vector output from the first bandpass filter 120 and an index corresponding to the quantized linear prediction coefficient output from the linear prediction coefficient calculation circuit 170. Is entered. Further, the linear prediction synthesis filter 130 reads out the quantized linear prediction coefficient corresponding to the index from the table.
- a first reproduction signal (reproduction vector) is generated.
- the first reproduction vector is output to a first differentiator 180.
- the transfer function 1 / A (z) of the synthesis filter is expressed by the following equation (2).
- Equation 2 The input vector and the first reproduction vector output from the linear prediction synthesis filter 130 are input to the first differentiator 180 via the input terminal 10.
- the first difference unit 180 calculates the difference, and outputs the difference value as a first difference vector to the weighting filter 140 and the inverse linear prediction filter 230.
- the first difference vector output from the first differentiator 180 and the linear prediction coefficient output from the linear prediction coefficient calculation circuit 170 are input to the first weighting filter 140.
- the first weighting filter 140 generates a weighting filter W (z) corresponding to human auditory characteristics using the linear prediction coefficients, and calculates the weighting filter by the first difference vector. Drive. As a result, a first weighted difference vector is obtained. And the first weighted difference vector is a first minimizing circuit
- the first minimizing circuit 150 sequentially outputs all the first sound source vectors stored in the first sound source generating circuit 110: the corresponding indexes to the first sound source generating circuit 110, The indices corresponding to all the first gains stored in the first gain circuit 160 are sequentially output to the first gain circuit 160. Further, the first minimizing circuit 150 receives the first weighted difference vector sequentially output from the weighting filter 140, calculates the norm thereof, and calculates the norm such that the norm is minimized. One source vector and the first gain are selected, and the corresponding index is output to the code output circuit 290.
- the linear prediction inverse filter 230 has a table in which the quantized linear prediction coefficients are stored. An index corresponding to the quantized linear prediction coefficient output from the linear prediction coefficient calculation circuit 170 and the first difference The first output difference vector is input. Further, the inverse linear prediction filter 230 reads out the quantized linear prediction coefficient corresponding to the index from the table. By driving the inverse filter A (z) in which the quantized linear prediction coefficients are set by the first difference vector, a first residual vector is obtained. Then, the first residual vector is output to the orthogonal transform circuit 240.
- the transfer function A (z) of the inverse filter is represented by the following Equation 4.
- the first residual vector output from the inverse linear prediction filter 230 is input to the orthogonal transform circuit 240.
- the orthogonal transform circuit 240 performs an orthogonal transform on the first residual vector to generate a second residual vector.
- the second residual vector is output to the band selection circuit 250.
- DCT Discrete Cosine Transform
- the second residual vector output from the orthogonal transformation circuit 240 is input to the band selection circuit 250.
- Nsbv sub-vectors are generated using the components included in the second band.
- An arbitrary band can be set as the second band.
- the band from Fs2 [Hz] to Fe2 [Hz] is used.
- FsO Fs2 ⁇ Fe2 ⁇ FeO is used.
- the first band and the second band do not overlap, that is, Fel ⁇ Fs2.
- the band selection circuit 250 outputs the Nsbv sub-vectors to the orthogonal transform coefficient quantization circuit 260.
- Nsbv sub-vectors output from the band selection circuit 250 are input to the orthogonal transform coefficient quantization circuit 260.
- the orthogonal transform coefficient quantization circuit 260 stores a table in which a quantization value (shape code vector) for the shape of the sub-vector is stored, and a quantization value (quantization gain) for the gain of the sub-vector. It has a table. For each of the input Nsbv sub-vectors, the quantization error is minimized.
- the orthogonal transform coefficient quantization circuit 260 calculates the quantized value of the shape and The quantized value of the input is selected from the table, and the corresponding index is output to the code output circuit 290.
- the configuration of the orthogonal transform coefficient quantization circuit 260 will be supplemented with reference to FIG.
- Fig. 4 there are Nsbv blocks surrounded by dotted lines.
- the Nsbv sub-vectors are quantized in each block.
- the Nsbv sub-vectors are represented as in Equation 5 below.
- Equation 5 esb, o ( n ), ⁇ e sb , N sbv -l (n), n 2 0,..., L— 1
- L represents the vector length
- j represents the index.
- k represents an index.
- the quantization gain gO [k] corresponding to the index is read from the table.
- the difference vector 2640 is connected to the sub-vector e sb, 0 input through the input terminal 2650.
- the difference vector is sequentially input from the difference unit 2640, and its norm DO is calculated.
- the indexes corresponding to these are output to the index output circuit 2660. Similar processing is performed for the sub-vectors shown in Equation 6 below.
- the DO can be modified as shown in Equation 8 or Equation 9 below.
- An index corresponding to the quantized linear prediction coefficient output from the linear prediction coefficient calculation circuit 170 is input to the code output circuit 290.
- the code output circuit 290 receives an index output from the first minimizing circuit 150 and an index corresponding to each of the first excitation vector and the first gain.
- the code output circuit 290 receives a set of indices output from the orthogonal transform coefficient quantization circuit 260 and composed of an index of shape vectors and quantization gains for Nsbv subvectors. Is done. Then, as schematically shown in FIG. 31, each index is converted into a bit sequence code and output via the output terminal 20.
- the first embodiment described with reference to FIG. 3 is for the case where the number of bands is 2, but the case where the number of bands is extended to 3 or more will be described below.
- Figure 3 can be rewritten as in Figure 7.
- the first encoding circuit 1001 in FIG. 7 is equivalent to FIG.
- the second encoding circuit 1002 in FIG. 7 is equivalent to FIG.
- the blocks making up FIGS. 8 and 9 are the same as the blocks described in FIG.
- the second embodiment of the present invention is to extend the number of bands to 3 in the first embodiment. Is achieved.
- the configuration of the audio / music signal encoding apparatus according to the second embodiment of the present invention can be represented by a block diagram shown in FIG.
- the first encoding circuit 1001 is equivalent to FIG. 8
- the second encoding circuit 1002 is equivalent to FIG. 8
- the third encoding circuit 100 3 is equivalent to FIG.
- the index output from the linear prediction coefficient calculation circuit 170 is input to the code output circuit 2901
- the index output from the first coding circuit 1001 is input to the code output circuit 2901
- the second The index output from the encoding circuit 1002 is input
- the index set output from the third encoding circuit 103 is input.
- Each index is converted into a bit sequence code and output via an output terminal 20.
- the third embodiment of the present invention is realized by extending the number of bands to N in the first embodiment.
- the configuration of the audio / music signal encoding apparatus according to the third embodiment of the present invention can be represented by a block diagram shown in FIG.
- the first encoding circuit 1001 to the (N ⁇ 1) th encoding circuit 1004 are equivalent to FIG.
- the N-th encoding circuit 1005 is equivalent to FIG.
- the index output from the linear prediction coefficient calculation circuit 170 is input to the code output circuit 290 02, and the first (N— 1) th coding circuit 10 0 to the first coding circuit 100 1
- the index output from each of N.04 and N.04 is input, and the set of indices output from the N-th encoding circuit 1005 is input. Then, each index is converted into a bit sequence code and output via the output terminal 20.
- the first encoding circuit 1001 in FIG. 7 is based on an encoding system using an A-b-S (Analysis-by-Synthesis) method.
- an encoding method other than the A_b—S method can be applied to the first encoding circuit 1001.
- an encoding method using time frequency conversion is applied to first encoding circuit 1001, as an encoding method other than the A—b_S method.
- the fourth embodiment of the present invention is realized by applying an encoding method using time-frequency conversion in the first embodiment.
- the configuration of the audio / music signal encoding apparatus according to the fourth embodiment of the present invention can be represented by a block diagram shown in FIG.
- the first encoding circuit 101 is equivalent to FIG.
- Second encoding circuit 1 0 0 2 Is equivalent to FIG.
- the linear prediction inverse filter 230, the orthogonal transform circuit 240, the band selection circuit 250, and the orthogonal transform coefficient quantization circuit 260 are described in FIG. It is the same as each block.
- the orthogonal transform coefficient inverse quantization circuit 460 the orthogonal inverse transform circuit 440, and the linear predictive synthesis filter 13 1 are used in a ninth embodiment, which will be described later, to perform speech and music decoding corresponding to the first embodiment. It is the same as the block that constitutes the device.
- the code output circuit 290 3 receives the index output from the linear prediction coefficient calculation circuit 17 2, receives the index set output from the first coding circuit 101 1, A set of indexes output from the second encoding circuit 1002 is input. Then, each index is converted into a bit-sequence code and output via an output terminal 20.
- the fifth embodiment of the present invention is realized by extending the number of bands to three in the fourth embodiment.
- the configuration of the audio and music signal encoding apparatus according to the fifth embodiment of the present invention can be represented by a block diagram shown in FIG.
- the first encoding circuit 101 1 is equivalent to FIG. 12
- the second encoding circuit 110 12 is equivalent to FIG. 12
- the third encoding circuit 1 03 is equivalent to FIG.
- An index output from the linear prediction coefficient calculation circuit 170 is input to the code output circuit 290 4
- a set of indexes output from the first coding circuit 110 1 is input to the code output circuit 290 4.
- a set of indices output from the second encoding circuit 101 is input, and a set of indices output from the third encoding circuit 103 is input.
- Each index is converted into a bit sequence code and output via the output terminal 20.
- the sixth embodiment of the present invention is realized by extending the number of bands to N in the fourth embodiment.
- the configuration of the audio and music signal encoding apparatus according to the sixth embodiment of the present invention can be represented by a block diagram shown in FIG.
- each of the first encoding circuit 101 1 to the N ⁇ 1th encoding circuit 110 4 is equivalent to FIG.
- the N-th encoding circuit 1005 is equivalent to FIG.
- the index output from the linear prediction coefficient calculation circuit 17 ° is input to the code output circuit 2905, and the first coding circuit
- a set of indices output from each of the 0-1-1 and N-th encoding circuits 1 ⁇ 14 is input, and a set of indices output from the N-th encoding circuit 1005 is input. Is entered. Then, each index is converted into a bit sequence code and output via the output terminal 20.
- FIG. 16 is a block diagram showing a configuration of a speech and music signal encoding device according to a seventh embodiment of the present invention. Blocks surrounded by dotted lines in the figure are called pitch prediction filters. Fig. 16 is obtained by adding a pitch prediction filter to Fig. 3. In the following, the blocks different from those in FIG. 3 are the storage circuit 5 10, the pitch signal generation circuit 1 12, the third gain circuit 16 2, the adder 18 4, and the first minimization circuit 5 5 0, the sign output circuit 590 will be described.
- the storage circuit 510 receives the fifth sound source signal from the adder 184 and holds the fifth sound source signal.
- the storage circuit 510 outputs the fifth sound source signal that has been input and held in the past to the pitch signal generation circuit 1 12.
- the past fifth sound source signal held in the storage circuit 5 10 and the index output from the first minimization circuit 5 50 are input to the pitch signal generation circuit 1 12.
- the index specifies the delay d.
- the first pitch vector is L samples equivalent to the vector length from a point d samples past the start of the current frame. Is generated by cutting out the signal.
- d a signal of d samples is cut out, and the cut out d samples are repeatedly connected to generate a first pitch vector having a vector length of L samples.
- the pitch signal generation circuit 112 outputs the first pitch vector to a third gain circuit 162.
- the third gain circuit 16 2 has a table in which gain values are stored.
- the index output from the first minimizing circuit 550 and the first pitch vector output from the pitch signal generating circuit 112 are input to the third gain circuit 162.
- a third gain corresponding to the index is read from the table, a second pitch vector is generated by multiplying the third gain by the first pitch vector, and the generated second pitch vector is generated.
- the vector is output to adder 184.
- the second sound source vector output from the first gain circuit 160 and the second pitch vector output from the third gain circuit 162 are input to the adder 184. Is performed.
- the adder 184 calculates the sum of the second sound source vector and the second pitch vector, and uses this value as the fifth sound source vector to the first bandpass filter 120. Output.
- the indices corresponding to all the first sound source vectors stored in the first sound source generating circuit 110 are sent to the first sound source generating circuit 110. Output sequentially.
- the indices corresponding to all the delays d within the range defined in the pitch signal generation circuit 112 are sequentially output to the pitch signal generation circuit 112.
- the indices corresponding to all the first gains stored in the first gain circuit 160 are sequentially output to the first gain circuit 160.
- Indexes corresponding to all of the third gains stored in the third gain circuit 162 are sequentially output to the third gain circuit 162.
- the first minimizing circuit 550 sequentially inputs the first weighted difference vector output from the weighting filter 140, and calculates the norm thereof.
- the first minimizing circuit 550 selects the first sound source vector, the delay d, the first gain, and the third gain so as to minimize the norm, and corresponds to these. And outputs them to the code output circuit 590 collectively.
- an index corresponding to the quantized linear prediction coefficient output from the linear prediction coefficient calculation circuit 170 is input.
- the code output circuit 590 receives the index output from the first minimization circuit 550 and corresponds to each of the first excitation vector, the delay d, the first gain, and the third gain. Is done.
- the code output circuit 590 receives from the orthogonal transform coefficient quantization circuit 260 a set of indices composed of shape code vectors and quantization gain indices for Nsbv sub-vectors. Is done. Then, each index is converted into a bit sequence code and output via the output terminal 20.
- FIG. 17 is a block diagram showing a configuration of an audio / music signal encoding apparatus according to an eighth embodiment of the present invention.
- the downsample circuit 780, the first linear prediction coefficient calculation circuit 770, the first linear prediction synthesis filter 1332, and the third differentiator which are different blocks from FIG. 1 8 3
- Upsampling circuit 7 8 1, 1st differencer 1 80, a second linear prediction coefficient calculation circuit 771, a third linear prediction coefficient calculation circuit 772, a linear prediction inverse filter 730, and a sign output circuit 790 will be described.
- the down-sampling circuit 780 receives an input vector from the input terminal 10 and down-samples the input vector to obtain a second input vector having a first band, which is a first linear prediction coefficient calculating circuit 7. 7 0 and the third differentiator 18 3.
- the first band is from F sl [H z] to F el [H z] as in the first embodiment
- the band of the input vector is from F sO [H z] to F eO [H z]. (Third band).
- the configuration of the down-sample circuit is described in section 4.1.1 of the document entitled Multi Rate Systems and Filter Banks J by PP Vaidyanathan (Reference 6).
- the first linear prediction coefficient calculation circuit 770 receives a second input vector from the down-sampling circuit 780, performs a linear prediction analysis on the second input vector, and generates a first linear prediction coefficient having a first band. And further quantizes the first linear prediction coefficient to obtain a first quantized linear prediction coefficient.
- the first linear prediction coefficient calculation circuit 770 converts the first linear prediction coefficient into a first Output to the weighting filter 140 and calculate the index corresponding to the first quantized linear prediction coefficient, the first linear prediction synthesis filter 132, the linear prediction inverse filter 730 and the third linear prediction coefficient calculation Output to the circuit 772 and the sign output circuit 790.
- the first linear prediction synthesis filter 1 32 includes a table in which first quantized linear prediction coefficients are stored.
- the first linear prediction synthesis filter 1 32 includes a fifth sound source vector output from the adder 18 4 and a first quantization linear output from the first linear prediction coefficient calculation circuit 7 70. An index corresponding to the prediction coefficient is input.
- the first linear prediction synthesis filter 132 reads the first quantized linear prediction coefficient corresponding to the index from the table, and sets the first quantized linear prediction coefficient to the synthesis filter. By driving the evening with the fifth sound source vector, a first reproduction vector having a first band is generated. Then, the first reproduction vector is output to the third differentiator 183 and the up-sampling circuit 781.
- the third differentiator 183 outputs the first reproduction vector output from the first linear prediction synthesis filter 132 and the second input vector output from the downsampler 780. And calculate their difference, and weight this as the second difference vector Output to filter 140.
- the up-sampling circuit 780 1 receives the first reproduction vector output from the first linear prediction synthesis filter 13 2, and assembles the first reproduction vector to obtain a third reproduction signal having a third band. Generate a vector.
- the third band is from F sO [H z] to F eO [H z].
- the up-sampling circuit 781 outputs the third reproduction vector to the first differentiator 180.
- the configuration of the up-sampling circuit is described in section 4.1.1 of the document entitled "Multi rate Systems and Fiber Banks" by PP Vaidy Anathan (Reference 6).
- the first differentiator 180 receives the input vector via the input terminal 10 and the third reproduction vector output from the upsampler circuit 781, and calculates the difference between them. This is output to the linear prediction inverse filter 730 as a first difference vector.
- the second linear prediction coefficient calculation circuit 771 receives the input vector from the input terminal 10, performs a linear prediction analysis on the input vector, and calculates a second linear prediction coefficient having a third band. Determining: The second linear prediction coefficient is output to a third linear prediction coefficient calculation circuit 772.
- the third linear prediction coefficient calculation circuit 772 has a table in which the first quantized linear prediction coefficient is stored.
- the third linear prediction coefficient calculation circuit 772 includes a second linear prediction coefficient output from the second linear prediction coefficient calculation circuit 771, and an output from the first linear prediction coefficient calculation circuit 77
- the index corresponding to the first quantized linear prediction coefficient to be performed is input.
- the third linear prediction coefficient calculation circuit 772 reads out the first quantized linear prediction coefficient corresponding to the index from the table, converts the first quantized linear prediction coefficient into LSP, By converting this into a sampling frequency, a first LSP corresponding to the sampling frequency of the input signal is generated. Also, the third linear prediction coefficient calculation circuit 772 converts the second linear prediction coefficient into an LSP, and generates a second LSP.
- a third linear prediction coefficient calculation circuit 772 calculates a difference between the second LSP and the first LSP. This difference value is used as the third LSP.
- the conversion of the sampling frequency of the LSP is described in Japanese Patent Application No. 9-202475 (Reference 7).
- the third LSP is quantized, and the quantized third LSP is converted to linear prediction coefficients, and has a third band. Then, a third quantized linear prediction coefficient is generated.
- the index corresponding to the third quantized linear prediction coefficient is output to the linear prediction inverse filter ⁇ 30 and the sign output circuit 790.
- the inverse linear prediction filter 730 includes a first table in which first quantized linear prediction coefficients are stored and a second table in which third quantized linear prediction coefficients are stored.
- the linear prediction inverse filter 730 includes a first index corresponding to the first quantized linear prediction coefficient output from the first linear prediction coefficient calculation circuit 770, and a third linear prediction coefficient
- the second index corresponding to the third quantized linear prediction coefficient output from the calculation circuit 772 and the first difference vector output from the first differentiator 180 are input.
- the linear prediction inverse filter 730 reads out the first quantized linear prediction coefficient corresponding to the first index from the first table, converts it into LSF, and further converts this to a sampling frequency. And generating a first LSP corresponding to the sampling frequency of the input signal.
- a third quantized linear prediction coefficient corresponding to the second index is read from the second table and converted into LSP to generate a third LSP.
- the first LSP and the third LSP are added to generate a second LSP.
- the inverse linear prediction filter 730 converts the second LSP into linear prediction coefficients, and generates second quantized linear prediction coefficients.
- the linear prediction inverse filter 730 drives the inverse filter in which the second quantized linear prediction coefficient is set by the first difference vector, thereby converting the first residual vector. Generate.
- the first residual vector is output to the orthogonal transform circuit 240.
- the sign output circuit 790 includes an index corresponding to the first quantized linear prediction coefficient output from the first linear prediction coefficient calculation circuit 770, and a third linear prediction coefficient calculation circuit 772.
- An index corresponding to the output third quantized linear prediction coefficient and an output from the first minimizing circuit 5 ⁇ 0, the first sound source vector, the delay d, the first gain and the third An index corresponding to each of the gains and a set of indices output from the orthogonal transform coefficient quantization circuit 260 and composed of the shape code vectors for the Nsbv sub-vectors and the indices of the quantization gain are input.
- Each of the indices is converted into a bit-sequence code and output via an output terminal 20. And output.
- FIG. 18 is a block diagram showing a configuration of an audio / music signal decoding device corresponding to the first embodiment according to the ninth embodiment of the present invention.
- the code of the bit sequence is input from the input terminal 30 to the present decoding device.
- the code input circuit 410 converts the code of the bit sequence input from the input terminal 30 into an index.
- the index corresponding to the first sound source vector is output to first sound source generation circuit 110.
- the index corresponding to the first gain is output to the first gain circuit 160.
- the index corresponding to the quantized linear prediction coefficient is output to the linear prediction synthesis filter 130 and the linear prediction synthesis filter 131.
- a set of indices in which the index Nsbv corresponding to each of the shape code vector and the quantization gain for the sub-vectors is integrated into the number of sub-vectors is output to the orthogonal transform coefficient inverse quantization circuit 460.
- the first sound source generating circuit 110 receives the index output from the code input circuit 410, and stores a first sound source vector corresponding to the index into a table in which a plurality of sound source vectors are stored. And outputs it to the first gain circuit 160.
- the first gain circuit 160 has a table in which the quantization gain is stored.
- the first gain circuit 160 receives the index output from the code input circuit 410 and the first sound source vector output from the first sound source generation circuit 110, and receives the index, A corresponding first gain is read from the table, and the first gain is multiplied by the first sound source vector to generate a second sound source vector.
- the generated second sound source vector is output to the first bandpass filter 120.
- the second sound source vector output from the first gain circuit 160 is input to the first bandpass filter 120.
- the second sound source vector is band-limited to a first band by this filter, and generates a first excitation vector.
- the first bandpass filter 120 outputs the first excitation vector to the linear prediction synthesis filter 130.
- the configuration of the orthogonal transform coefficient inverse quantization circuit 460 will be described with reference to FIG. In FIG. 20, there are N sbv blocks surrounded by a dotted line. In each of those blocks The Nsbv quantization sub-vectors defined in the band selection circuit 250 in FIG. 3 are represented as the following Expression 11. The Nsbv quantization vectors are decoded.
- j and k represent indexes.
- the index input circuit 4630 is a set of an index consisting of the shape code vector and the quantization gain index for the Nsbv quantization sub-vectors output from the code input circuit 410 via the input terminal 4650.
- G Enter if. From the index set if, the index i sbs.O specifying the shape code vector cO [j] (n), n-0,..., L-1 and the quantization gain gO [k] are specified.
- the index i sbg.O to be determined is taken out, issbs, 0 is output to the table 4610, and isbg.O is output to the gain circuit 4620.
- Table 4610 shows that cO [j]
- the Nsbv quantized sub-vectors (Equation 11) are arranged in a second band defined by the band selection circuit 250 in FIG. 3, and other than the second band
- a second excitation vector corresponding to the whole band for example, the 8 kHz band when the sampling frequency of the reproduction signal is 16 kHz
- the excitation vector is output to the orthogonal inverse transform circuit 440 via the output terminal 4660.
- the orthogonal inverse transform circuit 440 receives the second excitation vector output from the orthogonal transform coefficient inverse quantization circuit 460, and orthogonally inverse transforms the second excitation vector to obtain a third excitation vector. Then, the third excitation vector is output to the linear prediction synthesis filter 131.
- the inverse orthogonal transform an inverse discrete cosine transform (IDCT) can be used.
- the linear prediction synthesis filter 130 includes a table in which the quantized linear prediction coefficients are stored.
- the first excitation vector output from the first bandpass filter 120 and an index corresponding to the quantized linear prediction coefficient output from the code input circuit 410 are input to the linear prediction synthesis filter 130. Further, the linear prediction synthesis filter 130 reads out the quantized linear prediction coefficient corresponding to the index from the table, and outputs the synthesis filter 1 / A (z) in which the quantized linear prediction coefficient is set to the first filter.
- the first regeneration vector is generated by driving with the excitation vector. Then, the first reproduction vector is output to the adder 182.
- the linear prediction synthesis filter 131 has a table in which the quantized linear prediction coefficients are stored.
- the linear prediction synthesis filter 131 includes a third excitation vector output from the orthogonal inverse transform circuit 440 and a quantized linear prediction output from the code input circuit 410. The index corresponding to the coefficient is input. Further, the linear prediction synthesis filter 13 1 reads out the quantized linear prediction coefficients corresponding to the indices from the table, and outputs the synthesis filter 1 ZA (z) in which the quantized linear prediction coefficients are set to the third filter.
- the second regeneration vector is generated by being driven by the excitation vector.
- the second reproduction vector is output to the adder 182.
- the adder 182 receives the first reproduction vector output from the linear prediction synthesis filter 130 and the second reproduction vector output from the linear prediction synthesis filter 131, The sum of these is calculated and output via the output terminal 40 as a third reproduction vector.
- the ninth embodiment described with reference to FIG. 18 is a case where the number of bands is two, but the case where the number of bands is extended to three or more will be described below.
- Figure 18 can be rewritten as in Figure 21.
- the first decoding circuit 105 in FIG. 21 is equivalent to FIG. 22 and the second decoding circuit 105 in FIG. 21 is equivalent to FIG.
- Each block constituting FIGS. 22 and 23 is the same as each block described in FIG.
- the tenth embodiment of the present invention is realized by extending the number of bands to three in the ninth embodiment.
- the configuration of the audio / music signal decoding apparatus according to the tenth embodiment of the present invention can be represented by a block diagram shown in FIG.
- the first decoding circuit 105 is equivalent to FIG. 22
- the second decoding circuit 105 is equivalent to FIG. 22
- the third decoding circuit 105 Is equivalent to FIG.
- the code input circuit 4101 converts the code of the bit sequence input from the input terminal 30 into an index, and converts the index corresponding to the quantized linear prediction coefficient into the first decoding circuit 1 0 5 1 and the second Output to the second decoding circuit 105 and the third decoding circuit 105, and the index corresponding to the sound source vector and the gain is output to the first decoding circuit 105 and the second decoding circuit 100. 52, and a set of shape codes corresponding to the sub-vectors and an index corresponding to the quantization gain are output to the third decoding circuit 105.
- the eleventh embodiment of the present invention is realized by extending the number of bands to N in the ninth embodiment.
- the configuration of the audio / music signal decoding apparatus according to the eleventh embodiment of the present invention can be represented by a block diagram shown in FIG.
- the first decoding circuit 1 0 5 Each of the 1st to (N ⁇ 1) th decoding circuits 105 is equivalent to FIG. 22, and the N-th decoding circuit 105 is equivalent to FIG.
- the code input circuit 4102 converts the code of the bit sequence input from the input terminal 30 into an index, and converts the index corresponding to the quantized linear prediction coefficient from the first decoding circuit 1051 to the N-1) to each of the decoding circuit 105 and the ⁇ -th decoding circuit 105, and the indices corresponding to the sound source vector and the gain are output from the first decoding circuit 1051 to the (N-1) is output to each of the decoding circuits 105, and the set of the shape code vector for the sub-vector and the index corresponding to the quantization gain is output to the N-th decoding circuit 105. Output to 5.
- the first decoding circuit 105 in FIG. 21 is based on a decoding system corresponding to an encoding system using the A-b-S method.
- a decoding method corresponding to an encoding method other than the A-b-S method can be applied to 1051.
- a decoding method corresponding to an encoding method using time-frequency conversion is applied to first decoding circuit 1051.
- the twelfth embodiment of the present invention is realized by applying a decoding system corresponding to an encoding system using time-frequency conversion in the ninth embodiment.
- the configuration of the audio and music signal decoding device according to the 12th embodiment of the present invention can be represented by a block diagram shown in FIG.
- the first decoding circuit 1061 is equivalent to FIG. 23, and the second decoding circuit 1052 is equivalent to FIG.
- the code input circuit 4103 converts the code of the bit sequence input from the input terminal 30 into an index, and converts the index corresponding to the quantized linear prediction coefficient into the first decoding circuit 1061 and the second decoding circuit.
- Output to decoding circuit 1052, and set of shape code vector for sub-vector and index corresponding to quantization gain to first decoding circuit 1061 and second decoding circuit 1052 Output.
- the thirteenth embodiment of the present invention is realized by extending the number of bands to three in the first and second embodiments.
- the configuration of the audio and music signal decoding apparatus according to the thirteenth embodiment of the present invention can be represented by a block diagram shown in FIG.
- the first decoding circuit 1061 is equivalent to FIG. 23
- the second decoding circuit 1062 is equivalent to FIG. 23
- the third decoding circuit 105 Is equivalent to FIG.
- the sign input circuit 4 1 0 4 The code of the bit sequence input from the input terminal 3 ⁇ is converted into an index, and the index corresponding to the quantized linear prediction coefficient is converted into the first decoding circuit 1061, the second decoding circuit 1062 and the third decoding circuit.
- the 14th embodiment of the present invention is realized by extending the number of bands to N in the 12th embodiment.
- the configuration of the audio / music signal decoding apparatus according to the fourteenth embodiment of the present invention can be represented by a block diagram shown in FIG.
- each of the first decoding circuit 106 1 to the N ⁇ 1th decoding circuit 106 4 is equivalent to FIG.
- the code input circuit 410 converts the code of the bit sequence input from the input terminal 30 into an index, and converts the index corresponding to the quantized linear prediction coefficient from the first decoding circuit 1061 to the N-th 1 to the decoding circuit 1 064 and the N-th decoding circuit 1 0 5 5, and outputs the first set of the shape code vector for the sub-vector and the index set corresponding to the quantization gain.
- the signal is output from the circuit 1061 to each of the (N-1) th decoding circuit 1064 and the Nth decoding circuit 105.
- FIG. 29 is a block diagram showing a configuration of a speech and music signal decoding device corresponding to the seventh embodiment according to the fifteenth embodiment of the present invention.
- blocks different from the ninth embodiment in FIG. 18 are a storage circuit 510, a pitch signal generation circuit 112, a third gain circuit 1622, an adder 1884 Since the sign input circuit 6 10, the power, storage circuit 5 10, pitch signal generation circuit 1 12, third gain circuit 16 2, and adder 18 4 are the same as in FIG. 16, The description will be omitted, and the sign input circuit 610 will be described.
- the code input circuit 610 converts the code of the bit sequence input from the input terminal 30 into an index.
- the index corresponding to the first sound source vector is output to first sound source generation circuit 110.
- the index corresponding to the delay d is output to the pitch signal generation circuit 112.
- the index corresponding to the first gain is output to the first gain circuit 160.
- the index corresponding to the third gain is output to the third gain circuit 162.
- the index corresponding to the quantized linear prediction coefficient is Output to the linear prediction synthesis filter 13 0 and the linear prediction synthesis filter 13 1.
- a set of indices in which the indices corresponding to each of the shape code vector and the quantization gain for the sub-vectors are combined into Nsbv sub-vectors is sent to the orthogonal transform coefficient inverse quantization circuit 460. Is output.
- FIG. 30 is a block diagram showing a configuration of a speech and music signal decoding device corresponding to the eighth embodiment according to the sixteenth embodiment of the present invention.
- the blocks different from those in FIG. 29 are the code input circuit 8 10, the first linear prediction coefficient synthesis filter 13 2, the up-sampling circuit 7 8 1, and the second linear prediction synthesis filter 8 3 1 Is explained.
- the code input circuit 810 converts the code of the bit sequence input from the input terminal 30 into an index.
- the index corresponding to the first sound source vector is output to first sound source generation circuit 110.
- the index corresponding to the delay d is output to the pitch signal generation circuit 112.
- the index corresponding to the first gain is output to first gain circuit 160.
- the index corresponding to the third gain is output to third gain circuit 162.
- the index corresponding to the first quantized linear prediction coefficient is output to first linear prediction synthesis filter 1332 and second linear prediction synthesis filter 831.
- the index corresponding to the third quantized linear prediction coefficient is output to second linear prediction synthesis filter 831.
- a set of indices obtained by combining N sbv sub-vectors with indices corresponding to the shape code vector and the quantization gain for each sub-vector is output to the orthogonal transform coefficient inverse quantization circuit 460 .
- the first linear prediction synthesis filter 1 32 includes a table in which first quantized linear prediction coefficients are stored.
- the first linear prediction synthesis filter 1 32 has a fifth excitation vector output from the adder 18 4 and an index corresponding to the first quantized linear prediction coefficient output from the sign input circuit 8 10. Enter In addition, a first quantized linear prediction coefficient corresponding to the index is read out from the table, and a synthesis filter in which the first quantized linear prediction coefficient is set is determined by the fifth sound source vector. By driving, a first reproduction vector having a first band is obtained. Then, the first reproduction vector is output to the up-sampling circuit 781.
- the up-sampling circuit 780 1 receives the first reproduction vector output from the first linear prediction synthesis filter 13 2, and up-samples this to perform a third reproduction having a third band. Get the vector. Then, the third reproduction vector is output to the first adder 182.
- the second linear prediction synthesis filter 831 is composed of a first table having a first quantized linear prediction coefficient having a first band, and a third quantization having a third band. And a second table in which linear prediction coefficients are stored.
- the second linear prediction synthesis filter 831 includes a third excitation vector output from the orthogonal inverse transform circuit 4400 and a first quantized linear prediction output from the code input circuit 810. A first index corresponding to the coefficient and a second index corresponding to the third quantized linear prediction coefficient are input.
- the second linear prediction synthesis filter 831 reads out the first quantized linear prediction coefficient corresponding to the first index from the first table, converts this into an LSP, and samples this.
- a first LSP corresponding to the sampling frequency of the third reproduction vector is generated.
- a third quantized linear prediction coefficient corresponding to the second index is read from the second table, and is converted into LSP to generate a third LSP.
- the second LSP obtained by adding the first LSP and the third LSP is converted into a linear prediction coefficient, and a second linear prediction coefficient is generated.
- the second linear prediction synthesis filter 831 drives the synthesis filter in which the second linear prediction coefficient is set by the third excitation vector, so that the second reproduction having the third band is performed. Generate the vector. Then, the second reproduction vector is output to the adder 18.
- the adder 182 receives the third reproduction vector output from the up-sampling circuit 781, and the second reproduction vector output from the second linear prediction synthesis filter 831, Then, the sum of these is calculated, and this is output via an output terminal 40 as the fourth reproduction vector.
- a speech music signal can be encoded favorably over the whole band.
- the input signal is generated by a sound source signal having a band characteristic corresponding to a low band of the input signal.
- Generating a first reproduction signal by driving a linear prediction synthesis filter obtained from the signal, and driving an inverse filter of the linear prediction synthesis filter by a difference signal between the input signal and the first reproduction signal. Since the residual signal is generated by using the encoding method based on the orthogonal transform, and the high frequency component of the residual signal is encoded, the encoding performance for the high frequency component of the input signal is improved. is there.
Landscapes
- Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Signal Processing (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Compression, Expansion, Code Conversion, And Decoders (AREA)
Description
Claims
Priority Applications (4)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| EP99925329A EP1087378B1 (en) | 1998-06-15 | 1999-06-15 | Voice/music signal encoder and decoder |
| US09/719,826 US6865534B1 (en) | 1998-06-15 | 1999-06-15 | Speech and music signal coder/decoder |
| DE69941259T DE69941259D1 (de) | 1998-06-15 | 1999-06-15 | Sprach/musik signal kodierer und dekodierer |
| CA002335284A CA2335284A1 (en) | 1998-06-15 | 1999-06-15 | Speech and music signal coder/decoder |
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP16657398A JP3541680B2 (ja) | 1998-06-15 | 1998-06-15 | 音声音楽信号の符号化装置および復号装置 |
| JP10/166573 | 1998-06-15 |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO1999066497A1 true WO1999066497A1 (en) | 1999-12-23 |
Family
ID=15833779
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/JP1999/003185 Ceased WO1999066497A1 (en) | 1998-06-15 | 1999-06-15 | Voice/music signal encoder and decoder |
Country Status (6)
| Country | Link |
|---|---|
| US (1) | US6865534B1 (ja) |
| EP (1) | EP1087378B1 (ja) |
| JP (1) | JP3541680B2 (ja) |
| CA (1) | CA2335284A1 (ja) |
| DE (1) | DE69941259D1 (ja) |
| WO (1) | WO1999066497A1 (ja) |
Families Citing this family (18)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US7072832B1 (en) * | 1998-08-24 | 2006-07-04 | Mindspeed Technologies, Inc. | System for speech encoding having an adaptive encoding arrangement |
| FI119576B (fi) * | 2000-03-07 | 2008-12-31 | Nokia Corp | Puheenkäsittelylaite ja menetelmä puheen käsittelemiseksi, sekä digitaalinen radiopuhelin |
| US6686193B2 (en) | 2000-07-10 | 2004-02-03 | Vertex Pharmaceuticals, Inc. | High throughput method and system for screening candidate compounds for activity against target ion channels |
| US7240001B2 (en) * | 2001-12-14 | 2007-07-03 | Microsoft Corporation | Quality improvement techniques in an audio encoder |
| US6934677B2 (en) | 2001-12-14 | 2005-08-23 | Microsoft Corporation | Quantization matrices based on critical band pattern information for digital audio wherein quantization bands differ from critical bands |
| US7752052B2 (en) | 2002-04-26 | 2010-07-06 | Panasonic Corporation | Scalable coder and decoder performing amplitude flattening for error spectrum estimation |
| US7299190B2 (en) * | 2002-09-04 | 2007-11-20 | Microsoft Corporation | Quantization and inverse quantization for audio |
| JP4676140B2 (ja) | 2002-09-04 | 2011-04-27 | マイクロソフト コーポレーション | オーディオの量子化および逆量子化 |
| US7502743B2 (en) * | 2002-09-04 | 2009-03-10 | Microsoft Corporation | Multi-channel audio encoding and decoding with multi-channel transform selection |
| JP3881943B2 (ja) * | 2002-09-06 | 2007-02-14 | 松下電器産業株式会社 | 音響符号化装置及び音響符号化方法 |
| US7486719B2 (en) * | 2002-10-31 | 2009-02-03 | Nec Corporation | Transcoder and code conversion method |
| US7844451B2 (en) | 2003-09-16 | 2010-11-30 | Panasonic Corporation | Spectrum coding/decoding apparatus and method for reducing distortion of two band spectrums |
| WO2006062202A1 (ja) * | 2004-12-10 | 2006-06-15 | Matsushita Electric Industrial Co., Ltd. | 広帯域符号化装置、広帯域lsp予測装置、帯域スケーラブル符号化装置及び広帯域符号化方法 |
| US7539612B2 (en) * | 2005-07-15 | 2009-05-26 | Microsoft Corporation | Coding and decoding scale factor information |
| CN101548318B (zh) * | 2006-12-15 | 2012-07-18 | 松下电器产业株式会社 | 编码装置、解码装置以及其方法 |
| US9602127B1 (en) * | 2016-02-11 | 2017-03-21 | Intel Corporation | Devices and methods for pyramid stream encoding |
| US10957331B2 (en) | 2018-12-17 | 2021-03-23 | Microsoft Technology Licensing, Llc | Phase reconstruction in a speech decoder |
| US10847172B2 (en) * | 2018-12-17 | 2020-11-24 | Microsoft Technology Licensing, Llc | Phase quantization in a speech encoder |
Citations (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPH05265492A (ja) * | 1991-03-27 | 1993-10-15 | Oki Electric Ind Co Ltd | コード励振線形予測符号化器及び復号化器 |
| JPH08263096A (ja) * | 1995-03-24 | 1996-10-11 | Nippon Telegr & Teleph Corp <Ntt> | 音響信号符号化方法及び復号化方法 |
| JPH0946233A (ja) * | 1995-07-31 | 1997-02-14 | Kokusai Electric Co Ltd | 音声符号化方法とその装置、音声復号方法とその装置 |
| JPH09127987A (ja) * | 1995-10-26 | 1997-05-16 | Sony Corp | 信号符号化方法及び装置 |
| JPH09127994A (ja) * | 1995-10-26 | 1997-05-16 | Sony Corp | 信号符号化方法及び装置 |
| JPH09130260A (ja) * | 1995-10-26 | 1997-05-16 | Victor Co Of Japan Ltd | 音響信号の符号化装置及び復号化装置 |
Family Cites Families (14)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US4956871A (en) | 1988-09-30 | 1990-09-11 | At&T Bell Laboratories | Improving sub-band coding of speech at low bit rates by adding residual speech energy signals to sub-bands |
| JP3249144B2 (ja) | 1991-03-29 | 2002-01-21 | 株式会社東芝 | 音声符号化装置 |
| JP3264679B2 (ja) | 1991-08-30 | 2002-03-11 | 沖電気工業株式会社 | コード励振線形予測符号化装置及び復号化装置 |
| JP3089769B2 (ja) | 1991-12-03 | 2000-09-18 | 日本電気株式会社 | 音声符号化装置 |
| US5526464A (en) * | 1993-04-29 | 1996-06-11 | Northern Telecom Limited | Reducing search complexity for code-excited linear prediction (CELP) coding |
| JP3186489B2 (ja) | 1994-02-09 | 2001-07-11 | ソニー株式会社 | ディジタル信号処理方法及び装置 |
| JPH09127995A (ja) | 1995-10-26 | 1997-05-16 | Sony Corp | 信号復号化方法及び信号復号化装置 |
| TW321810B (ja) * | 1995-10-26 | 1997-12-01 | Sony Co Ltd | |
| JPH09127985A (ja) | 1995-10-26 | 1997-05-16 | Sony Corp | 信号符号化方法及び装置 |
| US5778335A (en) * | 1996-02-26 | 1998-07-07 | The Regents Of The University Of California | Method and apparatus for efficient multiband celp wideband speech and music coding and decoding |
| JPH09281995A (ja) | 1996-04-12 | 1997-10-31 | Nec Corp | 信号符号化装置及び方法 |
| JP3092653B2 (ja) | 1996-06-21 | 2000-09-25 | 日本電気株式会社 | 広帯域音声符号化装置及び音声復号装置並びに音声符号化復号装置 |
| JP3357795B2 (ja) | 1996-08-16 | 2002-12-16 | 株式会社東芝 | 音声符号化方法および装置 |
| US6345246B1 (en) * | 1997-02-05 | 2002-02-05 | Nippon Telegraph And Telephone Corporation | Apparatus and method for efficiently coding plural channels of an acoustic signal at low bit rates |
-
1998
- 1998-06-15 JP JP16657398A patent/JP3541680B2/ja not_active Expired - Fee Related
-
1999
- 1999-06-15 WO PCT/JP1999/003185 patent/WO1999066497A1/ja not_active Ceased
- 1999-06-15 EP EP99925329A patent/EP1087378B1/en not_active Expired - Lifetime
- 1999-06-15 US US09/719,826 patent/US6865534B1/en not_active Expired - Fee Related
- 1999-06-15 DE DE69941259T patent/DE69941259D1/de not_active Expired - Lifetime
- 1999-06-15 CA CA002335284A patent/CA2335284A1/en not_active Abandoned
Patent Citations (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPH05265492A (ja) * | 1991-03-27 | 1993-10-15 | Oki Electric Ind Co Ltd | コード励振線形予測符号化器及び復号化器 |
| JPH08263096A (ja) * | 1995-03-24 | 1996-10-11 | Nippon Telegr & Teleph Corp <Ntt> | 音響信号符号化方法及び復号化方法 |
| JPH0946233A (ja) * | 1995-07-31 | 1997-02-14 | Kokusai Electric Co Ltd | 音声符号化方法とその装置、音声復号方法とその装置 |
| JPH09127987A (ja) * | 1995-10-26 | 1997-05-16 | Sony Corp | 信号符号化方法及び装置 |
| JPH09127994A (ja) * | 1995-10-26 | 1997-05-16 | Sony Corp | 信号符号化方法及び装置 |
| JPH09130260A (ja) * | 1995-10-26 | 1997-05-16 | Victor Co Of Japan Ltd | 音響信号の符号化装置及び復号化装置 |
Non-Patent Citations (1)
| Title |
|---|
| See also references of EP1087378A4 * |
Also Published As
| Publication number | Publication date |
|---|---|
| JP2000003193A (ja) | 2000-01-07 |
| EP1087378B1 (en) | 2009-08-12 |
| US6865534B1 (en) | 2005-03-08 |
| CA2335284A1 (en) | 1999-12-23 |
| EP1087378A4 (en) | 2005-10-26 |
| JP3541680B2 (ja) | 2004-07-14 |
| EP1087378A1 (en) | 2001-03-28 |
| DE69941259D1 (de) | 2009-09-24 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| JP3134817B2 (ja) | 音声符号化復号装置 | |
| US6694292B2 (en) | Apparatus for encoding and apparatus for decoding speech and musical signals | |
| EP2041745B1 (en) | Adaptive encoding and decoding methods and apparatuses | |
| JP4005359B2 (ja) | 音声符号化及び音声復号化装置 | |
| WO1999066497A1 (en) | Voice/music signal encoder and decoder | |
| EP4336500B1 (en) | Methods, encoder and decoder for linear predictive encoding and decoding of sound signals upon transition between frames having different sampling rates | |
| US20040111257A1 (en) | Transcoding apparatus and method between CELP-based codecs using bandwidth extension | |
| JP2005515486A (ja) | Celpによる音声符号間のトランスコーディング・スキーム | |
| JP2002268686A (ja) | 音声符号化装置及び音声復号化装置 | |
| JP4108317B2 (ja) | 符号変換方法及び装置とプログラム並びに記憶媒体 | |
| EP1326237A2 (en) | Excitation quantisation in noise feedback coding | |
| JP3147807B2 (ja) | 信号符号化装置 | |
| JP3268750B2 (ja) | 音声合成方法及びシステム | |
| JP3916934B2 (ja) | 音響パラメータ符号化、復号化方法、装置及びプログラム、音響信号符号化、復号化方法、装置及びプログラム、音響信号送信装置、音響信号受信装置 | |
| JP4293005B2 (ja) | 音声音楽信号の符号化装置および復号装置 | |
| JPH09127987A (ja) | 信号符号化方法及び装置 | |
| JP2010256932A (ja) | 音声信号標本値の符号化または復号化のための方法並びに符号化器ないし復号化器 | |
| JP2002073097A (ja) | Celp型音声符号化装置とcelp型音声復号化装置及び音声符号化方法と音声復号化方法 | |
| US20250356867A1 (en) | Methods, Encoder And Decoder For Linear Predictive Encoding And Decoding Of Sound Signals Upon Transition Between Frames Having Different Sampling Rates | |
| JPH0612097A (ja) | 音声の予測符号化方法および装置 | |
| JP2775533B2 (ja) | 音声の長期予測装置 | |
| JP3277090B2 (ja) | ゲイン量子化方法及び装置、音声符号化方法及び装置並びに音声復号化方法及び装置 | |
| HK40104768B (en) | Methods, encoder and decoder for linear predictive encoding and decoding of sound signals upon transition between frames having different sampling rates | |
| HK40104768A (en) | Methods, encoder and decoder for linear predictive encoding and decoding of sound signals upon transition between frames having different sampling rates | |
| HK40036813A (en) | Methods, encoder and decoder for linear predictive encoding and decoding of sound signals upon transition between frames having different sampling rates |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| AK | Designated states |
Kind code of ref document: A1 Designated state(s): CA US |
|
| AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): DE FI FR GB NL SE |
|
| DFPE | Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101) | ||
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
| WWE | Wipo information: entry into national phase |
Ref document number: 1999925329 Country of ref document: EP |
|
| ENP | Entry into the national phase |
Ref document number: 2335284 Country of ref document: CA |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 09719826 Country of ref document: US |
|
| WWP | Wipo information: published in national office |
Ref document number: 1999925329 Country of ref document: EP |