[go: up one dir, main page]

TW201145263A - Method and apparatus for decoding an audio signal - Google Patents

Method and apparatus for decoding an audio signal Download PDF

Info

Publication number
TW201145263A
TW201145263A TW100109444A TW100109444A TW201145263A TW 201145263 A TW201145263 A TW 201145263A TW 100109444 A TW100109444 A TW 100109444A TW 100109444 A TW100109444 A TW 100109444A TW 201145263 A TW201145263 A TW 201145263A
Authority
TW
Taiwan
Prior art keywords
channel
audio signal
spatial
information
spatial information
Prior art date
Application number
TW100109444A
Other languages
Chinese (zh)
Other versions
TWI485698B (en
Inventor
Yang-Won Jung
Hee-Suk Pang
Hyen-O Oh
Dong-Soo Kim
Jae-Hyun Lim
Original Assignee
Lg Electronics Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Lg Electronics Inc filed Critical Lg Electronics Inc
Publication of TW201145263A publication Critical patent/TW201145263A/en
Application granted granted Critical
Publication of TWI485698B publication Critical patent/TWI485698B/en

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/008Multichannel audio signal coding or decoding using interchannel correlation to reduce redundancy, e.g. joint-stereo, intensity-coding or matrixing

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Computational Linguistics (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Stereophonic System (AREA)

Abstract

An apparatus for decoding an audio signal and method thereof are disclosed. The present invention includes receiving the audio signal and spatial information, identifying a type of modified spatial information, generating the modified spatial information using the spatial information, and decoding the audio signal using the modified spatial information, wherein the type of the modified spatial information includes at least one of partial spatial information, combined spatial information and expanded spatial information. Accordingly, an audio signal can be decoded into a configuration different from a configuration decided by an encoding apparatus. Even if the number of speakers is smaller or greater than that of multi-channels before execution of downmixing, it is able to generate output channels having the number equal to that of the speakers from a downmix audio signal.

Description

201145263 六、發明說明: 【發明所屬之技術領域】 本發明係關於音頻訊號處理,尤其係關於一種音頻訊號之解 碼裝置及其方法。雖然本發明適合廣泛的應用,但是尤其適合用 於解碼音頻訊號。 【先前技術】 通常,當編碼器編碼音頻訊號時,如果待編碼的音頻訊號為 多通道音頻訊號,此多通道音頻訊號被降混為兩個通道或一個通 道以產生降混音頻訊號,並且從多通道音頻訊號中擷取空間資 訊。空間資訊係為可用於升混來自降混訊號的多通道音頻訊號的 資訊。同時,編碼器係依照預定的樹狀組態,降混多通道音頻訊 號。這個實例巾’預定的樹狀組態可為音頻訊號解碼如及音頻 訊號編碼關所議定之結構^制是,如果識师訊指出所屬預 定樹狀組態類型之-,解碼㈣能夠知道已升混的音頻訊號結 構,例如通道的數目,每個通道的位置等等。 因此,如果編碼器依照預定的樹狀組態降混多通道訊號,這 個過程擷取的空間資訊也由此結構決定。所以,如果解碼器使用 此、”。構所核的m資訊升混此降混音賴號,則依照此結構之 多通道音頻訊號即被產生。 也就疋說,如果解碼器使用編碼器所產生的空間資訊,則升 混過程係依照編碼器與解碼ϋ間之議定結構而完成。所以,無法 產生不遵循献結構的輸出通道音頻訊號。例如,無法升混訊號 201145263 依照議定結構所決定的 【發明内容】 馨於以上的問題, 之解碼裝置及其方法, 致的一或多個問題。 本發明的主要目的在於提供一種音頻訊號 實質上消除由於習知技術之限制及缺點導 本發騎目的在於提供—種音頻訊號之解碼裝置及其方法, 解姑頻訊號,從而使得其結構不同於編所蚊的結構。 、、η务月的另目的在於提供一種音頻訊號之解碼裝置及其方 法可藉由修正編碼時所產生的先前空間資訊從而產生新的空間 資訊以進行音敏號之解碼。 、本發明其它的特徵和優點將在如下的制書巾加以闡述,並 且透^本發明如下的·得崎分地理解或者可以從本發明的實 ^中付出。本發明的目的和其它優點可贿過本發明所記載的說 明書和巾請專職圍巾制指_結構並結合圖式部份,得以實 現和獲得。 為了獲得本發明的這些目的和其他優點,現對本發明作具體 化和概括性的描述,本發明的—種音頻訊號之解财法包含接收 均頻訊號及空間資訊(spatial informati〇n),識別修正式空間資訊的 類型,使用空間資訊產生修正式空㈣訊,以及使祕正式空間 資訊解碼音頻峨,其巾修正式m訊的翻包含部分空間資 訊、組合式空間資訊以及擴充空間資訊至少其中之一。 201145263 為了進-步獲得本發明的這些目的和其他優點,本發明的一 種音頻訊號之解碼方法包含接收空間資訊,使用此空間資訊產生 組合式空間資訊,以及使用此組合式郎魏解碼音頻訊號其 中透過組合包含於空間資訊中的空間參數產生組合式空間資訊。 為了進一步獲得本發明的這些目的和其他優點,本發明的一 種音頻訊號之解碼方法包含接收空間資訊及空間濾波器資訊,其 中空間資訊至少包含—空間參數,空間渡波器資訊至少包含一渡 波器參數;透·合如錄减波n參數產生包含環繞效果的 組合式空間魏;以及組合式郎換音親號為虛擬 環繞訊號。 為了進-步獲得本發明的這些目的和其他優點,本發明的一 種音頻訊號之解碼方法包含魏音麵號;接收包含樹狀組態資 訊及空間參數的空間資訊;透過增加延伸空間資·空間資訊以 產生修正式空間賴;以及使祕正式空間資訊升混音頻訊號, 其中包含根據朗資贿音齡賴換為她升齡頻訊號,以 及根據延伸空間資訊將初級升混音頻峨轉換為二次升混音頻訊 號。 可以理解岐,如上所述的本發明之概括綱和隨後所述的 本發明之詳細_均是射代絲和轉性的綱,並且是為了 進一步揭示本發明之申請專利範圍。 【實施方式】 以下’將結合圖式部份對本發明的較佳實施方式作詳細說明。 6 S: 201145263 本發明使用的術語係選擇當前全球使用的常用術語。並且, 針對特定的實例,申請者任意地選擇術語,其中本發明之較佳實 施例的6兑明中將洋盡地解釋詳細含義。因此,應該根據術語的含 義而並非根據術語的名字來理解本發明。 首先,本發明使用空間資訊產生修正式空間資訊,然後使用 產生的修正式空間資訊解碼音頻訊號。這個實例中,空間資訊係 為依照預定樹狀組態於降混過程中所擷取的空間資訊,並且修正 式空間為訊係為使用空間資訊所新產生的空間資訊。 下面結合「第1圖」詳細解釋本發明。 「第1圖」係為本發明實施例之音頻訊號編碼裝置以及音頻 訊號解碼裝置之方塊圖。 請參考「第1圖」,音頻訊號之編碼裝置1〇〇包含降混單元11〇 以及空間資訊娜單元12〇。音頻訊號之解碼裝置2⑻包含輸出通 道產生單元210以及修正式空間資訊產生單元22〇。 編碼裝置100之降混單元U〇透過降混多通道音頻訊號 產生降混音頻域d。降混音頻簡d可為降混單元m降混多通 道曰頻訊號IN_M所產生的訊號,或者細者任意地降混多通道 音頻訊號IN一Μ所產生的任意降混音頻訊號。 編碼裝置之如資賴取單元12Q從多通道音頻訊號 IN—M中擷㈣間#訊。這個實例中’空間資訊係將降混音頻訊號 4升混為多通道音頻訊號IN_M所需要的資訊。 同時,空間資訊可為紐預定的樹狀組態在降混多通道音頻 201145263 訊號IN一Μ的過程中所擷取的資訊。這個實例中’此樹狀組態可 以對應於音頻訊號解碼以及編碼裝置之間議定的樹狀組態,但本 發明並非限制於此。 空間資訊可包含樹狀組態資訊、指示元以及空間參數等。樹 狀組態資訊係為樹狀組態類型的資訊。所以,若干個多通道以及 各通道降混序列等係依照樹狀組態類型而變化。指示元係用於指 示是否出現延伸空間資訊。此外,空間參數可包含將至少兩個通 道降混為至多兩個通道過程中的通道位階差值(channel levd difference ; CLD)、通道間相關或同調(inter_channelc〇ITelati〇n〇r coherence ; ICC )以及通道預測係數(channd predicti〇n coefficients ; CPC)等。 同時,空間-貝訊操取單元120更能夠操取延伸空間資訊以及 空間資訊。這個實例中,延伸空職訊係為需要被額外地延伸之 具有已升混空間參數的降混音頻訊號d之資訊。延伸空間資訊可 包含延伸通道組態資訊以及延伸空間參數。以後將解釋延伸空間 資訊’並非限制於空間資訊擷取單元120所擷取的延伸空間資訊。 此外,編碼裝置100更包含核心編解碼器(c〇dec)之編碼單 元(圖中未表示),用以透過解碼降混音頻訊號d產生降混音頻位 7C流;㈣資訊編碼單元(圖巾未表示),用以透過編碼空間資訊 S產生空間資訊位it流;以及多I料(圆中未表示),用以透過 多工降混音頻位元流及空間資訊位元流產生音頻_位元流,本 發明並非限制於此。 201145263 解碼裝置200更可包含解多工單元(圖中未表示),用以分離 音頻訊號位元流為降混音頻位元流及空間資訊位元流;核心編解 碼器之解碼單元(圖中未表示),用以解碼降混音頻位元流;以及 空間資訊解碼單元(圖中未表示),用以解碼空間資訊位元流,本 發明並非限制於此。 解碼裝置200之修正式空間資訊產生單元22〇使用空間資訊 識別修正式空間資訊的類型,然後基於所識別出之空間資訊產生 某個類型的修正式雜資訊s,。這個實例巾,此空間資訊可為轉 換自編碼襄置刚的空間資訊s。修正式空間資訊係為使用此空間 資訊所新產生的資訊。 此外,還存在多麵型的修正式空間資訊。修正式空間資訊 的類型至少包含以下之—:a)部分雜資訊,b)岭技間資訊, 以及c)延伸空間資訊,本發明並非限制於此。 部分空間資訊包含部分㈣畔數;組合式調資訊係透過201145263 VI. Description of the Invention: [Technical Field] The present invention relates to audio signal processing, and more particularly to an audio signal decoding apparatus and method thereof. Although the present invention is suitable for a wide range of applications, it is particularly suitable for decoding audio signals. [Prior Art] Generally, when an encoder encodes an audio signal, if the audio signal to be encoded is a multi-channel audio signal, the multi-channel audio signal is downmixed into two channels or one channel to generate a downmix audio signal, and Capture spatial information from multi-channel audio signals. Spatial information is information that can be used to upmix multichannel audio signals from downmix signals. At the same time, the encoder downmixes the multichannel audio signal according to a predetermined tree configuration. This example towel 'predetermined tree configuration can be used for audio signal decoding and audio signal coding. The structure is determined. If the teacher information indicates the type of the predetermined tree configuration, the decoding (4) can know that it has been upgraded. Mixed audio signal structures, such as the number of channels, the location of each channel, and so on. Therefore, if the encoder downmixes the multi-channel signal according to the predetermined tree configuration, the spatial information captured by this process is also determined by this structure. Therefore, if the decoder uses this, the m-information of the core of the nucleus is mixed with the downmix lag, then the multi-channel audio signal according to this structure is generated. That is, if the decoder uses the encoder The spatial information generated, the upmixing process is completed according to the agreed structure between the encoder and the decoding. Therefore, the output channel audio signal that does not follow the structure cannot be generated. For example, the upmix signal 201145263 cannot be as determined according to the agreed structure. SUMMARY OF THE INVENTION One or more problems are caused by the above problems, the decoding apparatus and the method thereof. The main object of the present invention is to provide an audio signal that substantially eliminates the limitations and disadvantages of the prior art. The purpose is to provide a decoding device and method for the audio signal, and to solve the structure of the squad, so that the structure is different from the structure of the mosquito. The other purpose of the MV is to provide an audio signal decoding device and method thereof. The new spatial information can be generated by correcting the previous spatial information generated during the encoding to decode the audio sensitive number. Other features and advantages will be set forth in the following book towel, and can be understood from the following aspects of the present invention or can be paid from the practice of the present invention. The object and other advantages of the present invention can be bribed. The present specification and general description of the present invention will now be realized and obtained in the light of the description and the appended claims. The audio signal method of the present invention comprises receiving a uniform frequency signal and spatial information (spatial informati〇n), identifying a type of corrected spatial information, using spatial information to generate a modified spatial (four) message, and making secret official spatial information. Decoding audio 峨, the tamper-modified m-message includes at least one of partial spatial information, combined spatial information, and extended spatial information. 201145263 In order to further obtain these and other advantages of the present invention, an audio of the present invention The signal decoding method includes receiving spatial information, using the spatial information to generate combined spatial information, and using the group The Lang Wei decoding audio signal generates combined spatial information by combining spatial parameters included in the spatial information. To further obtain the objectives and other advantages of the present invention, an audio signal decoding method of the present invention includes receiving spatial information and space. Filter information, wherein the spatial information includes at least a spatial parameter, and the spatial waver information includes at least one waver parameter; the transparent and combined negative wave n parameter generates a combined space Wei containing a surrounding effect; and the combined Lang sounding pro The number is a virtual surround signal. In order to further obtain the above objects and other advantages of the present invention, an audio signal decoding method of the present invention includes a Weiyin face number; receiving spatial information including tree configuration information and spatial parameters; Adding space and space information to generate a revised space; and making the secret space information upmixed with audio signals, including the conversion of the age-based frequency signal to the age-based frequency signal, and the primary information based on the extended space information The upmix audio is converted to a second upmix audio signal. It is to be understood that the summary of the invention as described above and the details of the invention as described hereinafter are all examples of the generation of filaments and the properties of the invention, and are intended to further disclose the scope of the invention. [Embodiment] Hereinafter, preferred embodiments of the present invention will be described in detail in conjunction with the drawings. 6 S: 201145263 The terminology used in the present invention is a selection of commonly used terms that are currently used globally. Also, the applicant arbitrarily selects terms for a specific example, and the preferred embodiment of the preferred embodiment of the present invention will explain the detailed meaning. Therefore, the invention should be understood in accordance with the meaning of the terms and not by the names of the terms. First, the present invention uses spatial information to generate modified spatial information, and then uses the generated modified spatial information to decode the audio signal. In this example, the spatial information is the spatial information acquired during the downmixing process according to the predetermined tree configuration, and the modified space is the spatial information newly generated by the information system for using the spatial information. The present invention will be explained in detail below with reference to "Fig. 1". Fig. 1 is a block diagram of an audio signal encoding apparatus and an audio signal decoding apparatus according to an embodiment of the present invention. Please refer to "Figure 1". The audio signal encoding device 1〇〇 includes a downmixing unit 11〇 and a spatial information unit 12〇. The audio signal decoding device 2 (8) includes an output channel generating unit 210 and a modified space information generating unit 22A. The downmixing unit U of the encoding device 100 generates a downmix audio domain d by downmixing the multichannel audio signal. The downmix audio simple d can be the signal generated by the downmix unit m downmixing the multi-channel frequency signal IN_M, or the arbitrary downmix audio signal generated by the multi-channel audio signal IN. The encoding device is dependent on the unit 12Q from the multi-channel audio signal IN-M. In this example, the spatial information is the information needed to mix the audio signal 4 liters into the multi-channel audio signal IN_M. At the same time, the spatial information can be the information that the predetermined tree configuration in the process of downmixing the multi-channel audio 201145263 signal IN. In this example, the tree configuration can correspond to the audio signal decoding and the agreed tree configuration between the encoding devices, but the invention is not limited thereto. Spatial information can include tree configuration information, indicator elements, and spatial parameters. The tree configuration information is information of the tree configuration type. Therefore, several multi-channels and each channel downmix sequence are changed according to the tree configuration type. The indicator element is used to indicate whether or not the extended space information appears. In addition, the spatial parameters may include channel levd difference (CLD), channel-to-channel correlation or homology (inter-channelc〇ITelati〇n〇r coherence; ICC) during downmixing of at least two channels into at most two channels. And channel prediction coefficients (channd predicti〇n coefficients; CPC) and so on. At the same time, the space-bein operation unit 120 is more capable of operating extended space information as well as spatial information. In this example, the extended empty message system is information that needs to be additionally extended with the downmix audio signal d having the upmixed spatial parameters. The extended space information can include extended channel configuration information as well as extended space parameters. The extended space information 'will be explained later is not limited to the extended space information captured by the spatial information capturing unit 120. In addition, the encoding apparatus 100 further includes a coding unit (not shown) of the core codec (c) for generating a downmix audio bit 7C stream by decoding the downmix audio signal d; (4) an information coding unit ( The towel is not shown) for generating a spatial information bit stream through the encoded spatial information S; and a multi-material (not shown in the circle) for generating the multiplexed downmix audio bit stream and the spatial information bit stream Audio_bit stream, the invention is not limited thereto. The decoding device 200 may further include a demultiplexing unit (not shown) for separating the audio signal bit stream into a downmix audio bit stream and a spatial information bit stream; a decoding unit of the core codec (Fig. The data is not used to decode the downmix audio bit stream; and the spatial information decoding unit (not shown) is used to decode the spatial information bit stream, and the present invention is not limited thereto. The modified spatial information generating unit 22 of the decoding device 200 uses the spatial information to identify the type of the corrected spatial information, and then generates a certain type of modified miscellaneous information s based on the identified spatial information. For this example towel, this spatial information can be the spatial information s just converted from the coded device. The revised spatial information is the new information generated by using this spatial information. In addition, there are multi-faceted modified spatial information. The type of the modified spatial information includes at least the following: a) partial information, b) ridge information, and c) extended space information, and the present invention is not limited thereto. Part of the spatial information contains part (4) of the number of beaches;

組合空間參數而產生;延伸空_訊職由空間資訊 間資訊而產生。 T 修正式空間資訊產生單元220用以產生修正式空間資訊其 產生方式可依雕正式雜#__㈣化。 釋 各種類型的修正式郎資訊的產生方法。 釋 樹狀==於決定修正式空間資訊的參考可對應於空間資訊之 態資訊以及指示元可包含於維麟里 、心崎撒組 U於編碼裝置之扣㈣St。輸出通道 201145263 -貝訊係用於與解碼裝置200互相連接之揚聲器,其可包含若干輸 出通道以及每雜ifcit道之位置資訊等。輸岐道資訊可由製= 業者事先輸入或者由使用者輸入。 下面將詳細解釋藉由這些資訊決定修正式空間資訊之類型之 方法。 解碼裝置200之輸出通道產生單元21〇使用修正式空間資訊 s以產生降混音頻訊號d的輸出通道音頻訊號〇υτ N。 空間遽波器資訊230係為聲音路徑之資訊,且提供至修正式 空間資訊產生單元220。如果修正式空間資訊產生單元22〇產生包 含環繞音效的組合式空職訊,則可使用此空間紐器資訊。 在下文中,依照修正式空間資訊的類型產生修正式空間資 Λ ’這種方式的音頻訊號解碼方法之解釋依序如下:⑴部分空間 資訊,(2)組合式空間資訊,以及⑶擴充空間資訊。 (1)部分空間資訊 由於二間參數係依照預定樹狀_在乡通道音頻訊號的降混 過程中被用算’因此如果使用完整的m參數以解碼降混音 頻訊號’將可重建降雜_始多猶音頻城。如果嘗試使得 輸出通道音親號㈣目N小於多通道音觀通道數目 Μ ’將能夠藉由應卿分的空間參數以解碼降混音頻訊號。 這種方法可依照編碼裝置中多通道音頻訊號之降混次序以及 I方法而變化,例如,樹狀組態的類型。此外,可使用空間資 代的樹狀組態資訊鶴樹狀組態醜型^這種方法可依照輸出通 201145263 道的數目而變化 的數目。 此外,還能夠使用輸出通道資訊查詢輪出通道 在下文中’#果輸出通道音頻訊號的通道數目小於 頻訊號的通道數目,則可採用一種應用包含部分空間參數之= 空間資訊的音頻峨解碼方法,下文將結合多個樹狀組態二 解釋這種方法。 (1)-1樹狀組態之第一實例(5_2_5樹狀組態) 「第2圖」係為應用部分空間資訊之實施例示意圖。 請參考「第2圖」的左部’圖中表示將多通道音頻訊號降混 為立體降混通道Lo及R〇的次序以及乡通道音触賴空間參數 之間的關係,其中多通道音頻訊號的通道數目為6 (左前通道L、 左環繞通道Ls、中央通道C、低頻通道LFE、右前通道R及右環 繞通道Rs )。 首先’完成左前通道L與左環繞通道Ls之間的降混、中央通 道C與低頻通道LFE之間的降混以及右前通道r與右環繞通道 Rs之間的降混。這個初級降混過程中,產生左總通道Lt、中央總 通道Ct以及右總通道Rt。此外,在此初級降混過程中計算的空間 參數包含CLD2 (包含ICC2)、CLD1 (包含ICC1)及CLD0 (包 含ICC0)等。 接續初級降混過程的二次降混過程中,左總通道Lt、中央總 通道Ct以及右總通道Rt —同降混以產生左通道Lo以及右通道 Ro。第二降混過程中所計算的的空間參數可包含CLDTTT、 201145263 CPCTTT 及 ICCTTT 等。 換言之,具有六個通道之多通道音頻訊號可以上述次序方式 降混產生立體降混之左右通道L〇以及R〇。 如果使用上述次序方式所計算的空間參數(CLD2、CLD1、 CLD0及CLDTTT等)’按照與降混次序相反的次序升混則產生多 通道音頻訊號,其中通道數目為6 (左前通道L、左環繞通道ls、 中央通道C、低頻通道LFE、右前通道R及右環繞通道Rs)。 請參考「第2圖」的右部,如果部分空間資訊對應於空間參 數(CLD2、CLD卜 CLD0 及 CLDTTT 等)中的 CLDTTT,則升 混為左總通道Lt、中央總通道Ct以及右總通道Rt。如果選擇左總 通道Lt以及右總通道Rt作為輸出通道音頻訊號’則能夠產生兩個 通道Lt及Rt的輸出通道音頻訊號。如果選擇左總通道Lt、中央 總通道Ct以及右總通道Rt作為輸出通道音頻訊號,則能夠產生 二個通道Lt、Ct以及Rt的輸出通道音頻訊號。使用附加的空間參 數CLD1完成升混後,如果選擇左總通道Lt、右總通道见、中央 通道C以及低頻通道LFE,則能夠產生四個通道(Lt、Rt、C以 及LFE)的輸出通道音頻訊號。 (1)-2樹狀組態之第二實例(5小5樹狀組態) 「第3圖」係為應用部分空間資訊之另一實例之示意圖。 凊參考「第3圖」的左部’圖中表示將多通道音頻訊號降混 為單一降混音頻訊號Μ的次序以及多通道音頻訊號與空間參數之 間的關係’其中多通道音頻訊號的通道數目為6 (左前通道L、左 12 201145263 環繞通道Ls、中央通道C、低頻通道LFE、右前通道R及右環繞 通道Rs)。 首先,類似於第一實例,完成左前通道L與左環繞通道Ls之 間的降混、中央通道C與低頻通道LFE之間的降混以及右前通道 R與右環繞通道Rs之間的降混。在初級降混過程中,產生左總通 道Lt、中央總通道Ct以及右總通道Rt。初級降混過程中計算的空 間參數包含CLD3 (包含ICC3 )、CLD4 (包含ICC4 )及CLD5 (包 含ICC5)等(在此實例中,CLDx以及ICCx有別於先前第一實例 中的CLDx)。 接續初級降混過程的二次降混過程中,左總通道Lt以及右總 通道Rt—同降混以產生左中央通道lc ;中央總通道ct以及右總 通道Rt —同降混以產生右中央通道^^。二次降混過程中計算的 空間參數可包含CLD2 (包含ICC2 )及CLD1 (包含ICC1)等。 接著,三次降混過程中,降混左中央通道1/:以及右中央通道 RC以產生單一降混音頻訊號M。三次降混過程中計算的空間參數 包含CLD0 (包含ICC0)等。 請參考「第3圖」的右部,如果部分空間資訊對應於空間參 數(CLD3、CLD4、CLD5、CLD1、CLD2 及 CLD0 等)中的 CLD0, 則產生左中央通道LC以及右中央通道狀。如果選擇左中央通道 LC以及右中央通道Rc作為輸出通道音頻訊號,則能夠產生兩個 通道LC及RC的輸出通道音頻訊號。 同時’如果部分空間資訊對應於空間參數(CLD3、CLD4、 13 201145263 CLD5、CLD卜 CLD2 及 CLD0 等)中的 CLDO、CLD1 以及 CLD2, 則產生左總通道Lt、中央總通道Ct以及右總通道Rt。 如果選擇左總通道Lt以及右總通道Rt作為輸出通道音頻訊 號,則肖夠產生兩個通道Lt以及Rt的輸出通道音頻訊號。如果選 擇左總通道Lt、中央總通道Ct以及右總通道Rt作為輸出通道音 頻訊號,則能夠產生三個通道Lt、Ct以及Rt的輸出通道音頻訊號。 如果部分空間資訊尚包含空間參數CLD4,中央通道c以及 低頻通道LFE完成升混後,如果選擇左總通道Lt、右總通道见、 中央通道C以及低頻通道LFE作為輸出通道音頻訊號,則能夠產 生四個通道(左總通道Lt、右總通道Rt、中央通道c以及低頻通 道LFE)的輸出通道音頻訊號。 (1)-3樹狀組態之第三實例(54—5樹狀組態) 第4圖」係為應用部分空間資訊之再一實例之示意圖。 «月參考「第4圖」的左部,圖中表示將多通道音頻訊號降混 為單一降混音頻訊號M的次序以及多通道音頻訊號與空間參數之 間的關係,其中多通道音頻訊號的通道數目為6(左前通道L、左 環繞通道Ls、中央通道c、低頻通道LFE、右前通道R、右環繞 通道Rs)。 類似於第一或者第二實例,首先完成左通道L·與左環繞通道 U之間的降混、中央通道C與低頻通道LFE之間的降混以及右通 道R與右環繞通道rs之間的降混。在初級降混過程中,產生左總 通道Lt、中央總通道Ct以及右總通道Rt。初級降混過程中計算的 201145263 空間參數包含CLD1 (包含ICCl )、CLD2 (包含ICC2)及CLD3 (包含ICC3 )等(在此實例中,CLDx及ICCx有別於先前第一或 第二實例中的CLDx及ICCx )。 接續初級降混過程的二次降混過程中,一同降混左總通道 Lt、中央總通道Ct以及右總通道Rt以產生左中央通道LC以及右 通道R。並且計算了空間參數CLDTTT (包含ICCTTT )。Generated by combining spatial parameters; extended space_messages are generated by information between spatial information. The T-corrected spatial information generating unit 220 is configured to generate the modified spatial information, and the manner of generating the modified spatial information can be etched according to the formalization #__(4). Explain how to generate various types of revised lang information. The tree shape == The reference for determining the modified spatial information may correspond to the spatial information status information and the indicator element may be included in the Weilin, and the heart-splitting group U in the coding device (four) St. Output Channel 201145263 - The speaker is used for a speaker interconnected with the decoding device 200, which may include a number of output channels and location information for each of the ifcit channels. The information of the sampan can be input in advance by the operator = or input by the user. The method of determining the type of corrected spatial information by such information will be explained in detail below. The output channel generating unit 21 of the decoding device 200 uses the modified spatial information s to generate an output channel audio signal 〇υτ N of the downmix audio signal d. The spatial chopper information 230 is information of the sound path and is supplied to the corrected spatial information generating unit 220. If the modified spatial information generating unit 22 generates a combined empty service containing surround sound, the space information can be used. In the following, the interpretation of the audio signal decoding method in accordance with the type of the modified spatial information is explained as follows: (1) partial spatial information, (2) combined spatial information, and (3) expanded spatial information. (1) Partial spatial information is calculated because the two parameters are calculated according to the predetermined tree shape _ in the downmixing process of the audio channel of the home channel. Therefore, if the complete m parameter is used to decode the downmix audio signal, it will be rebuilt. _ Start more than the audio city. If you try to make the output channel number (four) M less than the multi-channel number of channels Μ ', you will be able to decode the downmix audio signal by the spatial parameters of the division. This method can vary depending on the downmixing order of the multichannel audio signals in the encoding device and the I method, for example, the type of tree configuration. In addition, you can use the space-based tree configuration information crane tree configuration ugly type ^ This method can be changed according to the number of output channels 201145263. In addition, it is also possible to use the output channel information to query the round-out channel. In the following, the number of channels of the audio channel of the output channel is smaller than the number of channels of the frequency signal, and an audio decoding method using space information including partial spatial parameters may be used. This method will be explained below in conjunction with a plurality of tree configurations. (1)-1 The first example of the tree configuration (5_2_5 tree configuration) "2" is a schematic diagram of an embodiment of applying partial space information. Please refer to the left part of "Figure 2" for the relationship between the multi-channel audio signal downmixed into the stereo downmixing channels Lo and R〇 and the relationship between the spatial channel and the spatial parameters of the channel, where the multi-channel audio signal The number of channels is 6 (left front channel L, left surround channel Ls, central channel C, low frequency channel LFE, right front channel R, and right surround channel Rs). First, the downmix between the left front channel L and the left surround channel Ls, the downmix between the center channel C and the low frequency channel LFE, and the downmix between the right front channel r and the right surround channel Rs are completed. In this primary downmixing process, the left total channel Lt, the central total channel Ct, and the right total channel Rt are generated. In addition, the spatial parameters calculated during this primary downmixing process include CLD2 (including ICC2), CLD1 (including ICC1), and CLD0 (including ICC0). During the secondary downmixing process of the primary downmixing process, the left total channel Lt, the central total channel Ct, and the right total channel Rt are downmixed to produce a left channel Lo and a right channel Ro. The spatial parameters calculated during the second downmixing process may include CLDTTT, 201145263 CPCTTT, and ICCTTT. In other words, the multi-channel audio signal having six channels can be downmixed in the above-described order to produce the left and right channels L〇 and R〇 of the stereo downmix. If the spatial parameters (CLD2, CLD1, CLD0, and CLDTTT, etc.) calculated using the above-described order mode are mixed up in the reverse order of the downmixing order, a multi-channel audio signal is generated, in which the number of channels is 6 (left front channel L, left surround) Channel ls, central channel C, low frequency channel LFE, right front channel R and right surround channel Rs). Please refer to the right part of "2nd picture". If part of the spatial information corresponds to the CLDTTT in the spatial parameters (CLD2, CLD, CLD0, CLDTTT, etc.), the upmix is the left total channel Lt, the central total channel Ct, and the right total channel. Rt. If the left total channel Lt and the right total channel Rt are selected as the output channel audio signal ', the output channel audio signals of the two channels Lt and Rt can be generated. If the left total channel Lt, the central total channel Ct, and the right total channel Rt are selected as the output channel audio signals, the output channel audio signals of the two channels Lt, Ct, and Rt can be generated. After the upmixing is completed using the additional spatial parameter CLD1, if the left total channel Lt, the right total channel, the central channel C, and the low frequency channel LFE are selected, the output channel audio of the four channels (Lt, Rt, C, and LFE) can be generated. Signal. (1)-2 Second example of tree configuration (5 small 5 tree configuration) "3rd picture" is a schematic diagram of another example of applying partial space information.凊 Refer to the left part of “Figure 3” for the sequence of down-mixing multi-channel audio signals into a single downmix audio signal and the relationship between multi-channel audio signals and spatial parameters. The number of channels is 6 (left front channel L, left 12 201145263 surround channel Ls, central channel C, low frequency channel LFE, right front channel R, and right surround channel Rs). First, similar to the first example, the downmix between the left front channel L and the left surround channel Ls, the downmix between the center channel C and the low frequency channel LFE, and the downmix between the right front channel R and the right surround channel Rs are completed. In the primary downmixing process, the left total channel Lt, the central total channel Ct, and the right total channel Rt are generated. The spatial parameters calculated during the primary downmix process include CLD3 (including ICC3), CLD4 (including ICC4), and CLD5 (including ICC5) (in this example, CLDx and ICCx are different from CLDx in the previous first example). During the secondary downmixing process of the primary downmixing process, the left total channel Lt and the right total channel Rt are downmixed to produce the left central channel lc; the central total channel ct and the right total channel Rt are downmixed to produce the right center Channel ^^. The spatial parameters calculated during the second downmixing process may include CLD2 (including ICC2) and CLD1 (including ICC1). Next, during the three downmixing processes, the left central channel 1/: and the right central channel RC are downmixed to produce a single downmix audio signal M. The spatial parameters calculated during the three downmixing process include CLD0 (including ICC0) and so on. Please refer to the right part of "3rd picture". If part of the spatial information corresponds to CLD0 in the spatial parameters (CLD3, CLD4, CLD5, CLD1, CLD2, CLD0, etc.), the left central channel LC and the right central channel are generated. If the left central channel LC and the right central channel Rc are selected as the output channel audio signals, the output channel audio signals of the two channels LC and RC can be generated. Meanwhile, if part of the spatial information corresponds to CLDO, CLD1, and CLD2 in the spatial parameters (CLD3, CLD4, 13 201145263 CLD5, CLD, CLD2, and CLD0, etc.), the left total channel Lt, the central total channel Ct, and the right total channel Rt are generated. . If the left total channel Lt and the right total channel Rt are selected as the output channel audio signals, then Xiao can generate the output channel audio signals of the two channels Lt and Rt. If the left total channel Lt, the central total channel Ct, and the right total channel Rt are selected as the output channel audio signals, the output channel audio signals of the three channels Lt, Ct, and Rt can be generated. If part of the spatial information still includes the spatial parameter CLD4, after the central channel c and the low frequency channel LFE are upmixed, if the left total channel Lt, the right total channel see, the central channel C, and the low frequency channel LFE are selected as the output channel audio signals, it can be generated. Output channel audio signal for four channels (left total channel Lt, right total channel Rt, central channel c, and low frequency channel LFE). (1)-3 The third example of the tree configuration (54-5 tree configuration) Fig. 4 is a schematic diagram of another example of applying partial space information. «Monthly reference to the left part of Figure 4, which shows the sequence of downmixing multi-channel audio signals into a single downmix audio signal M and the relationship between multi-channel audio signals and spatial parameters, among which multi-channel audio signals The number of channels is 6 (left front channel L, left surround channel Ls, central channel c, low frequency channel LFE, right front channel R, right surround channel Rs). Similar to the first or second example, first, the downmix between the left channel L· and the left surround channel U, the downmix between the center channel C and the low frequency channel LFE, and the right channel R and the right surround channel rs are completed. Downmix. In the primary downmixing process, the left total channel Lt, the central total channel Ct, and the right total channel Rt are generated. The 201145263 spatial parameters calculated during the primary downmixing process include CLD1 (including ICl), CLD2 (including ICC2), and CLD3 (including ICC3), etc. (In this example, CLDx and ICCx are different from those in the previous first or second example. CLDx and ICCx). In the second downmixing process following the primary downmixing process, the left total channel Lt, the central total channel Ct, and the right total channel Rt are downmixed together to produce a left central channel LC and a right channel R. And the spatial parameter CLDTTT (including ICCTTT) is calculated.

順序地’三次降混過程中,降混左中央通道LC以及右通道R 以產生單一降混音頻訊號Μ。並且計算出空間參數CLD0 (包含 ICC0 )。 請參考「第4圖」的右部’如果部分空間資訊對應於空間參 數(CLD 卜 CLD2、CLD3、CLDTTT 及 CLD0 等)中的 CLD0 以 及CLDTTT,將產生左總通道Lt、中央總通道ct以及右總通道 Rt〇 如果選擇左總通道Lt以及右總通道Rt作為輸出通道音頻訊 號,則能夠產生兩個通道Lt以及Rt的輸出通道音頻訊號。 如果選擇左總通道Lt、中央總通道Ct以及右總通道Rt作為 輸出通道音頻訊號’則能夠產生三個通道Lt、Ct以及Rt的輸出通 道音頻訊號。 如果部分空間資訊尚包含空間參數CLD2,中央通道C以及 低頻通道LFE完成升混後,如果選擇左總通道Lt、右總通道Rt、 中央通道C以及低頻通道LFE作為輪出通道音頻訊號,則能夠產 生四個通道(左總通道Lt、右總通道、中央通道匚以及低頻通 15 201145263 道LFE)的輪出通道音頻訊號。 上面的描述中,解釋了應用部分的空間參數產生輸出通道音 頻訊號_程’但是讎賴三麵狀_作為㈣。此外,像 部分空間資訊—樣,還能夠應用組合式空間資訊或者延伸空間資 Λ如此f+於應用修正式空間資訊至音頻訊號的過程,則能夠 Ρ皆層式地或者組合式地且合成式地加以控制。 (2)組合式空間資訊 由於二間資訊係依照預定的樹狀組態在多通道音頻訊號的降 混過程中計算得出’因此如果按__方式朗空間資訊的空 間參數以解碼降混音頻減,射重建降混前的初始多通道音頻 訊號。如果多通道音頻訊號的通道數目Μ不同於輸出通道音頻訊 號的通C數目Ν,透過組合空間資訊產生新的纟 1合式空間資訊, 然後則可使用產生的資訊升混此降混音頻訊號。尤其是,應用空 間參數至轉換公^,則能夠產生組合式空間參數。 這種方法可依照編碼裝置巾多通道音頻訊號的降混次序及方 法而變化。並且,㈣制空間資訊之樹狀組態資訊查詢降混的 人序及方’去這種方法可依照輪出通道的數目而變化。此夕卜,還 此夠使用輸出通道資訊查詢輸出通道的數目等。 下面的描述中’將解釋修正空間資訊的方法的詳細實施例以 及獲得虛擬3-D效果的實施例。 (2)-1 —般組合式空間資訊 本發明提供-種藉由組合空間f訊之空間參數而產生組合式 201145263 二間參數之方法,用以根據不同於降混過程的樹狀組態進行升混 過程。所以,無論樹狀組態資訊係根據何種樹狀組態,這種方法 均可應用於各種降混音頻訊號。 如果多通道音頻訊號為5.1通道且降混音頻訊號為丨通道(單 通道)’則結合以下兩種實例解釋兩個通道的輸出通道音頻訊號 的產生方法。 (2)-M樹狀組態之第四實施例(5+51樹狀組態) 「第5圖」係為應用組合式空間資訊之實例之示意圖。 晴參考「第5圖」的左部,CLD0至CLD4以及ICC0至ICC4 (圖中未表示)可稱為空間參數,係在5.1通道的多通道音頻訊號 之降混過程中計算獲得。例如,空間參數中,左通道訊號L以及 右通道訊號R之間的通道間位階差值為CLD3,L與R之間的通 道間相關為ICC3。左環繞通道Ls以及右環繞通道Rs之間的通道 間位階差值為CLD2,Ls以及RS之間的通道間相關為ICC2。 另一方面,請參考「第5圖」的右部,如果應用組合式空間 參數CLDa及ICOx至單一降混音頻訊號m,產生左通道訊號Lt 及右通道訊號Rt,則能夠從單一通道音頻訊號m中直接地產生立 體輸出通道音頻訊號Lt及Rt。這個實例中,透過組合空間參數 CLD0至CLD4以及ICC0至ICC4,可計算出組合式空間參數CLDa 及 ICCa。 下面,首先解釋透過組合空間參數CLD0至CLD4以計算出 組合式空間參數中CLDa的過程,然後解釋透過組合空間參數 5 17 201145263 CLD0至CLD4以及ICC0至ICC4以計算出組合式空間參數中 ICCa的過程。 (2)-l-l-a空間參數CLDa的推導 首先,由於空間參數CLDa係為左輸出訊號Lt及右輸出訊號 Rt之間的位階差值,輸入左輸出訊號Lt及右輸出訊號Rt至空間 參數CLD的定義公式之結果如下所示: [公式1] CLDa= 10*logl0(PLt/PRt) 其中PLt係為左輸出訊號Lt的功率,PRt係為右輸出訊號Rt 的功率。 [公式2] CLDa= 10*logl0(PLt+a/PRt+a) 其中PLt係為左輸出訊號Lt的功率,PRt係為右輸出訊號Rt 的功率,‘a’係為非常小的一個常數。 因此,空間參數CLD α由公式1或者公式2定義。 同時’為了使用空間參數CLD0至CLD4表示PLt及PRt,則 需要輸出通道音頻1峨的左輸出峨Lt、触通道音頻訊號的右 輸出訊號Rt與多通道訊號L、Ls、R、Rs、c以及啦之間的關 係公式。對應的關係式可定義如下: [公式3]In the sequential 'three downmixing process, the left central channel LC and the right channel R are downmixed to produce a single downmix audio signal. And calculate the spatial parameter CLD0 (including ICC0). Please refer to the right part of "Fig. 4". If part of the spatial information corresponds to CLD0 and CLDTTT in the spatial parameters (CLD, CLD2, CLD3, CLDTTT, CLD0, etc.), the left total channel Lt, the central total channel ct, and the right will be generated. Total channel Rt〇 If the left total channel Lt and the right total channel Rt are selected as the output channel audio signals, the output channel audio signals of the two channels Lt and Rt can be generated. If the left total channel Lt, the central total channel Ct, and the right total channel Rt are selected as the output channel audio signals', the output channel audio signals of the three channels Lt, Ct, and Rt can be generated. If part of the spatial information still includes the spatial parameter CLD2, after the center channel C and the low frequency channel LFE are upmixed, if the left total channel Lt, the right total channel Rt, the central channel C, and the low frequency channel LFE are selected as the round-out channel audio signals, The round-out channel audio signal is generated for four channels (left total channel Lt, right total channel, central channel 匚, and low-frequency pass 15 201145263 LFE). In the above description, it is explained that the spatial parameters of the application section generate an output channel audio signal _程' but rely on the three-sided _ as (4). In addition, as part of the spatial information, it is also possible to apply the combined spatial information or the extended space resource so that the process of applying the modified spatial information to the audio signal can be layered or combined and synthesized. Control it. (2) Combined spatial information, because the two information systems are calculated in the downmixing process of multi-channel audio signals according to the predetermined tree configuration. Therefore, if the spatial parameters of the spatial information are compressed by __ method, the downmixing is decoded. Frequency reduction, shot reconstruction of the initial multi-channel audio signal before downmixing. If the number of channels of the multi-channel audio signal is different from the number of C channels of the audio signal of the output channel, a new space information is generated by combining the spatial information, and then the downmix audio signal can be upmixed using the generated information. In particular, applying spatial parameters to the conversion public can generate combined spatial parameters. This method can be varied in accordance with the downmixing order and method of the multi-channel audio signal of the encoding device. Moreover, (4) the tree-shaped configuration information of the spatial information query, the descending order of the person and the square, can vary according to the number of round-out channels. In addition, it is enough to use the output channel information to query the number of output channels, and the like. In the following description, a detailed embodiment of a method of correcting spatial information and an embodiment of obtaining a virtual 3-D effect will be explained. (2)-1 General Combination Space Information The present invention provides a method for generating two parameters of the combined 201145263 by combining the spatial parameters of the space f signal for performing a tree configuration different from the downmix process. Upmixing process. Therefore, regardless of the tree configuration, the tree configuration information can be applied to various downmix audio signals. If the multi-channel audio signal is 5.1 channels and the downmix audio signal is 丨 channel (single channel), the following two examples are used to explain the output signal of the output channels of the two channels. (2) Fourth embodiment of the -M tree configuration (5+51 tree configuration) "5th picture" is a schematic diagram of an example of applying combined spatial information. The left side of the "5th figure", CLD0 to CLD4 and ICC0 to ICC4 (not shown) can be called spatial parameters, which are calculated in the downmixing process of the 5.1 channel multi-channel audio signal. For example, in the spatial parameter, the channel-to-channel difference between the left channel signal L and the right channel signal R is CLD3, and the channel-to-channel correlation between L and R is ICC3. The channel difference between the left surround channel Ls and the right surround channel Rs is CLD2, and the channel-to-channel correlation between Ls and RS is ICC2. On the other hand, please refer to the right part of "5th picture". If the combined spatial parameters CLDa and ICOx are applied to the single downmix audio signal m, the left channel signal Lt and the right channel signal Rt are generated, and the audio can be from a single channel. The stereo output channel audio signals Lt and Rt are directly generated in the signal m. In this example, the combined spatial parameters CLDa and ICCa can be calculated by combining the spatial parameters CLD0 to CLD4 and ICC0 to ICC4. In the following, the process of calculating the CLDa in the combined spatial parameters by combining the spatial parameters CLD0 to CLD4 is first explained, and then the process of calculating the ICCa in the combined spatial parameters by combining the spatial parameters 5 17 201145263 CLD0 to CLD4 and ICC0 to ICC4 is explained. . (2) Derivation of the -lla spatial parameter CLDa First, since the spatial parameter CLDa is the difference between the left output signal Lt and the right output signal Rt, the definition of the input left output signal Lt and the right output signal Rt to the spatial parameter CLD The result of the formula is as follows: [Formula 1] CLDa = 10*logl0(PLt/PRt) where PLt is the power of the left output signal Lt and PRt is the power of the right output signal Rt. [Formula 2] CLDa = 10 * logl0 (PLt + a / PRt + a) where PLt is the power of the left output signal Lt, PRt is the power of the right output signal Rt, and 'a' is a very small constant. Therefore, the spatial parameter CLD α is defined by Equation 1 or Equation 2. At the same time, in order to use the spatial parameters CLD0 to CLD4 to represent PLt and PRt, it is necessary to output the left output 峨Lt of the channel audio 1 右, the right output signal Rt of the touch channel audio signal and the multi-channel signals L, Ls, R, Rs, c and The relationship formula between the two. The corresponding relationship can be defined as follows: [Formula 3]

Lt = L + Ls + C/V2 + LFE/^2 ^ = R + Rs + C/V2 + LFE/V2 201145263 的關係式可依照如何定義料通道音頻訊 因為類似於公式3Lt = L + Ls + C / V2 + LFE / ^ 2 ^ = R + Rs + C / V2 + LFE / V2 201145263 The relationship can be defined according to how to define the material channel audio because it is similar to formula 3

號而變化,所以也可科同於公式3的公式定義。例如,⑽或 者LFE/W中的可為‘〇’或者‘1’。 S 公式3可導出如下的公式4 : [公式4] PLt = PL + PLs + PC/2 + PLFE/2 PRt = PR + PRs + PC/2 + PLFE/2 依照公式1或者公式2,能夠使用PLt及pRt表示空間參數 CLDα。依照公式4 ’可使用PL、PLs、PC、PLFE、PR以及PRS 表示PLt以及PRt。所以,需要找到一個關係式,能夠使用空間 參數 CLD0 至 CLD4 表示 PL、PLs、PC、PLFE、PR 以及 PRs。 同時,如果樹狀組態如「第5圖」所示,多通道音頻訊號(l、 R、C、LFE、Ls及RS)與單一降混通道訊號功之間的關係如下 所示: [公式5] 'L ' ciomci〇mcu〇TT〇 R dr c2,omc\,orr\c\y〇Tro C Dc •ft — Cl,OTT4 C2 ,〇m C\ ,〇7TO LFE DlFE ffl 一 C2,O7T4C2,OTT\Ch〇TT0 Ls C\,OTT2C2.O7T0 Rs . C2,OTT2C2,OTTQ _ 其中, )11+101° , vi+io 10 公式5可導出如下的公式6 : 19 201145263 其中The number changes, so it can also be defined in the formula of Equation 3. For example, (10) or LFE/W may be '〇' or '1'. S Equation 3 can be derived as Equation 4 below: [Formula 4] PLt = PL + PLs + PC/2 + PLFE/2 PRt = PR + PRs + PC/2 + PLFE/2 According to Equation 1 or Equation 2, PLt can be used. And pRt represents the spatial parameter CLDα. PLt and PRt can be expressed in accordance with Equation 4' using PL, PLs, PC, PLFE, PR, and PRS. Therefore, you need to find a relationship that can use the spatial parameters CLD0 to CLD4 to represent PL, PLs, PC, PLFE, PR, and PRs. At the same time, if the tree configuration is as shown in Figure 5, the relationship between the multi-channel audio signals (l, R, C, LFE, Ls, and RS) and the signal of the single downmix channel is as follows: 5] 'L ' ciomci〇mcu〇TT〇R dr c2,omc\,orr\c\y〇Tro C Dc •ft — Cl,OTT4 C2 ,〇m C\ ,〇7TO LFE DlFE ffl a C2,O7T4C2, OTT\Ch〇TT0 Ls C\, OTT2C2.O7T0 Rs . C2, OTT2C2, OTTQ _ where, ) 11+101° , vi+io 10 Equation 5 can be derived as Equation 6: 19 201145263 where

CXCSTx 1 1 + 10 10 'Pl ' (Cl,0773^1,0771^1,OTTO ) Pr (C2t07T3Cl(07TlCl,07T〇) Pc (C\tOTT4C2tOTT\CU〇7TO ) Plfe (C2,O7T4C2,O7TlCl,O7T0 ) Pu (C1,07T2C2,07T〇) (C2,07T2C2,〇7T〇) — 'σττ> 特別是,透過輸入公式6至公式4,並且輸入公式4至公式j 或者公式2,則能夠藉由組合空間參數CLDO至CLD4的方式表示 組合式空間參數CLDct。 同時’輸入公式6至公式4的PC/2 + PLFE/2的擴充結果如公 式7所示: [公式7] PC/2 + PLFE/2 = [(cl,OTT4)2 + (c2,〇TT4)2] *(c2,OTTl*cl,OTTO)2 * m2 /2 這個實例中,依照cl及c2的定義(請參考公式5),因為(cl,x)2 + (c2,x)2 =卜得出(cl,OTT4)2 + (c2,OTT4)2 =卜 因此,公式7可簡化如下: [公式8] PC/2 + PLFE/2 = (c2,OTTl*cl,OTTO)2 * m2/2 因此,輸入公式8以及公式6至公式4,並且輸入公式4至公 式1,則能夠透過組合空間參數0030至CLD4的方式表示組合式 空間參數CLDa ° 201145263 (2)-l-l-b通道間相關ICCa的導出 首先’因為通道間相關ICCa係為左輸出訊號Lt與右輸出訊 號Rt之間的相關,輸入左輸出訊號Lt與右輸出訊號至對應的 定義公式之結果如下所示: [公式9] ICCa= : ▽. ,其中。 公式9中,PLt及PRt可使用公式4、公式6以及公式8中的 空間參數CLD0至CLD4表示。PLtPRt可藉由公式10的方式展開。 [公式10] PLtRt = PLR + PLsRs + PC/2 + PLFE/2 公式10中,‘PC/2 + PLFE/2’可依照公式6表示為空間參數 CLD0至CLD4。PLR及PLsRs可依照ICC的定義展開如下: [公式11] ICC,P ㈣ 公式11中,如果移項或者^^,則得到公式12。 [公式12] plr=icc、* 诉FrCXCSTx 1 1 + 10 10 'Pl ' (Cl,0773^1,0771^1,OTTO ) Pr (C2t07T3Cl(07TlCl,07T〇) Pc (C\tOTT4C2tOTT\CU〇7TO ) Plfe (C2,O7T4C2,O7TlCl,O7T0 ) Pu (C1,07T2C2,07T〇) (C2,07T2C2,〇7T〇) — 'σττ> In particular, by inputting Equation 6 to Equation 4, and entering Equation 4 to Equation j or Equation 2, it is possible to combine The spatial parameter CLDO to CLD4 represents the combined spatial parameter CLDct. At the same time, the expansion result of PC/2 + PLFE/2 of Equation 6 to Equation 4 is as shown in Equation 7: [Formula 7] PC/2 + PLFE/2 = [(cl, OTT4)2 + (c2, 〇 TT4) 2] *(c2, OTTl*cl, OTTO)2 * m2 /2 In this example, according to the definition of cl and c2 (please refer to Equation 5), because (cl,x)2 + (c2,x)2 =b gives (cl,OTT4)2 + (c2,OTT4)2 =b Therefore, Equation 7 can be simplified as follows: [Formula 8] PC/2 + PLFE/ 2 = (c2, OTTl*cl, OTTO)2 * m2/2 Therefore, by inputting Equation 8 and Equation 6 to Equation 4, and inputting Equation 4 to Equation 1, it is possible to express the combination by combining the spatial parameters 0030 to CLD4. Spatial parameters CLDa ° 201145263 (2)-llb channel-related ICCa export First, because the inter-channel correlation ICCa is the correlation between the left output signal Lt and the right output signal Rt, the result of inputting the left output signal Lt and the right output signal to the corresponding definition formula is as follows: [Equation 9] ICCa= :公式. , where. In Equation 9, PLt and PRt can be expressed using the spatial parameters CLD0 to CLD4 in Equation 4, Equation 6, and Equation 8. PLtPRt can be expanded by the method of Equation 10. [Formula 10] PLtRt = PLR + PLsRs + PC/2 + PLFE/2 In Equation 10, 'PC/2 + PLFE/2' can be expressed as spatial parameters CLD0 to CLD4 according to Equation 6. PLR and PLsRs can be expanded as follows according to the definition of ICC: [Equation 11] ICC, P (4) In Equation 11, if the term or ^^ is shifted, Equation 12 is obtained. [Formula 12] plr=icc,* v Fr

PlsRs = ICC ^PijPrs 公式12中,PL、PR、PLs以及PRs可依照公式6表示為空間 參數CLD0至CLD4。輸入公式6至公式12的結果對應於公式13。 21 201145263 [公式B] PLR=ICC3 *cl,OTT3 *c2,OTT3 *(cl,OTTl*cl,OTTO)2 *m2 PLsRs=ICC2 *cl,OTT2 *c2,OTT2 *(c2,OTTO)2 *m2 總之,輸入公式6以及公式13至公式10,輸入公式以及 公式4至公式9,則能夠以空間參數CLD0至CLD3、ICC2以及 ICC3表示組合式空間參數iCCa。 (2)-1-2樹狀組態之第五實施例(5-1-52樹狀組態) 「第6圖」係為應用組合式空間資訊之另一實例之示意圖。 請參考「第6圖」的左部,CLD0至CLD4以及ICC0至ICC4 (圖中未表示)可稱為空間參數’其可在51通道的多通道音頻訊 號的降混過程中用於計算。 空間參數中,左通道訊號L與左環繞通道訊號Ls之間的通道 間位階差值為CLD3,左通道訊號L與左環繞通道訊號Ls之間的 通道間相關為ICC3。右通道R與右環繞通道rs之間的通道間位 階差值為CLD4,R與Rs之間的通道間相關為ICC4。 另一方面’請參考「第6圖」的右部’如果透過制組合式 空間參數〇Χ)β以及iccp至單-降混音頻訊號m,產生左通道訊 號Lt以及右通道訊號Rt,則能夠直接地從單一降混音頻訊號瓜 中產生立體輸出通道音頻訊號以及扮。這個實例中,可藉由組合 空間參數CLD0至CLD4以及ICC〇至ICC4計算出組合式空間參 數CLD0以及κχβ。 下面,首先解釋透過組合空間參數CLD〇至CLD4以計算出 S; 22 201145263 組合式空間參數中的CLDP的過程,然後解釋透過組合空間參數 CLD0至CLD4以及ICC0至ICC4以計算出組合式空間參數中的 Ι(Χβ的過程。 (2)-l_2-a空間參數CLDp的推導 首先,由於空間參數CLDP係為左輸出訊號Lt與右輸出訊號 Rt之間的位階差值,輸入左輸出訊號Lt與右輸出訊號玢至口;^ 的定義公式的結果如下所示: [公式14] CLDP= 10*logl0(PLt/PRt) 其中’ PLt係為左輸出訊號Lt的功率,PRt係為右輸出訊號 Rt的功率。 [公式15] CLDP= 10*logl0(PLt+a/PRt+a) 其中’ PLt係為左輸出§凡號Lt的功率,PRt係為右輸出訊號 Rt的功率,V係為非常小的數字。 因此,空間參數〇Χ)β由公式14或者公式15定義。 同時,為了使用空間參數CLD0至CLD4表示PLt以及PRt, 需要輸出通道音頻訊號的左輸出訊號Lt、輸出通道音頻訊號的右 輸出訊號财與多通道訊號:、1_(8、尺、118、(3以及1^丑之間的關 係式•對應的關係式可定義如下: [公式16]PlsRs = ICC ^PijPrs In Equation 12, PL, PR, PLs, and PRs can be expressed as spatial parameters CLD0 to CLD4 according to Equation 6. The result of inputting Equation 6 to Formula 12 corresponds to Formula 13. 21 201145263 [Formula B] PLR=ICC3 *cl,OTT3 *c2,OTT3 *(cl,OTTl*cl,OTTO)2 *m2 PLsRs=ICC2 *cl,OTT2 *c2,OTT2 *(c2,OTTO)2 *m2 In summary, by inputting Equation 6 and Equations 13 to 10, inputting the formula, and Equations 4 to 9, the combined spatial parameter iCCa can be expressed by the spatial parameters CLD0 to CLD3, ICC2, and ICC3. (5) The fifth embodiment of the -1-2 tree configuration (5-1-52 tree configuration) "Fig. 6" is a schematic diagram of another example of applying the combined spatial information. Please refer to the left part of "Picture 6". CLD0 to CLD4 and ICC0 to ICC4 (not shown) can be called spatial parameters' which can be used for calculation in the downmixing process of 51 channels of multichannel audio signals. In the spatial parameter, the channel-to-channel difference between the left channel signal L and the left surround channel signal Ls is CLD3, and the channel-to-channel correlation between the left channel signal L and the left surround channel signal Ls is ICC3. The channel-to-channel difference between the right channel R and the right surround channel rs is CLD4, and the channel-to-channel correlation between R and Rs is ICC4. On the other hand, please refer to the right part of 'Fig. 6', if the combined spatial parameter β) and iccp to the single-downmix audio signal m, the left channel signal Lt and the right channel signal Rt are generated. The stereo output channel audio signal and the appearance can be directly generated from a single downmix audio signal. In this example, the combined spatial parameters CLD0 and κχβ can be calculated by combining the spatial parameters CLD0 to CLD4 and ICC〇 to ICC4. In the following, first, the process of calculating the CLDP in the combined spatial parameters by calculating the spatial parameters CLD〇 to CLD4 is explained, and then interpreting the combined spatial parameters CLD0 to CLD4 and ICC0 to ICC4 to calculate the combined spatial parameters. The process of (Χ) (2)-l_2-a spatial parameter CLDp First, since the spatial parameter CLDP is the difference between the left output signal Lt and the right output signal Rt, the left output signal Lt and the right are input. The output signal 玢 to the mouth; the result of the definition formula of ^ is as follows: [Formula 14] CLDP = 10*logl0(PLt/PRt) where 'PLt is the power of the left output signal Lt, and PRt is the right output signal Rt [Equation 15] CLDP = 10*logl0(PLt+a/PRt+a) where 'PLt is the power of the left output § Lt, and PRt is the power of the right output signal Rt, the V system is very small Therefore, the spatial parameter 〇Χ) β is defined by Equation 14 or Equation 15. At the same time, in order to use the spatial parameters CLD0 to CLD4 to represent PLt and PRt, it is necessary to output the left output signal Lt of the channel audio signal, the right output signal of the output channel audio signal and the multi-channel signal: 1, 1 (8, ft, 118, (3) And the relationship between 1^ ugly • The corresponding relation can be defined as follows: [Formula 16]

Lt = L + Ls + C/V2 + LFE/V2 23 201145263Lt = L + Ls + C/V2 + LFE/V2 23 201145263

Rt = R + Rs + C/V2 +LFE/V2 因為類似於公式16的關係式可依照如何定義輸出通道音步員 號而變化,所以還可以用不同於公式16的方式定義。例 或者LFE/W中的‘1/石,可為‘〇’或者‘r。 公式16可導出如下的公式17 : [公式17] PLt = PL + PLs + PC/2 + PLFE/2 PRt = PR + PRS + PC/2 + PLFE/2 依照公式14或者公式15,能夠使用PLt以及PRt表示空間參 數CLDp。依照公式15,可使用PL、PLs、PC、PLFE、PR以及 pRs表示pLt以及PRt。所以,需要找到一個關係式,能夠使用空 間參數 CLD0 至 CLD4 表示 PL、PLs、Pc、PLFE、PR 以及 PRs。 同時’如果樹狀組態如「第6圖」所示,則多通道音頻訊號 、R、c、LFE、Ls、Rs)與單一降混通道訊號m之間的關係 如下所示: [公式18;)Rt = R + Rs + C/V2 + LFE/V2 Since the relationship similar to Equation 16 can be varied depending on how the output channel tone step number is defined, it can also be defined in a different manner than Equation 16. For example, ‘1/stone in LFE/W can be ‘〇’ or ‘r. Equation 16 can derive the following formula 17: [Formula 17] PLt = PL + PLs + PC/2 + PLFE/2 PRt = PR + PRS + PC/2 + PLFE/2 According to Equation 14 or Equation 15, PLt can be used as well PRt represents the spatial parameter CLDp. According to Equation 15, pLt and PRt can be expressed using PL, PLs, PC, PLFE, PR, and pRs. Therefore, it is necessary to find a relation that can express PL, PLs, Pc, PLFE, PR, and PRs using the spatial parameters CLD0 to CLD4. At the same time, if the tree configuration is as shown in Figure 6, the relationship between the multi-channel audio signal, R, c, LFE, Ls, Rs) and the single downmix channel signal m is as follows: [Equation 18 ;)

LL

LsLs

RR

RsRs

CC

LFE CU〇mC\tOTr\C\,OTTQ Du C2,OTn>C\,OTr\C\,OTTO C\,O7T4^2,OTT\Ch〇TT0 m - C2,Oir4C2,OTT\C\,OTTO Dc C\tOTT2C2,OTTO _ C2,O772C2,O7T0 _ ΓΧ&ν 10~LFE CU〇mC\tOTr\C\, OTTQ Du C2, OTn>C\, OTr\C\, OTTO C\, O7T4^2, OTT\Ch〇TT0 m - C2, Oir4C2, OTT\C\, OTTO Dc C\tOTT2C2, OTTO _ C2, O772C2, O7T0 _ ΓΧ & ν 10~

CLDX 10CLDX 10

其中 1 + 101 + 10 of them

1 + 10^ 201145263 公式18導出如下的公式19 : [公式19] -PL- (cu〇mcwmcu〇Tr〇) Pu (C2,07T3C1,07T1C1,〇7T〇) Pr (C\y〇TT4C2f〇rr\C\ ,ΟΤΤΟ ) (C2,OTT4C2,OTT\C\,〇TTO ) Pc iC\,OTT2C2t〇rr〇^ J*LFE- (C2,OTT2C2,O7T0^ _ ίο 10 [~~ϊ ^·σΓΓ* = \ ϋΓ 气吨=-5ΞΓ 其中 ΙΙι+ηΠ7- , Vi+ioi-。 尤其地,輸入公式19至公式π,並且輸入公式π至公式14 或者公式15,則能夠透過組合空間參數CLD0至CLD4的方式表 示組合式空間參數CLDP。 同時,輸入公式19至公式π中的PL + PLs,得到的展開公 式如公式20所示: [公式20] PL + PLs = [(cl,OTT3)2 + (c2,OTT3)2] (cl,OTTl*cl,〇TTO)2 *m2 這個實例中,依照cl以及C2的定義(比較公式5),由於(ci,x)2 + (c2,x)2 =1,所以得到(ci,〇TT3)2 + (c2,〇TT3)2 = 1。 所以’公式20可簡化如下: [公式21] PL_ = PL + PLs = (cl,OTTl*cl,OTTO)2 *m2 另一方面,輸入公式19至公式17中的PR+ PRs,得到的展 25 201145263 開公式如公式22所示: [公式22] PR + PRs = [(cl,〇TT4)2 + (c2,OTT4)2] (cl,OTTl*cl,OTTO)2 *m2 這個實例中’依照Cl以及C2的定義(請參考公式5),由於 (cl,x)2 + (c2,x)2=l,所以得到(cl,〇几4)2+ (c2,OTT4)2=l。 所以’公式22可簡化如下。 [公式23] PR_ = PR + PRs =(c2,OTTl*cl,〇TTO)2 *m2 另一方面,輸入公式19至公式π的PC/2 + PLFE/2,展開公 式的結果如公式24所示: [公式24] PC/2 + PLFE/2 = [(cl,〇TT2)2 + (c2,OTT2)2] (C2,OTTO)2 *m2 /2 這個實例中’依照cl以及C2的定義(請參考公式5),由於 (cl,x)2 + (c2,x)2 =1,所以得到(ci,〇tx2)2 + (c2,〇TT2)2 = 1。 所以,公式24可簡化如下: [公式25] PC/2 + PLFE/2 = (c2,OTTO)2 *m2 /2 因此’輸入公式21、公式23以及公式25至公式π,並且輸 入公式17至公式14或者公式15,則能夠透過組合空間參數CLD0 至CLD4的方式表示組合式空間參數CLDp。 26 a 201145263 (2)-l-2-b空間參數ICCp的推導 首先,由於空間參數ICCp係為左輸出訊號Lt與右輸出訊號 Rt之間的通道間相關,輸入左輸出訊號Lt與右輸出訊號Rt至對 應的定義公式的結果如下: [公式26] ICC = _ A ,其中 。 公式26中,依照公式19,可使用空間參數CLD0至CLD4 表示PLt以及PRt。PLtPRt可藉由公式27的方式展開如^ : [公式27] PLtRt = PL_R_ + PC/2 + PLFE/2 公式27中,‘PC/2 + PLFE/2’可依照公式9表示為空間參數 CLD0至CLD4。PL_R_可依照ICC的定義展開如下: [公式28] ICC, = PL R "(PL PR ) 如果移項,則得到公式29 : [公式29] 公式29中,依照公式21以及公式23,PL一以及ρι^可表示 為空間參數CLD0至CLD4。輸入公式21以及公式23至公式29 得到的公式對應公式30 :1 + 10^ 201145263 Equation 18 derives the following formula 19: [Formula 19] -PL- (cu〇mcwmcu〇Tr〇) Pu (C2,07T3C1,07T1C1,〇7T〇) Pr (C\y〇TT4C2f〇rr\ C\ ,ΟΤΤΟ ) (C2, OTT4C2, OTT\C\, 〇TTO) Pc iC\, OTT2C2t〇rr〇^ J*LFE- (C2, OTT2C2, O7T0^ _ ίο 10 [~~ϊ ^·σΓΓ* = \ ϋΓ ton = -5 ΞΓ where ΙΙι+ηΠ7- , Vi+ioi-. In particular, enter the formula 19 to the formula π, and enter the formula π to the formula 14 or the formula 15, then the way to combine the spatial parameters CLD0 to CLD4 Indicates the combined spatial parameter CLDP. At the same time, enter formula 19 to PL + PLs in the formula π, and the resulting expansion formula is as shown in Equation 20: [Formula 20] PL + PLs = [(cl, OTT3) 2 + (c2, OTT3)2] (cl, OTTl*cl, 〇TTO)2 *m2 In this example, according to the definition of cl and C2 (compare formula 5), since (ci, x) 2 + (c2, x) 2 =1, So get (ci, 〇 TT3) 2 + (c2, 〇 TT3) 2 = 1. So 'Formula 20 can be simplified as follows: [Formula 21] PL_ = PL + PLs = (cl, OTTl*cl, OTTO)2 *m2 On the other hand, enter PR + PRs in Equation 19 to Equation 17, and get the exhibition 25 201145263 As shown in Equation 22: [Formula 22] PR + PRs = [(cl, 〇 TT4) 2 + (c2, OTT4) 2] (cl, OTTl * cl, OTTO) 2 * m2 In this example 'According to Cl and C2 The definition (please refer to Equation 5), since (cl, x) 2 + (c2, x) 2 = l, so (cl, 〇 4) 2+ (c2, OTT4) 2 = l. So 'Formula 22 It can be simplified as follows: [Formula 23] PR_ = PR + PRs = (c2, OTTl * cl, 〇 TTO) 2 * m2 On the other hand, enter Equation 19 to PC/2 + PLFE/2 of the formula π, and expand the result of the formula As shown in Equation 24: [Formula 24] PC/2 + PLFE/2 = [(cl, 〇 TT2) 2 + (c2, OTT2) 2] (C2, OTTO) 2 * m2 /2 In this example 'according to cl And the definition of C2 (please refer to Equation 5). Since (cl,x)2 + (c2,x)2 =1, we get (ci,〇tx2)2 + (c2,〇TT2)2 = 1. Therefore, Equation 24 can be simplified as follows: [Formula 25] PC/2 + PLFE/2 = (c2, OTTO) 2 * m2 /2 Therefore 'Enter Equation 21, Equation 23, and Equation 25 to Equation π, and enter Equation 17 to Equation 14 or Equation 15 can express the combined spatial parameter CLDp by combining the spatial parameters CLD0 to CLD4. 26 a 201145263 (2)-l-2-b Spatial parameter ICCp derivation First, since the spatial parameter ICCp is the channel-to-channel correlation between the left output signal Lt and the right output signal Rt, the left output signal Lt and the right output signal are input. The result of Rt to the corresponding definition formula is as follows: [Formula 26] ICC = _ A , where. In Equation 26, according to Equation 19, the spatial parameters CLD0 to CLD4 can be used to represent PLt and PRt. PLtPRt can be expanded by the method of Equation 27 as: [Formula 27] PLtRt = PL_R_ + PC/2 + PLFE/2 In Equation 27, 'PC/2 + PLFE/2' can be expressed as the spatial parameter CLD0 according to Equation 9. CLD4. PL_R_ can be developed as follows according to the definition of ICC: [Equation 28] ICC, = PL R "(PL PR ) If the term is shifted, Equation 29 is obtained: [Formula 29] In Equation 29, according to Equation 21 and Equation 23, PL And ρι^ can be expressed as spatial parameters CLD0 to CLD4. The formula obtained by inputting the formula 21 and the formula 23 to the formula 29 corresponds to the formula 30:

[公式30J 27 201145263 PL-R- = ICC1 *cl,0TTl *cl,〇TT0 *c2,OTTl *cl,OTTO *m2 總之’輸入公式30至公式27,並且輸入公式27以及公式Π 至公式26,則能夠以空間參數CLD0至CLD4以及ICC1表示組 合式空間參數iccp。 ±$解__參數.正方法僅僅為—個實補。得到ρχ 或者Pxy的過程巾,透過考慮各通道咖及附加的減能量之間 的通道間_ (例如ICC()等),顯紅面轉的公式可變化為多 種形式。 (2)-2具有環繞效果的組合式空間資訊 首先,如果考慮聲音路徑,透過組合空間資訊以產生組合式 空間資訊’則能夠產生虛擬的環繞效果。 虛擬的環繞效果或者虛擬的3D效果能夠產生下述效果,在沒 有%繞通道的揚聲器的條件下產生實質上存在有環繞通道的揚聲 器效果。例如,5.1通道音頻訊號透過兩個立體揚聲器而輸出。 聲音路徑可對應於空間濾波器資訊。空間濾波器資訊能夠使 用一種名為頭部相關轉移函數(head-related transfer fimeticm ; HRTF)之函數,但本發明並非限制於此。空間濾波器資訊可包含 濾波器參數。輸入濾波器參數以及空間參數至轉換公式,則能夠 產生組合式空間參數。產生的組合式空間參數可包含級器係數。 下文中,假設多通道音頻訊號為5通道,並且產生三通道的 輸出通道音頻訊號,一種考慮聲音路徑以產生具有環繞效果的組 合式空間資訊的方法解釋如下。 S: 28 201145263 「第7圖」係為揚聲器至收聽者的聲音路徑的示意圖,圖中 顯示出揚聲器的所在位置。 請參考「第7圖」,三個揚聲器SPK1、SPK2以及SPK3的位 置分別位於左前方L、中央C以及右前方R。虛擬環繞通道的位 置分別為左環繞Ls以及右環繞RS。 圖中分別表示三個揚聲器的位置L、C及R以及虛擬環繞通 道的位置Ls及Rs到收聽者的左右耳的位置r及1的聲音路徑。 ‘Gx_y’表示位置x到位置y的聲音路徑。例如‘GL_r’表示左前位置 L到收聽者的右耳位置Γ的聲音路徑。 如果五個位置存在揚聲器(例如,揚聲器也處於左環繞Ls以 及右環繞Rs的位置),並且收聽者處於「第γ圖」所示的位置, 則進入收聽者左耳的訊號L0以及進入收聽者右耳的訊號R〇可由 公式31表示: [公式31] L0 = L*GL_1 + C*GC J + R*GR_1 + Ls*GLs_l + Rs*GRs_l R0= L*GL一r + C*GC_r + R*GR_r + Ls*GLs_r + Rs*GRs_r, 其中L、C、R、Ls以及RS分別為各位置的通道,Gx_y表示 位置x到位置y的聲音路徑,表示卷積。 但是’如上所述’如果僅三個位置L、c以及R存在揚聲器, 則進入收聽者左耳的訊號L0_real以及進入收聽者右耳的訊號 R0_real分別表示如下: [公式32] 29 201145263 LOrea 卜 L*GLJ + C*GC—1 + R*GR 1 RO_real = L*GL_r + C*GC—r + R*GR r 由於公式32所示的訊號並沒有考慮到環繞通道訊號Ls以及 Rs ’所以無法產生虛擬環繞效果。^ 了產生虛擬環繞效果,可使 自揚聲器所在位置Ls’到達收聽者位置〇, r)的左環繞通道訊號 Ls等於自不同於初始位£Ls’之各揚聲器L、c以及R所在位置到 達收聽者位置(1,〇社環繞通道域Ls。也可躺於右環 繞通道訊號RS的例子中。 進一步地深究左環繞通道訊號Ls,如果左環繞通道訊號Ls 係輸出自以左環繞位置Ls,作為初始位置的揚聲器則到達收聽者 左右耳1以及r的訊號分別表示如下: [公式33] ‘Ls*GLs_l’,‘Ls*GLs_r, 如果右環繞通道訊號RS輸出自以右環繞位置Rs,作為初始位 置的揚聲H,則到達收聽者左右耳丨以及r的訊號分別表示如下: [公式34] <Rs*GRs一1’,‘Rs*GRs r, 如果到達收聽者左右耳丨以及Γ的訊號等於公式33以及公式 4的分量,即使經由任何位置的揚聲器(例如經由左前方的揚聲 器SPK1)輸出,收聽者皆能夠感覺到彷彿揚聲器係分別位於左右 環繞位置Ls,以及Rs,。 同時,如果公式33所示的分量輸出自左環繞位置Ls的揚聲 30 -[Formula 30J 27 201145263 PL-R- = ICC1 *cl, 0TTl *cl, 〇TT0 *c2, OTTl *cl, OTTO *m2 In summary 'Enter Equation 30 to Equation 27, and enter Equation 27 and Equation Π to Equation 26, The combined spatial parameter iccp can then be represented by spatial parameters CLD0 to CLD4 and ICC1. The ±$ solution __ parameter. The positive method is just a real complement. The process towel of ρχ or Pxy can be changed into a plurality of forms by considering the inter-channel _ (for example, ICC(), etc.) between each channel and the additional energy reduction. (2)-2 Combined spatial information with surround effects First, if a sound path is considered, a virtual surround effect can be generated by combining spatial information to generate combined spatial information. A virtual surround effect or a virtual 3D effect can produce the effect of a speaker that substantially has a surround channel without the speaker surrounding the channel. For example, a 5.1 channel audio signal is output through two stereo speakers. The sound path may correspond to spatial filter information. The spatial filter information can use a function called head-related transfer fimeticm (HRTF), but the invention is not limited thereto. Spatial filter information can include filter parameters. By inputting filter parameters and spatial parameters to the conversion formula, combined spatial parameters can be generated. The resulting combined spatial parameters may include level coefficients. In the following, assuming that the multi-channel audio signal is 5 channels and generating a three-channel output channel audio signal, a method of considering the sound path to produce a combined spatial information with a surround effect is explained below. S: 28 201145263 "Figure 7" is a schematic diagram of the sound path from the speaker to the listener, showing the location of the speaker. Please refer to Figure 7. The positions of the three speakers SPK1, SPK2 and SPK3 are located at the left front L, the center C and the right front R, respectively. The positions of the virtual surround channels are left surround Ls and right surround RS, respectively. The figure shows the positions L, C and R of the three speakers and the positions Ls and Rs of the virtual surround channels to the positions r and 1 of the listener's left and right ears, respectively. 'Gx_y' represents the sound path from position x to position y. For example, 'GL_r' represents the sound path from the left front position L to the listener's right ear position Γ. If the speaker is present at five locations (for example, the speaker is also in the position of the left surround Ls and the right surround Rs), and the listener is in the position shown in the "γ map", the signal L0 entering the listener's left ear and entering the listener are entered. The signal R〇 of the right ear can be expressed by Equation 31: [Equation 31] L0 = L*GL_1 + C*GC J + R*GR_1 + Ls*GLs_l + Rs*GRs_l R0= L*GL-r + C*GC_r + R *GR_r + Ls*GLs_r + Rs*GRs_r, where L, C, R, Ls, and RS are the channels of each position, and Gx_y represents the sound path from position x to position y, indicating convolution. However, as described above, if only three positions L, c, and R have speakers, the signal L0_real entering the listener's left ear and the signal R0_real entering the listener's right ear are respectively expressed as follows: [Formula 32] 29 201145263 LOrea *GLJ + C*GC—1 + R*GR 1 RO_real = L*GL_r + C*GC—r + R*GR r Since the signal shown in Equation 32 does not take into account the surround channel signals Ls and Rs ', it cannot be generated. Virtual surround effect. ^ The virtual surround effect is generated so that the position Ls' from the speaker reaches the listener position 〇, r) the left surround channel signal Ls is equal to the position of each speaker L, c and R different from the initial position £Ls' Location (1, 〇 环绕 surround channel domain Ls. Can also lie in the example of the right surround channel signal RS. Further study the left surround channel signal Ls, if the left surround channel signal Ls is output from the left surround position Ls, as The initial position of the speaker reaches the listener's left and right ears 1 and r, respectively, as follows: [Formula 33] 'Ls*GLs_l', 'Ls*GLs_r, if the right surround channel signal RS is output from the right surround position Rs, as an initial The position of the speaker H, the signal to the listener's left and right ear and r, respectively, is expressed as follows: [Formula 34] <Rs*GRs-1', 'Rs*GRs r, if the listener reaches the left and right ear and the squat The signal is equal to the components of Equation 33 and Equation 4. Even if the speaker is output via any position of the speaker (for example, via the left front speaker SPK1), the listener can feel as if the speaker system is located respectively. Surround position Ls, and Rs, at the same time, if the component shown in Equation 33 is output from the left surround position Ls, the speaker 30 -

S 201145263 器’則其分別為到達收聽者左右耳丨以及r的訊號。所以,如果公 式33所示的分量完整輸出自左前位置的揚聲器SPKi,則到達收 聽者的左右耳1以及r的職可分職示如下: [公式35]S 201145263 ' is the signal to reach the listener's left and right ear and r. Therefore, if the component shown in the formula 33 is completely output from the speaker SPKi at the left front position, the positions of the left and right ears 1 and r reaching the listener are as follows: [Form 35]

Ls GLs l*GL一1’ ’ ‘Ls*GLs一r*GL_r, 深究公式35 ’對應於自左前位置L到收聽者的左耳1 (或者 右耳Γ)的聲音路徑之分量说_1,(。1*证_〇則被增加。 到達收^者的左右耳1以及Γ的訊號應該為公式33而非公式 35所:分量。如果左前位置[處的揚聲器輪出的聲音到達收聽 GL'n ° ,33 二^自左_ L的_ SPK1,_音路徑中應該考慮 刀I —(或者‘GL、r,)的反函數ή,(或者“证―^,)。換古 之,如果對應於公式33的分量輸出自左前位置聲^ 則應該修正為如下的公式: SPK1 [公式36]Ls GLs l*GL-1' 'Ls*GLs-r*GL_r, the formula 35' corresponds to the component of the sound path from the left front position L to the listener's left ear 1 (or right ear). (.1*证_〇 is added. The signal to the left and right ears 1 and Γ of the receiver should be the formula 33 instead of the formula 35: component. If the left front position [the sound of the speaker at the ring reaches the listening GL' n ° , 33 2 ^ From the left _ L _ SPK1, _ sound path should consider the inverse function 刀 of the knife I — (or 'GL, r,), (or “certificate ^ ^,). The component output corresponding to Equation 33 from the left front position should be corrected to the following formula: SPK1 [Formula 36]

Ls GLsJ*GL_>i’,‘Ls*GLS-r*GL_r_r 的揚聲器Ls GLsJ*GL_>i’, ‘Ls*GLS-r*GL_r_r speakers

如果對應於公式34的分量輸出自左前仇置L SPK1,職該修正為如下的公式: [公式37]If the component corresponding to the formula 34 is output from the left front hatching L SPK1, the correction is as follows: [Formula 37]

Rs GRsJ*glJ.p , ^Rs*GRs r*GL_l-r •所以,輸出自左前位置L的揚聲器SPK1的訊號l,可總括如 31 201145263 [公式38] L’=L + Ls*GLs」*GL」-1+rs*gRs 1*GL 1-1 — _ (省略分置 Ls*GLs_r*GL_j-l 以及 Rs*GRs_r*GL_l-l。) 如果公式38所示之訊號係輸出自左前位置L的揚聲器spK1 而到達收聽者的左耳的位置卜聲音路翻數‘GLJ,則被增加。 所以,公式38中的‘GL—1’則互相抵消,最終保留公式%以及公 式34所示的因數。 「第8圖」所稀為_自各揚聲驗置輪出的訊號以產生 虛擬環繞效果之示意圖。 月參考第8圖」’如果考慮聲音路挺,輪出自環繞位置a 以及Rs,的環繞通道訊仙以及Rs包含於輸出自揚聲器腦位 置的訊號L’中,則對應於公式%。 公式38中,GLsJ^簡寫為hls上因此可表示如下: [公式39]Rs GRsJ*glJ.p , ^Rs*GRs r*GL_l-r • Therefore, the signal l of the speaker SPK1 output from the left front position L can be summarized as 31 201145263 [Formula 38] L'=L + Ls*GLs”* GL"-1+rs*gRs 1*GL 1-1 — _ (Omit Ls*GLs_r*GL_j-l and Rs*GRs_r*GL_l-l.) If the signal shown in Equation 38 is output from the left front position L The speaker spK1 arrives at the listener's left ear position and the sound path counts 'GLJ' is increased. Therefore, 'GL-1' in Equation 38 cancels each other, and finally retains the formula % and the factor shown by Formula 34. The "Fig. 8" is sparsely _ the signal from the sound detectors to generate a virtual surround effect. Referring to Fig. 8"", if the sound path is considered to be good, the surround channel from the surrounding positions a and Rs, and the Rs are included in the signal L' output from the brain position of the speaker, corresponding to the formula %. In Equation 38, GLsJ^ is abbreviated as hls and can therefore be expressed as follows: [Formula 39]

V= L + Ls*HLs_L + Rs*HRs_L 例如’中央位置C處的揚聲HSPK2輸㈣峨C,總括如下: [公式40]V= L + Ls*HLs_L + Rs*HRs_L For example, the speaker HSPK2 at the central position C is transmitted (four) 峨C, which is summarized as follows: [Formula 40]

C’= C . Ls*HLs—C + Rs*HRsC 例如’右前位置R處的揚聲HSPK3輸叫訊號r,總括如下。 [公式41]C'= C . Ls*HLs—C + Rs*HRsC For example, the speaker HSPK3 input signal r at the right front position R is summarized as follows. [Formula 41]

R - R + Ls*HLs_R + Rs*HRs_R 「第9圖」聯係級_條公式%、公式39或者公式 201145263 40的5通道訊號以產生3通道訊號的方法之概念亓·意圖。 如果使用5通道訊號產生2通道訊號R’以及L’,或者中央通 道訊號C’中不包含環繞通道訊號Ls或Rs,則HLs_C或HRs c 變為0。 為了便於實現,Hx_y可透過多種方式修正,ϋ如以Gx_y取 代Hx_y,或者使用Hx_y時考慮串音。 上述詳細解釋係關於具有環繞效果的組合式空間資訊之實 例。顯然可依照空間濾波器資訊的應用方法而變化。如前所述, 依照上述過程經由揚聲器(上述實例中,左前通道L,、右前通道 R’以及中央通道C’)輸出的訊號可透過組合式空墦資訊產生自降 混音頻訊號,其中尤其是使用了組合式空間參數。 (3)擴充空間資訊 首先,透過增加延伸空間資訊至空間資訊,則能夠產生擴充 空間資訊。然後則可使用延伸空間資訊升混音頻訊號。對應的升 混過程中’基於空間資訊’音頻訊號被轉換為初級升混音頻訊號, 然後基於延伸空間資訊,初級升混音頻訊號則被轉換為二次升混 音頻訊號。 這個實例中,延伸空間資訊能夠包含延伸通道組態資訊、延 伸通道對映資訊以及延伸空間參數。 延伸通道組態資訊係為用於可配置通道的資訊,通道可藉由 空間資訊的樹狀組態資訊而組態。延伸通道組態資訊可包含劃分 識別碼以及非劃分識別碼至少其中之一,下面將詳細解釋。延伸 33 201145263 通道對映資訊係為配驗伸通道的每鹏道的位置資訊。延伸空 間參數可升混-個通道轻少兩個财。延輕間參數可= 含通道間位階差值。 上面轉的延伸郎錄經由編碼裝置⑴或者解碼裝置本 身(ϋ)產生後,可能包含於空間資訊中。如果延伸空間資訊由編 碼裝置產生,_伸_#狀存在或者不存在可取決於空間資 訊的指示元。如果延伸空間資訊由解碼裝置本身產生,則延伸空 間-貝§fl之延伸空間參數可使用空間資訊之空間參數計算得出。 同時,利用基於空間資訊以及延伸空間資訊所產生的擴充空 間資訊以升混音頻訊號的過程可被連續地且階層式地或者組合式 地且合成式地執行。如果基於如資誠及延輕間資訊計算擴 充空間資訊得到-她陣,則可使耻轉組合式地且直接地將 一降混音頻訊號升混為多通道音頻訊號。這個實例中,可依照空 間參數以及延伸空間參數定義組態矩陣的因數。 下面,透過編碼裝置產生延伸空間資訊的實例解釋完成後, 將解釋透過解碼裝置本身產生延伸空間資訊的實例。 (3) ·1透過編碼裝置產生延伸空間資訊的實例:任意樹狀組 態 首先,透過增加延伸空間資訊至空間資訊,編碼裝置產生擴 充空間資訊。然後解釋解碼裝置接收此延伸空間資訊之實例。此 外,延伸空間資訊可擷取於編碼裝置降混多通道音頻訊號的過程 中。R - R + Ls*HLs_R + Rs*HRs_R "Figure 9" Contact level _ formula %, formula 39 or formula 201145263 40 5-channel signal to generate a 3-channel signal concept 意图 Intention. If the 5-channel signal R' and L' are generated using the 5-channel signal, or the surround channel signal Ls or Rs is not included in the central channel signal C', HLs_C or HRsc becomes 0. For ease of implementation, Hx_y can be modified in a number of ways, such as replacing Hx_y with Gx_y or considering crosstalk when using Hx_y. The above detailed explanation is an example of combined spatial information with a surround effect. Obviously, it can be changed according to the application method of spatial filter information. As described above, the signal output through the speaker (the left front channel L, the right front channel R', and the central channel C' in the above example) can generate a self-downmix audio signal through the combined air information according to the above process, especially The combined spatial parameters are used. (3) Expanding spatial information First, by adding extended space information to spatial information, expanded spatial information can be generated. Then you can use the extended space information to upmix the audio signal. In the corresponding upmixing process, the 'space-based information' audio signal is converted into a primary upmix audio signal, and then based on the extended space information, the primary upmix audio signal is converted into a second upmix audio signal. In this example, the extended space information can include extended channel configuration information, extended channel mapping information, and extended spatial parameters. The extended channel configuration information is information for the configurable channel, and the channel can be configured by the tree configuration information of the spatial information. The extended channel configuration information may include at least one of a division identification code and a non-partition identification code, which will be explained in detail below. Extension 33 201145263 The channel mapping information is the location information of each passway of the test extension channel. The extended space parameter can be upmixed - one channel is lighter and less expensive. The parameter between delays can be = with the difference between the channels. The extended extension recorded above may be included in the spatial information after being generated by the encoding device (1) or the decoding device itself (ϋ). If the extended space information is generated by the encoding device, the presence or absence of the extension may depend on the indicator of the spatial information. If the extended spatial information is generated by the decoding device itself, the extended space parameter of the extended space - §fl can be calculated using the spatial parameters of the spatial information. At the same time, the process of upmixing the audio signals using the expanded spatial information generated based on the spatial information and the extended spatial information can be performed continuously and hierarchically or in combination and in a synthetic manner. If the space is calculated based on the information of the resources such as Zicheng and Yanliang, the video can be combined and directly mixed into a multi-channel audio signal. In this example, the factors of the configuration matrix can be defined in terms of spatial parameters and extended space parameters. In the following, after the example interpretation of the extended space information by the encoding device is completed, an example of generating extended space information through the decoding device itself will be explained. (3) Example of generating extended space information by the encoding device: arbitrary tree configuration First, by adding extended space information to spatial information, the encoding device generates expanded spatial information. An example of the decoding device receiving this extended space information is then explained. In addition, the extended space information can be taken from the process of downmixing the multi-channel audio signal by the encoding device.

34 S 201145263 如前所述,延伸空間資訊包含延伸通道組態資訊、延伸通道 對映資訊以及延伸空間參數。這個實例中,延伸通道組態資訊可 能包含劃分識別碼以及非劃分識別碼至少其中之一。下文中,將 詳細解釋根據劃分以及非劃分識別碼的陣列以配置延伸通道的過 程。 「第ίο圖」係為基於延伸通道組態資訊以配置延伸通道之實 施例示意圖。 請參考「第10圖」的下端,〇以及1係重複地排列成一序列。 這個實例中’ G表示非劃分識別碼,丨表示劃分識別碼。非劃分識 別碼0位於第-順位⑴,匹配於第—順位的非劃分識別碼的通 道係為左通道L,處於最上端。所以,匹配於非劃分識別碼的左 通道L被選擇作為輸出通道,而非_分。第二順位⑵中,存 在劃分識別碼1。匹配於劃分識別碼的通道係為左環繞通道Ls, 鄰接於左通道L。所以,眺_分姻碼1的左環繞通道Ls被 劃分為兩個通道。 班 ' 、)以及第四順位⑷中存在非劃分識別碼〇, 左環繞通道Lst彳分㈣_通道被完整地卿作為輸出通道,而 卜被^刀上述過程重複到最後的順位(10),則能夠配置出整個 延伸通道。 m複通道劃分過程的次數等於劃分識別碼1的數目,並且重 1二所通道作為輸出通道的過程次數等於非劃分識㈣〇的數 。以’通道分單S 以及AT1的數目等於劃分識別碼! 35 201145263 的數目⑵,延伸通道(L、Lfs、Ls、R、Rfs、Rs、c^LFE) 的數目等於非劃分識別碼〇的數目⑻。 同時延伸通道經過配置後,則能夠使用延伸通道對映資訊 對映每個輸出通道的位置。「第1G圖」的實例巾,完成對映的順 序如下左削通道L、左前側通道Lfs、左環繞通道Ls、右前通道 R右别側通道Rfs、右環繞通道Rs、中央通道C以及低頻通道 LFE。 如上所述’可基於延伸通道組態資訊以配置延伸通道。因此, 需要劃分-個通道為至少兩個通道的通_分單元。劃分一個通 道為至少兩個通道時,通分單元可使祕伸空間參數。因為 延伸工間參數的數目等於通道齡單元的數目,所以也等於劃分 削馬的數目。因此’延伸空間參數賴取次數可等於劃分識別 碼的數目。 第11圖」係為解釋「第1〇圖」所示之延伸通道組態以及 與延伸空間參數之關係示意圖。 "月參考「第11圖」,應用於兩個通道劃分單元ΑΤ0以及ATI 上的延伸空間參數ATDQ以及ATm分別如圖所示。 如果延伸空間參數係為通道間位階差值,通道劃分單元則能 夠使用延伸空間參數決定兩個劃分通道的位階。 因此,透過增加延伸空間參數以完成升混的過程中,延伸空 間參數不需要完全被應用,*是部分被應用。 (3)-2產生延伸空間資訊的實例:内插法/外插法34 S 201145263 As mentioned earlier, the extended space information includes extended channel configuration information, extended channel mapping information, and extended space parameters. In this example, the extended channel configuration information may include at least one of a partition identification code and a non-partition identification code. Hereinafter, the process of configuring the extended channel according to the division and the array of non-partitioned identification codes will be explained in detail. "Fig. ίο" is a schematic diagram of an embodiment based on configuration information of an extended channel to configure an extended channel. Please refer to the lower end of "Picture 10", and the 〇 and 1 series are repeatedly arranged in a sequence. In this example, 'G denotes a non-divided identification code, and 丨 denotes a division identification code. The non-divided identification code 0 is located at the first-order (1), and the channel matching the non-partitioned identification code of the first-order is the left channel L, which is at the uppermost end. Therefore, the left channel L matching the non-divided identification code is selected as the output channel instead of _. In the second order (2), there is a division identification code 1. The channel matched to the division identification code is the left surround channel Ls adjacent to the left channel L. Therefore, the left surround channel Ls of the 眺_marriage code 1 is divided into two channels. There is a non-divided identification code 班 in the class ', ' and the fourth order (4), and the left surround channel Lst ( (4) _ channel is completely cleared as the output channel, and the above process is repeated to the last order (10). Then the entire extension channel can be configured. The number of m-channel division processes is equal to the number of division identification codes 1, and the number of processes that emphasizes the channel as the output channel is equal to the number of non-partitions (four)〇. The number of channels and the number of AT1 is equal to the division ID! 35 The number of 201145263 (2), the number of extended channels (L, Lfs, Ls, R, Rfs, Rs, c^LFE) is equal to the number of non-partitioned identification codes ( (8). At the same time, the extended channel is configured to map the position of each output channel using the extended channel mapping information. For the example towel of "1G", the order of mapping is as follows: left clipping channel L, left front channel Lfs, left surround channel Ls, right front channel R right side channel Rfs, right surround channel Rs, central channel C, and low frequency channel LFE. As described above, information can be configured based on the extended channel to configure the extended channel. Therefore, it is necessary to divide the channel into a pass_minute unit of at least two channels. When dividing a channel into at least two channels, the pass-by unit can make a secret space parameter. Since the number of extension work parameters is equal to the number of channel age units, it is also equal to the number of divided horses. Therefore, the number of times the extension space parameter is taken may be equal to the number of division identification codes. Figure 11 is a schematic diagram showing the relationship between the configuration of the extended channel and the parameters of the extended space as shown in the "Figure 1". "Monthly reference "Figure 11", applied to the two channel division unit ΑΤ0 and the extended space parameters ATDQ and ATm on ATI are shown in the figure. If the extended spatial parameter is the inter-channel step difference, the channel dividing unit can determine the order of the two divided channels using the extended space parameter. Therefore, by adding the extended space parameter to complete the upmixing process, the extended space parameter does not need to be fully applied, and * is partially applied. (3)-2 Example of generating extended space information: interpolation/extrapolation

36 S 201145263 扣首先,藉由增加延伸空間資訊至空間資訊,將能夠產生擴充 空間資訊。τ面將解釋使用空間資訊產生延伸空間資訊的實例。 尤其疋’使用空間資訊之空間參數’能夠產生延伸空間資訊。整 個實例中’可使軸插法、外插法或其_似方法。 (3)-2-1延伸至6.1通道 如果多通道音頻訊號為5.1通道,則結合如下的實例解釋產生 6.1通道的輸出通道音頻訊號的情況。 「第12圖」係為5.丨通道的多通道音頻訊號的位置以及61 通道的輸出通道音頻訊號的位置之示意圖。 請參考「第12A圖」’可以看到5>1通道的多通道音頻訊號的 通道位置分別為左前通道L、右前通道R、中央通道c、低頻通道 (圖中未表示)LFE、左環繞通道Ls以及右環繞通道Rs。 在5.1通道的多通道音頻訊號係為降混音頻訊號情況下,如果 空間參數被應用至降混音頻訊號,則此降混音頻訊號將再次被升 混為5.1通道的多通道音頻訊號。 然而’如「第12B圖」所示之後中央RC的通道訊號應該更 進一步地被產生’以將降混音頻訊號升混為6丨通道的多通道音頻 訊號。 可使用兩個後通道(左環繞通道LS以及右環繞通道RS )相關 的空間參數產生後中央RC的通道訊號。尤其是,空間參數中的通 道間位階差值(CLD)指示兩個通道之間的位階差值。所以,透 過調整兩個通道之間的位階差值,則能夠改變兩個通道之間的虛 37 201145263 擬聲音源的位置。 虛擬聲音源依照兩個通道之間的位置差值而變化的原則解釋 如下。 第13圖」係為虛擬聲音源位置與兩個通道間的位階差值之 間的關係示意圖,其中左右環繞通道Ls以及以的位階分別為‘a, 以及‘b’。 請參考「第13A圖」,在左環繞通道Ls的位階a大於右環繞 通道Rs的位階b之情況下,則可看出虛擬聲音源vs的位置較接 近左環繞通道Ls的位置而非右環繞通道Rs的位置。 如果音頻訊號係輸出自兩個通道,收聽者則感覺到虛擬聲音 源vs實質上存在於兩個通道之間。這個實例中,虛擬聲音源vs 的位置較接近具有較高位階的通道位置。 在第13B圖」的情況下’由於左環繞通道ls的位階a幾乎 等於右環繞通道RS的位階b,因此收聽者感覺虛擬聲音源vs的 位置處於左ϊ衣繞通道Ls以及右環繞通道Rs之間的中央位置。 因此,能夠使用上述原則決定後中央位置的位階。 第14圖」係為兩個後通道的位階以及一個後中央通道的位 階之不意圖。 請參考「第14圖」,透過内插左環繞通道Ls的位階a以及右 環繞通道RS的位階b之間的差值,能夠計算出後中央通道尺(:的 位階c。這個實例中,計算時可使用非線性内插法或者線性内插法。 透過下面的公式,可根據線性内插法計算出兩個通道(例如36 S 201145263 First, by adding extended space information to spatial information, it will be able to generate expanded spatial information. The τ plane will explain an example of using spatial information to generate extended spatial information. In particular, the 'space parameter using spatial information' can generate extended space information. In the entire example, 'axis interpolation, extrapolation, or a similar method can be used. (3) -2-1 extends to 6.1 channel If the multi-channel audio signal is 5.1 channel, the case of generating the output channel audio signal of 6.1 channel is explained by the following example. Figure 12 is a schematic diagram of the location of the multichannel audio signal of the channel and the location of the audio signal of the output channel of the 61 channel. Please refer to "Fig. 12A" to see that the channel position of the multichannel audio signal of channel 1 is left front channel L, right front channel R, central channel c, low frequency channel (not shown) LFE, left surround channel Ls and right surround channel Rs. In the case where the 5.1 channel multi-channel audio signal is a downmix audio signal, if the spatial parameter is applied to the downmix audio signal, the downmix audio signal will be upmixed again into a 5.1 channel multi-channel audio signal. . However, the channel signal of the central RC should be further generated after the "Fig. 12B" is shown to 'mix the downmix audio signal into a multichannel audio signal of 6 channels. The channel parameters associated with the rear central RC can be generated using the spatial parameters associated with the two rear channels (left surround channel LS and right surround channel RS). In particular, the inter-channel level difference (CLD) in the spatial parameter indicates the step difference between the two channels. Therefore, by adjusting the step difference between the two channels, it is possible to change the position of the virtual sound source between the two channels. The principle that the virtual sound source varies according to the position difference between the two channels is explained below. Fig. 13 is a diagram showing the relationship between the position of the virtual sound source and the step difference between the two channels, wherein the left and right surround channels Ls and the steps are 'a, and 'b', respectively. Please refer to "Fig. 13A". In the case where the level a of the left surround channel Ls is larger than the level b of the right surround channel Rs, it can be seen that the position of the virtual sound source vs is closer to the position of the left surround channel Ls than the right surround The location of the channel Rs. If the audio signal is output from two channels, the listener perceives that the virtual sound source vs essentially exists between the two channels. In this example, the virtual sound source vs is located closer to the channel position with a higher level. In the case of Fig. 13B", since the level a of the left surround channel ls is almost equal to the level b of the right surround channel RS, the listener perceives that the position of the virtual sound source vs is in the left and right surround channels Ls and the right surround channel Rs. Central location between the two. Therefore, the above-described principle can be used to determine the level of the rear central position. Figure 14 is the intention of the order of the two rear channels and the order of one rear central channel. Please refer to "Fig. 14". By interpolating the difference between the level a of the left surround channel Ls and the level b of the right surround channel RS, the post-central channel ruler (: level c can be calculated. In this example, the calculation Non-linear interpolation or linear interpolation can be used. The following equations can be used to calculate two channels based on linear interpolation (for example)

38 S 20114526338 S 201145263

Ls以及Rs)之間的新通道(例如後中央通道尺匸)位階c : [公式40] c = a*k + b*(l-k) 其令’ a以及‘b’各自為兩個通道的位階,‘k’係為位階&的 通道、位階b的通道以及位階c的通道之間的相對位置。 如果位階e的通道(例如後中央通道RC)位於位階&的通道 (例如Ls)以及位階b的通道以之間的中央位置,則‘k,為〇5。 如果‘k’為〇_5,則公式40將變為公式41。 [公式41] c = (a + b)/2 依照公式41,如果位階c的通道(例如後中央通道Rc)位於 位階a的通道(例如Ls)以及位階b的通道Rs之間的中央位置, 新通道的位階c則對應於之前通道的位階的平均值。此外, 么式40以及公式41僅是具有代表性的。因此,也可能重新調整 位階c的結果以及位階a及位階b的值。 (3)-2-2延伸至7.1通道 、备夕通道音頻訊號為5.1通道時,下面將解釋嘗試產生通 道的輸出通道音頻訊號的實例。 第b圖」係為解釋5.1通道的多通道音頻訊號的位 置以及 7.1通道的輸出通道音頻訊號的位置之示意圖。 請參考「第15A圖」,類似於「第12A圖」,可看出5」通道 的多通道音頻訊號的位置通道各自為左前通道L、右前通道r、中 39 201145263 央通道C、低頻通道(圖中未表示)LFE、左環繞通道Ls以及右 環繞通道Rs。 在5,1通道的多通道音頻訊號係為降混音頻訊號之情況下,如 果空間參數被應用至此降混音頻訊號,則此降混音頻訊號將再次 被升混為5.1通道的多通道音頻訊號。 然而,如「第15B圖」所示之左前側通道Lfs以及右前側通 道Rfs,更應該被產生以將降混音頻訊號升混為7丨通道的多通道 音頻訊號。 因為左前侧通道Lfs位於左前通道L與左環繞通道Ls之間, 則能夠使用左前通道L的位階以及左環繞通道Ls的位階透過内插 法決定左前側通道Lfs的位階。 「第16圖」係為解釋兩個左通道的位階以及一個左前側通道 (Lfs)的位階之示意圖。 請參考「第16圖」,可看出,左前侧通道Lfs的位階c係為基 於左前通道L的位階a以及左環繞通道Ls的位階b的線性内插值。 同時,雖然左前侧通道Lfs位於左前通道l以及左環繞通道 Ls之間,但是也可位於左前通道L、中央通道匚以及右前通道R 的外面。所以,可使用左前通道L、中央通道C以及右前通道R 的位階透過外插法決定左前側通道Lfs的位階。 「第17圖」係為解釋三個前通道的位階以及一個左前侧通道 的位階之示意圖。 请參考「第17圖」’如圖所示可以看出左前側通道Lfs的位階 201145263 d係為基於左前通道L的位階a、中央通道C的位階c以及右前通 道R的位階b的線性外插值。使用延伸空間資訊的升混過程中, 延伸的空間參數可不需要完全應用,而是部分應用。因此,應用 空間參數至音頻訊號的過程可順序地且階層式地或者組合式地且 合成式地執行。 工業應用 因此,本發明提供以下效果。 首先,本發明能夠產生一種音頻訊號,包含的組態不同於預 定的樹狀組態,從而產生各種組態的音頻訊號。 第二,因為產生的音頻訊號包含的組態可不同於預定樹狀組 態,即使執行降混前的多通道的數目小於或者大於揚聲器的數 目,仍然能夠自降混音頻訊號產生具有等於揚聲器數目的輪出通 道數目。 第二,在產生具有數目小於多通道數目的輸出通道之情況 下,由於多通道音頻訊號係直接地產生自降混音頻訊號,而非降 混來自多通道音頻訊號的輸出通道音頻訊號,其中多通道音頻訊 號係由升混-降混音舰麟產生,因此㈣頗為減少解碼音頻 訊號所需要的作業量。 第四’因為產纽合式郎資訊時需要考慮聲纽徑,本發 明可在無法實現環繞通道輸出條件下提供一種偽環繞效果。 雖然本發明以前述之較佳實施例揭露如上,並非用以限 定本發明。本領域之技術人員應當意識到在不脫離本發明所附之 201145263 申請專利範圍所揭示之本發明之精神和範圍的情況下,所作之更 動與潤飾’均屬本發明之專利保護範圍之内。 【圖式簡單說明】 第1圖所示係為本發明之音頻訊號編碼裝置以及音頻訊號解 碼裝置之方塊圖; 第2圖所示係為應用部分空間資訊之實施例之方塊圖; 第3圖所示係為應用部分空間資訊之另一實施例之方塊圖; 第4圖所示係為應用部分空間資訊之再一實施例之方塊圖; 第5圖所示係為應用組合空間資訊之實施例之方塊圖; 第6圖所示係為應用組合空間資訊之另一實施例之方塊圖; 第7圖所示係為揚聲器至收聽者的聲音路徑之示意圖,圖中 表示揚聲器的位置; 第8圖所示係為用於產生環繞效果的揚聲器的每個位置輸出 訊號之示意圖; 第9圖所示料制5通道喊產生3通道職的方法的概 念方塊圖; 第10嶋為基於延伸通道城¥誠配置延伸通道之實施 例示意圖; 第圖係為解釋第10圖所示之延伸通道組態以及與延伸空 間參數之關係示意圖; ' 第I2圖係為通道的多通道音頻訊號以及^通道的輸出 S1 42 201145263 通道音頻訊號的位置之示意圖; 第13圖係為虛擬聲音源位置與兩個通道間的位階差值之間 的關係示意圖; 第14圖係為解釋兩個後通道的位階以及一個後中央通道的 位階之示意圖; 第丨5圖係為解釋5丨通道的多通道音頻訊號的位置以及71 通道的輸出通道音頻訊號的位置之示意圖; 第16圖係為解釋兩個左通道的位階以及一個左前側通道 (Lfs)的位階之示意圖;以及 第Η圖係為解釋三個前通道的位階以及一個左前側通道 (Lfs)的位階之示意圖。 【主要元件符號說明】 100 編碼裝置 110 降混單元 120 空間資訊擷取單元 200 解碼裝置 210 輸出通道產生單元 220 修正式空間資訊產生單元 230 空間濾波器資訊 d 降混音頻訊號 S 空間資訊 43 201145263 IN_M 多通道音頻訊號 OUT—N 輸出通道音頻訊號 s’ 修正式空間資訊 L 左前通道 Ls 左環繞通道 C、RC 中央通道 LFE 低頻通道 R 右前通道 Rs 右環繞通道 Lt 左總通道 ct 中央總通道 Rt 右方總通道 L〇 ' R〇 立體降混通道 CLD 通道位階差值 ICC 通道間相關或同調 CPC 通道預測係數 M 多通道音頻訊號的通道數目 N 輸出通道音頻訊號的通道數目 m 單一降混音頻訊號 SPK1 ' SPK2 、SPK3 揚聲器 ATO、ATI 通道劃分單元 VS 虛擬聲音源 201145263A new channel between Ls and Rs) (for example, the rear central channel ruler) level c: [Formula 40] c = a*k + b*(lk) Let 'a and 'b' each be the order of two channels 'k' is the relative position between the channel of the level & the channel of the level b and the channel of the level c. If the channel of the level e (for example, the rear central channel RC) is located at the center position between the channel of the level & (e.g., Ls) and the channel of the level b, then 'k, 〇5. If 'k' is 〇_5, the formula 40 will become the formula 41. [Formula 41] c = (a + b)/2 According to Equation 41, if the channel of the level c (for example, the rear central channel Rc) is located at a central position between the channel of the step a (for example, Ls) and the channel Rs of the step b, The level c of the new channel corresponds to the average of the steps of the previous channel. Further, Equation 40 and Formula 41 are only representative. Therefore, it is also possible to readjust the result of the level c and the values of the level a and the level b. (3) When -2-2 is extended to 7.1 channels and the audio channel of the circum-channel is 5.1 channels, an example of the audio signal of the output channel attempting to generate the channel will be explained below. Figure b is a schematic diagram explaining the position of the 5.1 channel multi-channel audio signal and the position of the 7.1 channel output channel audio signal. Please refer to Figure 15A. Similar to Figure 12A, it can be seen that the position channels of the multi-channel audio signals of the 5" channel are left front channel L, right front channel r, medium 39 201145263 central channel C, low frequency channel ( The LFE, the left surround channel Ls, and the right surround channel Rs are not shown. In the case where the 5, 1 channel multi-channel audio signal is a downmix audio signal, if the spatial parameter is applied to the downmix audio signal, the downmix audio signal will be upmixed again to 5.1 channels. Channel audio signal. However, the left front channel Lfs and the right front channel Rfs as shown in Fig. 15B should be generated to upmix the downmix audio signal into a multichannel audio signal of 7 channels. Since the left front side channel Lfs is located between the left front channel L and the left surrounding channel Ls, the level of the left front side channel Lfs can be determined by interpolation using the level of the left front channel L and the level of the left surrounding channel Ls. Figure 16 is a schematic diagram explaining the order of two left channels and the order of a left front channel (Lfs). Referring to Fig. 16, it can be seen that the level c of the left front side channel Lfs is a linear interpolation value based on the level a of the left front channel L and the level b of the left surrounding channel Ls. Meanwhile, although the left front side passage Lfs is located between the left front passage 1 and the left surround passage Ls, it may be located outside the left front passage L, the center passage 匚, and the right front passage R. Therefore, the steps of the left front channel Lfs can be determined by extrapolation using the steps of the left front channel L, the central channel C, and the right front channel R. Figure 17 is a diagram explaining the steps of the three front channels and the steps of the left front channel. Please refer to "Fig. 17". As can be seen, the level of the left front channel Lfs 201145263 d is based on the linear extrapolation of the level a of the left front channel L, the level c of the central channel C, and the level b of the right front channel R. . In the upmixing process using extended space information, the extended spatial parameters may not need to be fully applied, but may be partially applied. Thus, the process of applying spatial parameters to audio signals can be performed sequentially and hierarchically or in combination and synthetically. Industrial Applicability Accordingly, the present invention provides the following effects. First, the present invention is capable of producing an audio signal that includes a configuration that is different from a predetermined tree configuration to produce various configured audio signals. Second, because the generated audio signal contains a configuration different from the predetermined tree configuration, even if the number of multi-channels before performing downmixing is smaller or larger than the number of speakers, the self-downmix audio signal can be generated with a speaker equal to The number of rounded out channels. Second, in the case of generating an output channel having a number smaller than the number of channels, since the multi-channel audio signal directly generates the self-downmix audio signal, instead of downmixing the output channel audio signal from the multi-channel audio signal, wherein The multi-channel audio signal is generated by the upmix-downmixer, so (4) considerably reduces the amount of work required to decode the audio signal. The fourth 'because of the need to consider the acoustic trajectory when producing the lang information, the present invention can provide a pseudo-surround effect without being able to achieve the surround channel output. Although the present invention has been disclosed above in the foregoing preferred embodiments, it is not intended to limit the invention. It will be appreciated by those skilled in the art that the modifications and variations of the present invention are intended to be within the scope of the invention as defined by the appended claims. BRIEF DESCRIPTION OF THE DRAWINGS FIG. 1 is a block diagram of an audio signal encoding apparatus and an audio signal decoding apparatus of the present invention; FIG. 2 is a block diagram of an embodiment of applying partial spatial information; Shown is a block diagram of another embodiment of applying partial spatial information; FIG. 4 is a block diagram of still another embodiment of applying partial spatial information; and FIG. 5 is an implementation of application combined spatial information. Figure 6 is a block diagram of another embodiment of the application of combined spatial information; Figure 7 is a schematic diagram of the sound path of the speaker to the listener, showing the position of the speaker; Figure 8 is a schematic diagram showing the output signal of each position of the speaker for generating the surround effect; Figure 9 is a conceptual block diagram of the method for generating a 5-channel job by the 5-channel shout; the 10th is based on the extended channel Schematic diagram of the embodiment of the city's configuration extension channel; the figure is to explain the extension channel configuration shown in Figure 10 and the relationship with the extended space parameters; 'I2 picture is the channel multi-channel Frequency signal and output of ^ channel S1 42 201145263 Schematic diagram of the position of the channel audio signal; Figure 13 is a schematic diagram of the relationship between the position of the virtual sound source and the difference between the two channels; Figure 14 is to explain two Schematic diagram of the level of the rear channel and the level of a rear central channel; Figure 5 is a schematic diagram explaining the position of the multi-channel audio signal of the 5 channel and the position of the audio signal of the output channel of the 71 channel; Figure 16 is an explanation A schematic diagram of the order of two left channels and the order of one left front channel (Lfs); and the second diagram is a schematic diagram explaining the steps of the three front channels and the order of one left front channel (Lfs). [Description of main component symbols] 100 encoding device 110 downmixing unit 120 spatial information capturing unit 200 decoding device 210 output channel generating unit 220 modified spatial information generating unit 230 spatial filter information d downmix audio signal S spatial information 43 201145263 IN_M Multi-channel audio signal OUT-N Output channel audio signal s' Corrected spatial information L Left front channel Ls Left surround channel C, RC Central channel LFE Low frequency channel R Right front channel Rs Right surround channel Lt Left total channel ct Central total channel Rt Right Total channel L〇' R〇 stereo downmix channel CLD channel step difference ICC channel correlation or homology CPC channel prediction coefficient M number of channels of multichannel audio signal n number of channels of output channel audio signal m single downmix audio signal SPK1 'SPK2, SPK3 speaker ATO, ATI channel division unit VS virtual sound source 201145263

Lfs 左前側通道 Rfs 右前側通道 a 左環繞通道的位階 b 右環繞通道的位階 c 中央通道的位階 R,、l,、σ 訊號 45Lfs Left front channel Rfs Right front channel a Left surround channel level b Right surround channel level c Center channel level R,, l, σ Signal 45

Claims (1)

201145263 七、申請專利範圍: 1_ 一種音頻訊號之解碼方法,包含有. 接收一空間資訊; 使用該空間資訊產生一組合式式空間資訊;以及 使用該組合式空間資訊解碼該音頻訊號, 其中雜合式m資訊係透過組合該空間資訊中包含的 複數個空間參數被產生。 2.如申4專她圍第1項所述之音頻訊號之解碼方法,其中該組 合式空間資訊係透過輸入該等空間參數至一轉換公式被產生。 3·如申睛專利細第2項所述之音頻訊號之解碼方法,其中該轉 換公式係依照該音頻訊號之一樹狀組態資訊被判定。 4‘如申請專利翻第1項所述之音雜號之解碼方法,其中該組 合式空間資訊係基於一輸出通道資訊被產生。 5·如申請專利範圍第1項所述之音頻訊號之解碼方法,其中該音 頻訊號係為降混一多通道音頻訊號所產生的一降混音頻訊 號’以及該4空間參數於該多通道音頻訊號之降混過程中依照 一預定樹狀組態被判定。 6.如申請專利範圍第1項所述之音頻訊號之解碼方法,其中該音 頻訊號係為降混一多通道音頻訊號所產生之一訊號,該等空間 參數包含該多通道音頻訊號之一通道間位階差值至少其一,以 及經過組合之該等空間參數之一通道間位階差值係透過全部 或部分地組合該多通道音頻訊號之該通道間位階差值被計算。 46 201145263 7. 如申請專利範圍第1項所述之音頻訊號之解碼方法,其中該音 頻訊號係為降混一多通道音頻訊號所產生之一訊號,該等空間 參數包含該多通道音頻訊號之一通道間相關值至少其一,以及 經過組合之該等空間參數之一通道間相關值係透過組合該多 通道音頻訊號之該通道間相關值至少其一被計算。 8. 如申請專利範圍第7項所述之音頻訊號之解碼方法,其中該等 空間參數更包含該多通道音頻訊號之一通道間位階差值至少 其一,經過組合之該等空間參數之該通道間相關值係透過組合 該多通道音頻訊號之該通道間相關值至少其—與該多通道音 頻訊號之該通道間位階差值至少其一被計算。 9. 如申請專利範圍第1項所述之音頻訊號之解碼方法,其中包含 該組合式空間資訊之一修正式空間資訊係使用該空間資訊被 產生,以及該修正式空間資訊包含一部分空間資訊與一延伸空 間資訊至少其一。 10. —種音頻訊號之解碼裝置,包含有: 修正式空間資訊產生單元,用以利用—空間資訊產生一 組合式空間資訊;以及 一輸出通道產生單元,用以_雜合式空間資訊解碼該 音頻訊號, 其中該組合式空間資訊係透過組合該空間資訊中包含的 複數個空間參數被產生。 11. 如申請專利範圍第1G項所述之音頻訊號之解钟置,其中該 201145263 組合式空間資訊係透過輸入該等空間參數至一轉換公式被產 生。 12. 如申請專利範圍第11項所述之音頻訊號之解碼裝置,其中該 轉換公式係依照該音頻訊號之一樹狀組態資訊被判定。 13. 如申請專利範圍第10項所述之音頻訊號之解碼裝置,其中該 組合式空間資訊係基於一輸出通道資訊被產生。 14. 如申請專利範圍第10項所述之音頻訊號之解碼裝置,其中該 音頻訊號係為降混一多通道音頻訊號所產生之一降混音頻訊 號,該等空間參數係於該多通道音頻訊號之降混過程中依照一 預定樹狀組態被判定。 如申請專利範圍第10項所述之音頻訊號之解碼裝置,其中該 音頻訊號係為降混一多通道音頻訊號所產生之一訊號,該等空 間參數包含該多通道音頻訊號之一通道間位階差值至少其 一,以及經過組合之該等空間參數之一通道間位準差值係透過 全部或部分地組合該多通道音頻訊號之該多通道位準差值被 計算。 16. 如申請專利範圍第10項所述之音頻訊號之解碼裝置,其中該 音頻訊號係為降混一多通道音頻訊號所產生之—訊號,該等空 間參數包含該多通道音頻訊號之一通道間相關值至少其一,以 及經過組合之該等空間參數之一通道間相關值係透過組合該 多通道音頻訊號之該通道間相關值至少其一被計算。 17. 如申請專利範圍第16項所述之音頻訊號之解碼裝置,其中該 S 48 201145263 等空間參數更包含該多通道音頻訊號之一通道間位階差值至 少其- ’經驗合之鱗郎參數之該通道間_值係透過組 合該多通道音頻訊號之該通道間相關值至少其一與該多通道 音頻訊號之該通道間位階差值至少其一被計算。 18.如申請專利範圍第1〇項所述之音頻訊號之解碼裝置,其中包 含s亥組合式空間資訊之一修正式空間資訊係使用該空間資訊 被產生,該修正式空間資訊包含一部分空間資訊與一延伸空間 資訊至少其一。 49201145263 VII. Patent application scope: 1_ A method for decoding an audio signal, comprising: receiving a spatial information; using the spatial information to generate a combined spatial information; and decoding the audio signal by using the combined spatial information, wherein the hybrid signal The m information system is generated by combining a plurality of spatial parameters included in the spatial information. 2. The method for decoding an audio signal as described in claim 1, wherein the combined spatial information is generated by inputting the spatial parameters to a conversion formula. 3. The method for decoding an audio signal as described in claim 2, wherein the conversion formula is determined according to tree configuration information of one of the audio signals. 4 'If the patent application is turned over, the decoding method of the audio code described in item 1, wherein the combined spatial information is generated based on an output channel information. 5. The method for decoding an audio signal according to claim 1, wherein the audio signal is a downmix audio signal generated by downmixing a multi-channel audio signal and the four spatial parameters are in the multi-channel The downmixing process of the audio signal is determined according to a predetermined tree configuration. 6. The method for decoding an audio signal according to claim 1, wherein the audio signal is a signal generated by downmixing a multi-channel audio signal, and the spatial parameters comprise one channel of the multi-channel audio signal. At least one of the inter-level difference values, and one of the spatial parameters of the combined spatial parameters, is calculated by dividing the inter-channel level difference value of the multi-channel audio signal in whole or in part. 46. The method for decoding an audio signal according to claim 1, wherein the audio signal is a signal generated by downmixing a multi-channel audio signal, and the spatial parameters comprise the multi-channel audio signal. At least one of the inter-channel correlation values, and one of the inter-channel correlation values of the combined spatial parameters is calculated by combining at least one of the inter-channel correlation values of the multi-channel audio signals. 8. The decoding method of the audio signal according to claim 7, wherein the spatial parameters further comprise at least one of the channel difference values of one of the multi-channel audio signals, and the spatial parameters are combined. The inter-channel correlation value is calculated by combining at least one of the inter-channel correlation values of the multi-channel audio signal with the inter-channel level difference of the multi-channel audio signal. 9. The method for decoding an audio signal according to claim 1, wherein the modified spatial information including the combined spatial information is generated using the spatial information, and the modified spatial information includes a part of spatial information and An extended space information is at least one of them. 10. A decoding device for an audio signal, comprising: a modified spatial information generating unit for generating a combined spatial information by using the spatial information; and an output channel generating unit for decoding the audio by the hybrid spatial information a signal, wherein the combined spatial information is generated by combining a plurality of spatial parameters included in the spatial information. 11. The solution of the audio signal described in claim 1G of the patent application, wherein the 201145263 combined spatial information is generated by inputting the spatial parameters to a conversion formula. 12. The decoding device for an audio signal according to claim 11, wherein the conversion formula is determined according to one of the audio signal tree configuration information. 13. The decoding device for an audio signal according to claim 10, wherein the combined spatial information is generated based on an output channel information. 14. The decoding device for an audio signal according to claim 10, wherein the audio signal is a downmix audio signal generated by downmixing a multi-channel audio signal, wherein the spatial parameters are in the multi-channel The downmixing process of the audio signal is determined according to a predetermined tree configuration. The decoding device for an audio signal according to claim 10, wherein the audio signal is a signal generated by downmixing a multi-channel audio signal, and the spatial parameters comprise an inter-channel level of the multi-channel audio signal. At least one of the differences, and one of the spatial parameters of the combined spatial parameters, is calculated by combining the multi-channel level differences of the multi-channel audio signals in whole or in part. 16. The decoding device for an audio signal according to claim 10, wherein the audio signal is a signal generated by downmixing a multi-channel audio signal, and the spatial parameters comprise one channel of the multi-channel audio signal. At least one of the inter-correlation values, and one of the inter-channel correlation values of the spatial parameters that are combined, is calculated by combining at least one of the inter-channel correlation values of the multi-channel audio signals. 17. The decoding device for an audio signal according to claim 16, wherein the spatial parameter such as S 48 201145263 further includes a step difference between the channels of the multi-channel audio signal, at least - the 'experienced scale parameter The inter-channel_value is calculated by at least one of the inter-channel correlation value of the multi-channel audio signal combined with the inter-channel level difference of the multi-channel audio signal. 18. The decoding device for an audio signal according to claim 1, wherein the modified spatial information including the sigma combined spatial information is generated using the spatial information, and the modified spatial information includes a part of the spatial information. At least one of the information with an extended space. 49
TW100109444A 2005-09-14 2006-09-14 Method and apparatus for decoding an audio signal TWI485698B (en)

Applications Claiming Priority (8)

Application Number Priority Date Filing Date Title
US71652405P 2005-09-14 2005-09-14
US75998006P 2006-01-19 2006-01-19
US76036006P 2006-01-20 2006-01-20
US77366906P 2006-02-16 2006-02-16
US77672406P 2006-02-27 2006-02-27
US78751606P 2006-03-31 2006-03-31
US81602206P 2006-06-22 2006-06-22
KR20060078300 2006-08-18

Publications (2)

Publication Number Publication Date
TW201145263A true TW201145263A (en) 2011-12-16
TWI485698B TWI485698B (en) 2015-05-21

Family

ID=44072944

Family Applications (2)

Application Number Title Priority Date Filing Date
TW095134072A TWI462086B (en) 2005-09-14 2006-09-14 Method and apparatus for decoding an audio signal
TW100109444A TWI485698B (en) 2005-09-14 2006-09-14 Method and apparatus for decoding an audio signal

Family Applications Before (1)

Application Number Title Priority Date Filing Date
TW095134072A TWI462086B (en) 2005-09-14 2006-09-14 Method and apparatus for decoding an audio signal

Country Status (3)

Country Link
JP (2) JP4728398B2 (en)
BR (1) BRPI0616057A2 (en)
TW (2) TWI462086B (en)

Families Citing this family (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2851132B2 (en) 1990-06-20 1999-01-27 株式会社東芝 Endoscope light guide
CA2621664C (en) 2005-09-14 2012-10-30 Lg Electronics Inc. Method and apparatus for decoding an audio signal
WO2007111568A2 (en) * 2006-03-28 2007-10-04 Telefonaktiebolaget L M Ericsson (Publ) Method and arrangement for a decoder for multi-channel surround sound
JP5298245B2 (en) * 2009-12-16 2013-09-25 ドルビー インターナショナル アーベー SBR bitstream parameter downmix
KR20120004909A (en) * 2010-07-07 2012-01-13 삼성전자주식회사 Stereo playback method and apparatus
KR102033071B1 (en) * 2010-08-17 2019-10-16 한국전자통신연구원 System and method for compatible multi channel audio
CH703771A2 (en) * 2010-09-10 2012-03-15 Stormingswiss Gmbh Device and method for the temporal evaluation and optimization of stereophonic or pseudostereophonic signals.
WO2014007095A1 (en) * 2012-07-02 2014-01-09 ソニー株式会社 Decoding device and method, encoding device and method, and program
CN103748629B (en) * 2012-07-02 2017-04-05 索尼公司 Decoding device and method, encoding device and method, and program
TWI517142B (en) * 2012-07-02 2016-01-11 Sony Corp Audio decoding apparatus and method, audio coding apparatus and method, and program
BR112014004128A2 (en) 2012-07-02 2017-03-21 Sony Corp device and decoding method, device and encoding method, and, program
EP2830052A1 (en) * 2013-07-22 2015-01-28 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Audio decoder, audio encoder, method for providing at least four audio channel signals on the basis of an encoded representation, method for providing an encoded representation on the basis of at least four audio channel signals and computer program using a bandwidth extension
CN107004421B (en) * 2014-10-31 2020-07-07 杜比国际公司 Parametric encoding and decoding of multi-channel audio signals
US10601322B1 (en) * 2019-04-04 2020-03-24 Ambiq Micro, Inc. Strong arm comparator
GB2592388A (en) * 2020-02-26 2021-09-01 Nokia Technologies Oy Audio rendering with spatial metadata interpolation

Family Cites Families (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5632005A (en) * 1991-01-08 1997-05-20 Ray Milton Dolby Encoder/decoder for multidimensional sound fields
JP2924539B2 (en) * 1993-01-29 1999-07-26 日本ビクター株式会社 Sound image localization control method
DE69428939T2 (en) * 1993-06-22 2002-04-04 Deutsche Thomson-Brandt Gmbh Method for maintaining a multi-channel decoding matrix
US6711266B1 (en) * 1997-02-07 2004-03-23 Bose Corporation Surround sound channel encoding and decoding
JP3572165B2 (en) * 1997-04-04 2004-09-29 株式会社デノン Video / audio signal reproducing apparatus and video / audio signal reproducing method
US7212872B1 (en) * 2000-05-10 2007-05-01 Dts, Inc. Discrete multichannel audio with a backward compatible mix
JP4304401B2 (en) * 2000-06-07 2009-07-29 ソニー株式会社 Multi-channel audio playback device
US7391869B2 (en) * 2002-05-03 2008-06-24 Harman International Industries, Incorporated Base management systems
EP1523862B1 (en) * 2002-07-12 2007-10-31 Koninklijke Philips Electronics N.V. Audio coding
US7447317B2 (en) * 2003-10-02 2008-11-04 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V Compatible multi-channel coding/decoding by weighting the downmix channel
US7394903B2 (en) * 2004-01-20 2008-07-01 Fraunhofer-Gesellschaft Zur Forderung Der Angewandten Forschung E.V. Apparatus and method for constructing a multi-channel output signal or for generating a downmix signal
WO2005081229A1 (en) * 2004-02-25 2005-09-01 Matsushita Electric Industrial Co., Ltd. Audio encoder and audio decoder

Also Published As

Publication number Publication date
TW200721112A (en) 2007-06-01
JP4740335B2 (en) 2011-08-03
TWI485698B (en) 2015-05-21
TWI462086B (en) 2014-11-21
JP2009508433A (en) 2009-02-26
JP4728398B2 (en) 2011-07-20
JP2009508432A (en) 2009-02-26
BRPI0616057A2 (en) 2011-06-07

Similar Documents

Publication Publication Date Title
RU2509442C2 (en) Method and apparatus for applying reveberation to multichannel audio signal using spatial label parameters
JP5587551B2 (en) Audio encoding
AU2007328614B2 (en) A method and an apparatus for processing an audio signal
US9747905B2 (en) Method and apparatus for decoding an audio signal
TW201145263A (en) Method and apparatus for decoding an audio signal
US20080235006A1 (en) Method and Apparatus for Decoding an Audio Signal
US20080221907A1 (en) Method and Apparatus for Decoding an Audio Signal
CN101454828B (en) Method and device for decoding audio signal
RU2380767C2 (en) Method and device for audio signal decoding
HK1126306B (en) Method and apparatus for decoding an audio signal

Legal Events

Date Code Title Description
MM4A Annulment or lapse of patent due to non-payment of fees