TW201143455A - Image encoding device, image decoding device, image encoding method, image decoding method - Google Patents
Image encoding device, image decoding device, image encoding method, image decoding method Download PDFInfo
- Publication number
- TW201143455A TW201143455A TW100104246A TW100104246A TW201143455A TW 201143455 A TW201143455 A TW 201143455A TW 100104246 A TW100104246 A TW 100104246A TW 100104246 A TW100104246 A TW 100104246A TW 201143455 A TW201143455 A TW 201143455A
- Authority
- TW
- Taiwan
- Prior art keywords
- direct
- vector
- image
- direct vector
- mode
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 96
- 239000013598 vector Substances 0.000 claims abstract description 760
- 230000033001 locomotion Effects 0.000 claims abstract description 377
- 230000002123 temporal effect Effects 0.000 claims abstract description 170
- 238000012545 processing Methods 0.000 claims abstract description 137
- 238000011156 evaluation Methods 0.000 claims abstract description 132
- 238000004364 calculation method Methods 0.000 claims abstract description 65
- 238000013139 quantization Methods 0.000 claims description 23
- 239000006185 dispersion Substances 0.000 claims description 14
- 241000282320 Panthera leo Species 0.000 claims description 2
- 238000010187 selection method Methods 0.000 claims 1
- 230000015654 memory Effects 0.000 description 36
- 238000010586 diagram Methods 0.000 description 17
- 230000006835 compression Effects 0.000 description 15
- 238000007906 compression Methods 0.000 description 15
- 239000000463 material Substances 0.000 description 10
- 230000000694 effects Effects 0.000 description 4
- 238000005259 measurement Methods 0.000 description 4
- 238000005516 engineering process Methods 0.000 description 3
- 239000012634 fragment Substances 0.000 description 3
- 239000004575 stone Substances 0.000 description 3
- 230000000295 complement effect Effects 0.000 description 2
- 210000000887 face Anatomy 0.000 description 2
- 239000000047 product Substances 0.000 description 2
- 239000004065 semiconductor Substances 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- 241000208340 Araliaceae Species 0.000 description 1
- 101100313164 Caenorhabditis elegans sea-1 gene Proteins 0.000 description 1
- 241000632511 Daviesia arborea Species 0.000 description 1
- 235000005035 Panax pseudoginseng ssp. pseudoginseng Nutrition 0.000 description 1
- 235000003140 Panax quinquefolius Nutrition 0.000 description 1
- RTAQQCXQSZGOHL-UHFFFAOYSA-N Titanium Chemical compound [Ti] RTAQQCXQSZGOHL-UHFFFAOYSA-N 0.000 description 1
- 239000000654 additive Substances 0.000 description 1
- 230000000996 additive effect Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 230000000903 blocking effect Effects 0.000 description 1
- 210000004556 brain Anatomy 0.000 description 1
- 238000009924 canning Methods 0.000 description 1
- 230000008602 contraction Effects 0.000 description 1
- 238000005202 decontamination Methods 0.000 description 1
- 230000003588 decontaminative effect Effects 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 230000035622 drinking Effects 0.000 description 1
- 230000008451 emotion Effects 0.000 description 1
- 235000008434 ginseng Nutrition 0.000 description 1
- 238000003702 image correction Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 239000008267 milk Substances 0.000 description 1
- 210000004080 milk Anatomy 0.000 description 1
- 235000013336 milk Nutrition 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000004091 panning Methods 0.000 description 1
- 238000003825 pressing Methods 0.000 description 1
- 210000001747 pupil Anatomy 0.000 description 1
- 239000012925 reference material Substances 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 238000003860 storage Methods 0.000 description 1
- 239000013589 supplement Substances 0.000 description 1
- 210000004243 sweat Anatomy 0.000 description 1
- 238000010257 thawing Methods 0.000 description 1
- 229910052719 titanium Inorganic materials 0.000 description 1
- 239000010936 titanium Substances 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/60—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
- H04N19/61—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
Description
201143455 • 六、發明說明: 【發明所屬之技術領域】 本發明係關於使用於影像壓縮編碼技術、壓縮影像資 料傳送技術等中之影像編碼裝置、影像解碼裝置、影像編 碼方法及影像解碼方法。 【先前技術】 例如 MPEG(Moving Picture Experts Group)及「πυ-Τ Η· 26x」等之國際標準影像編碼方式,採用的是··以妗合 亮度訊號16x16畫素、及與此亮度訊號16χ16畫素對應之 色差訊號8x8晝素而成之區塊資料(bl〇ckdata)(以下稱之 為「大區塊(macroblock)」)為一個單位,而根據移 技術及正交變換/變換像數量子化技術加以壓縮之方法。 影像編碼裝置及影像解碼裝置中之移動補償處理,係 參考前方或後方的晝面(picture),以大區塊為單位而進行 移動向量之檢測及預測影像之生成。 此寺’、參考一個晝面而進行晝面間預測編碼者稱為 P旦面(P picture) ’同時參考兩個晝面而進行晝面間預測 編碼者稱為B晝面(B Picture)。 、 國際標準編碼方式中之AVC/H. 264CISO/IEC 14496-10 I ITU-TH. 264) ’在進行B晝面的編碼之際,可選擇被稱 為直接模式(direct 之編碼模式(參照例如非專利文 獻1) 〇 、亦即,可選擇:編竭對象大區塊(將予以編碼之大區塊) 並不保有移動向量之編石馬資料,而是藉由使用已編碼之其 322763 4 201143455 他晝面(picture)的大區塊的移動向量、或使用周圍的大區 塊的移動向量之預定的演算處理,來生成編碼對象大區塊 的移動向量之編碼模式。 該直接模式有時間直接模式及空間直接模式兩種。 時間直接模式,係參考已編碼之其他晝面的移動向 量,依據已編碼晝面與編碼對象晝面之時間差而進行移動 向量之比例調整(sea 1 ing),來生成編碼對象大區塊的移動 向量。 空間直接模式,係參考位於編碼對象大區塊的周圍之 至少一個以上的已編碼大區塊的移動向量,從該等移動向 量來生成編碼對象大區塊的移動向量。 該直接模式,可藉由使用設於片段標頭(s 1 i ce header) 中之旗標 “direct_spatial_mv_pred_flag” ,而以片段 (si ice)為單位選擇時間直接模式及空間直接模式之任一 方。 第9圖係顯示以時間直接模式來生成移動向量之方法 之示意圖。 第9圖中,「P」表示P晝面,「B」表示B晝面。 此外,數字0-3表示晝面的顯示順序,代表時間TO, T1, T2,T3之顯示影像。 畫面之編碼處理,係以P0,P3,Bl,B2之順序進行。 假想為例如要以時間直接模式將晝面B2之中的大區 塊MB 1予以編碼之情況。 此情況,係使用在時間軸上位於晝面B2的後方之已編 5 322763 201143455 碼晝面之中最靠近畫面B2之晝面P3的移動向量,亦即使 用該晝面P3中之在空間上位於與大區塊MB1相同位置之大 區塊MB2的移動向量。 此移動向量MV係參考到晝面p〇,且進行大區塊 的編碼之際要用到的移動向量MVLO,MVL1,係以以下的 ⑴求出。 式 肅 λ Τ2-Τ0BACKGROUND OF THE INVENTION 1. Field of the Invention The present invention relates to an image encoding device, a video decoding device, an image encoding method, and a video decoding method used in image compression coding technology, compressed image data transmission technology, and the like. [Prior Art] For example, the international standard video coding method such as MPEG (Moving Picture Experts Group) and "πυ-Τ Η 26x" is used to combine the luminance signal 16x16 pixels and the luminance signal 16χ16 The block data (bl〇ckdata) (hereinafter referred to as "macroblock") formed by the color difference signal 8x8 is a unit, and the number of the image according to the shift technique and the orthogonal transform/transformation The method of compression by technology. The motion compensation processing in the video encoding device and the video decoding device refers to the front or rear picture, and performs motion vector detection and prediction image generation in units of large blocks. This temple's reference to a facet and the inter-plane predictive coding is called P-picture (P picture). At the same time, the two-faced reference is used for inter-plane prediction. The coder is called B picture. In the international standard coding method, AVC/H.264CISO/IEC 14496-10 I ITU-TH. 264) 'When performing B-plane coding, it can be called direct mode (direct coding mode (see for example Non-Patent Document 1) 〇, that is, you can choose: the large block of the object to be edited (the large block to be coded) does not retain the data of the moving vector, but by using the coded 322763 4 201143455 The motion vector of the large block of his picture, or the predetermined calculation process of the motion vector of the surrounding large block, to generate the coding mode of the motion vector of the large block of the coding object. The direct mode has time Direct mode and spatial direct mode. The time direct mode refers to the motion vector of other encoded faces, and adjusts the ratio of the moving vector according to the time difference between the encoded face and the face of the coded object (sea 1 ing). To generate a motion vector of a large block of the coding object. The spatial direct mode refers to a motion vector of at least one of the coded large blocks located around the large block of the coding object, from which The motion vector is used to generate a motion vector of the large block of the coding object. The direct mode can be obtained by using the flag "direct_spatial_mv_pred_flag" set in the slice header (s1 i ce header) and the slice (si ice) The unit selects either the time direct mode or the spatial direct mode. Fig. 9 is a schematic diagram showing a method of generating a motion vector in a time direct mode. In Fig. 9, "P" represents a P plane, and "B" represents a B node. In addition, the numbers 0-3 indicate the order in which the faces are displayed, representing the display images of the times TO, T1, T2, and T3. The encoding processing of the screens is performed in the order of P0, P3, B1, and B2. The case where the large block MB 1 in the face B2 is encoded in the time direct mode. In this case, the closest picture to the rear of the edited 5 322763 201143455 coded face located behind the face B2 on the time axis is used. The motion vector of the face P3 of B2, that is, the motion vector of the large block MB2 spatially located at the same position as the large block MB1 in the facet P3. The motion vector MV is referred to the face p〇, Large area To use the occasion of encoding the motion vectors MVLO, MVL1, to obtain the following system ⑴. Su formula λ Τ2-Τ0
MVL0---xMV Τ3-Γ0 λ/η/τ 1 — T3MVL0---xMV Τ3-Γ0 λ/η/τ 1 — T3
Mm=~ ^MV T3-TQ 因此,要以時間直接模式來求出編碼對象大區塊的移 動向量,需要一個畫面份的已編碼晝面的移動向量My,所 以需要有保持移動向量之記憶體(mem〇ry )。 第10圖係顯示以空間直接模式來生成移動向量之方 法之不意圖。 第10圖中,current MB表示編碼對象大區塊。 此時,假設編碼對象大區塊之橫向左邊的已編碼大區 塊A的移動向量為MVa,編碼對象大區塊之上方的已編: 大區塊B的移動向量為MVb,編碼對象大區塊之右上方的 已編碼大區塊C的移動向量為MVc,則如以下的式(2)所 示,將此等移動向量MVa,MVb,MVc的中央值(median)求 出’就可异出編碼對象大區塊的移動向量MV。 MV = ine d i a η (Μ V a、MVb、MVc) (2) 在空間直接模式中,雖然要針對前方及後方分別求出 移動向量,但不論是哪一方都可用上述的方法來求出。 322763 6 201143455 在AVC/H· 264中,要選擇空間直接模式或時間直接模 式的哪一方,係如上述參考設於片段標頭中之旗標 “direct一spatial_mv_pred_flag” ,如此一來,就必須以 片段(slice)為單位進行時間直接模式或空間直接模式之 選擇,所以無法逐一就每個大區塊切換及使用最適合的直 接模式。 [先前技術文獻] 〔非專利文獻〕 非專利文獻 1 : MPEG-4 AVC(IS0/IEC 14496-10)/ITU-T H. 264規格 【發明内容】 〔發明所欲解決之課題〕 以往的影像編碼裝置係如上述般構成,只要參考設於 片段標頭中之旗標 “direct_spatial_mv_pred_f lag”,就 可以以片段(siice)為單位進行時間直接模式及空間直接 模式之切換。不過,因為無法以大區塊為單位進行時間直 接模式及空間直接模式之切換,所以若對於屬於某一片段 之某一大區塊而言最適合的直接模式為例如空間直接模 式’但對應於該片段之直接模式卻經決定為時間直接模 式,就必須對該大區塊使用時間直接模式,而無法選擇最 適合的直接模式。在這樣的情況’因為無法選擇最適合的 直接模式,所以必須對不必要的移動向量進行編碼,而有 編碼量會增加等之課題。 本發明係為了解決如上述之課題而完成者,其目的在 322763 7 201143455 獲得能夠以預定的區塊為單位選擇最適合的直接模式之影 像編碼裝置、影像解碼裝置、影像編碼方法及影像解碼方 法。 〔解決課題之手段〕 本發明之影像編碼裝置,係設有:直接向量生成手段, 從位於編碼對象區塊的周圍之已編碼區塊的移動向量來生 成空間直接模式的空間直接向量,並且從在時間上處於編 碼對象區塊附近之已編碼晝面的移動向量來生成時間直接 模式的時間直接向量;評估值算出手段,使用藉由直接向 量生成手段而生成的空間直接向量,來算出空間直接模式 的評估值,並且使用該時間直接向量,來算出時間直接模 式的評估值;以及直接向量選擇手段,進行藉評估值算出 手段而算出的空間直接模式的評估值與時間直接模式的評 估值之比較,而選擇空間直接向量或時間直接向量的任一 方,且預測影像生成手段使用藉由直接向量選擇手段而選 出的直接向量來實施移動補償預測處理,從而生成預測影 像。 〔發明效果〕 依據本發明,由於係構成為設有:直接向量生成手段, 從位於編碼對象區塊的周圍之已編碼區塊的移動向量來生 成空間直接模式的空間直接向量,並且從在時間上處於編 碼對象區塊附近之已編碼晝面的移動向量來生成時間直接 模式的時間直接向量;評估值算出手段,使用藉由直接向 量生成手段而生成的空間直接向量,來算出空間直接模式 8 322763 201143455 的评估值並足使用該時間直接向量,來算出時間直接模 式的=估,;R及直接向量選擇手段,進行藉由評估值算 出手&而算出的空間直接模式的評估值與時間直接模式的 °平估值之比&、選擇空間直接向量或時間直接向量的任 方且預測影像生成手段使用藉由直接向量選擇手段而 選出的直接向量來實施移動補償預測處理,從而生成預測 影像仕所預定的區塊為單位選擇最適合的直接模 式,結果就具有可避免掉不必要的移動向量之編碼,而防 止編碼量增加之欵果。 【實施方式】 依照所附的圖式針 以下,為了更詳細地說明本發明 對實施本糾切態進行說明。 實施形態1. 構成林發明實施㈣丨巾的影絲碼裝置之 第1圖之影像編碼裝置,係說明使用AVC/H· 264規格 所採用的編碼方式之例。 第1圖中,移動向量記憶體1係用來儲存已編碼大區 塊(或者分割大區塊而成的次區塊(sub_macr〇b 1〇ck))的移 動向量之例如RAM等之記錄媒體。 移動補償預測部2實施以下處理:從儲存於圖框記憶 體(frame mem〇ry)9中之一圖框以上的移動補償預測用^ 參考影像之令選擇一圖框的參考影像,然後以構成輸入影 像之大區塊(或者分割大區塊而成的次區塊)為單位執/移 322763 9 201143455 動補償預測處理,而生成該大區塊(編碼對象大區塊)的移 動向量進而生成預測影像,並將針對各個大區塊的每一個 而選擇的參考影像的識別號碼、移動向量以及預測影像等 予以輸出。 不過,此處為了說明之方便,假設其為以大區塊為單 位而生成移動向量進而生成預測影像者。 但,移動補償預測部2在生成編碼對象大區塊的移動 向量進而生成預測影像之際,係針對構成輸入影像之各個 大區塊,實施以下處理:從位於該大區塊的周圍之已編碼 大區塊的移動向量(儲存於移動向量記憶體1中之移動向 量)來生成空間直接模式的空間直接向量,並且從在時間上 處於該大區塊附近之已編碼畫面的移動向量(儲存於移動 向量記憶體1中之移動向量)來生成時間直接模式的時間 直接向量。 然後,移動補償預測部2實施以下處理:使用該空間 直接向量來算出空間直接模式的評估值,以及使用該時間 直接向量來算出時間直接模式的評估值。 然後,移動補償預測部2實施以下處理:進行空間直 接模式的評估值與時間直接模式的評估值之比較,而選擇 空間直接向量或時間直接向量的任一方。 然後,移動補償預測部2實施以下處理:使用所選擇 的直接向量來實施移動補償預測處理,而生成預測影像。 減法器3實施以下處理:將移動補償預測部2所生成 的預測影像與輸入影像之差異影像(difference image)予 10 322763 201143455 •=二=表示該差異影像之預測差異訊號輸出至編 =核式判定部4實施以下處理:對於減法器3所輪 ㈣差異訊號的預測效率進行評估’從減法器3所“ 夕個以上的預測差異訊號之中選出預測效率最言 =二差異訊號,然後將與該預測差異訊號有關之曾在: 員預測部2的預測影像的生成中使用到的移動向量、 次區塊型式(包含例如,表示在該大區塊使用 者之-L :係為晝面間模式(inter mode)或直接模式的何 輸出參考f彡像的識職碼作為編碼模式資訊而 異訊號輸出至壓縮部5及將制效率最高之預測差 壓縮部5實施以下處理: 、 出的預測差, 。碼模式判定部4所輸 DCT㈣ T(離散餘弦變換)處理以笪Ψ 殿係數,而且將該⑽係 ^吳)處理以异出 的DCT係數(一種壓縮資 置子化’而將量子化後 度編碼部1〇。 μ至局部解碼部6及可變長 又,減法器3、編碼模式判 量子化手段。 疋。卩4及壓縮部5係構成 局部解碼部6實施以下處理. 縮資料予以逆量子化來求出DCT赵縮邛5所輸出的壓 施逆DCT(逆離散餘弦變換)處理「管^對該DCT係數實 部4所輸㈣闕差魏動目當㈣^賴式判定 加法器7實施以下處理··將局部解碼部6所算出的預 322763 11 201143455 :差訊號及由移動補償制部2所生成的表示預測影像 碼影:=r以相加’以生成表示局部解碼影一解 :遽波器(1_ filter)8實施以下處理:對於加法 償,炒f出的局部解碼影像訊號中所含的編碼失真進行補 的局影像訊號所表示 F麥亏衫像予以輸出至圖框記憶體9 » ^匡記憶體9係用來儲存從迴路遽波器 考衫像之RAM等記錄媒體。 τ和 輸出度編碼部1G實施以下處理:對於從_部5所 編號)進弁输, 弋移動向里、參考影像的識別 位元串'Ά. entRDPy)編碼’而生成表示該熵編碼結果之 以輪π reamx_資料)’然後將該位元串流予 可《長度編碼部1Q係構成可變長度編碼手段。 圖係顯不本發明實施形態J中之 移動補償_部2之構成圖。 〜像糾裝置的 為晝二資動:量搜尋部21在接收到表示 面間模式=(實:外 2,動向量,並將‘量輸:=^^ 直接向量生成部22在桩收5丨丨本-站 式之資訊時,實施不、扁碼模式為直接模 實施下處理:針對每一個編石馬對象大區塊’ 322763 12 201143455 生成空間直接模式的空間直接向量、及時間直接模式的時 間直接向量,並將該空間直接向量或時間直接向量的任一 方當作移動向量而輸出至移動補償處理部23。 移動補償處理部23實施以下處理:使用從移動向量搜 尋部21或直接向量生成部22所輸出之移動向量、及圖框 記憶體9中儲存的一圖框之參考影像來實施移動補償預測 處理,而生成預測影像之處理。又,移動補償處理部23係 構成預測影像生成手段。 第3圖係顯示構成移動補償預測部2之直接向量生成 部22之構成圖。 第3圖中,空間直接向量生成部31實施以下處理:從 移動向量記憶體1所儲存之已編碼大區塊的移動向量之 中,將位於編碼對象大區塊的周圍之已編碼大區塊的移動 向量予以讀出,從該移動向量來生成空間直接模式的空間 直接向量。 時間直接向量生成部32實施以下處理:從移動向量記 憶體1所儲存之已編碼大區塊的移動向量之中,將在時間 上處於編碼對象大區塊附近之已編碼晝面的移動向量,亦 即該已編碼畫面中之在空間上位於與編碼對象大區塊相同 位置之大區塊的移動向量予以讀出,從該移動向量來生成 時間直接模式的時間直接向量。 空間直接向量生成部31及時間直接向量生成部32係 構成直接向量生成手段。 直接向量判定部33實施以下處理:使用空間直接向量 13 322763 201143455 生成部31所生成的空間直接向量,來算出空間直接模式的 評估值,以及使用時間直接向量生成部32所生成的時間直 接向量,來算出時間直接模式的評估值,然後進行該空間 直接模式的評估值與時間直接模式的評估值之比較,而選 擇該空間直接向量或時間直接向量的任一方。 第4圖係顯示構成直接向量生成部22之直接向量判定 部33之構成圖。 第4圖中,移動補償部41實施以下處理:使用空間直 接向量生成部31所生成的空間直接向量,來生成空間直接 模式的前方預測影像及後方預測影像,以及使用時間直接 向量生成部32所生成的時間直接向量,來生成時間直接模 式的前方預測影像及後方預測影像。 類似度算出部42實施以下處理:算出空間直接模式的 前方預測影像與後方預測影像的類似度,以之作為空間直 接模式的評估值,以及算出時間直接模式的前方預測影像 與後方預測影像的類似度,以之作為時間直接模式的評估 值。 移動補償部41及類似度算出部42係構成評估值算出 手段。 直接向量選擇部43實施以下處理:進行類似度算出部 42所算出的空間直接模式中的前方預測影像與後方預測影 像的類似度、以及時間直接模式中的前方預測影像與後方 預測影像的類似度之比較,而選擇空間直接向量或時間直 接向量之中前方預測影像與後方預測影像的類似度較高之 14 322763 201143455 ίΠΐ的直接向量。直接向量選擇部43係構成直接向量 構成^圖軸示本翻實祕態1中的影像解碼裝置之 第5圖之影像解碼裝置明 所採用的編碼方式之例。 H 264規格Mm=~ ^MV T3-TQ Therefore, to obtain the motion vector of the large block of the coding object in the direct time mode, the motion vector My of the coded face of one picture is required, so the memory that holds the motion vector is needed. (mem〇ry). Figure 10 shows the intent of the method of generating a motion vector in spatial direct mode. In Fig. 10, current MB indicates a large block of the encoding object. At this time, it is assumed that the motion vector of the coded large block A on the left side of the large block of the coding object is MVa, and the coded object above the large block is encoded: the motion vector of the large block B is MVb, and the coding target area is large. If the motion vector of the coded large block C on the upper right of the block is MVc, the median value of the motion vectors MVa, MVb, and MVc is determined as shown in the following equation (2). The motion vector MV of the large block of the coding object is extracted. MV = ine d i a η (Μ V a, MVb, MVc) (2) In the spatial direct mode, the motion vector is obtained for each of the front and the rear, but either one can be obtained by the above method. 322763 6 201143455 In AVC/H·264, which side of the spatial direct mode or the time direct mode is to be selected, as described above with reference to the flag "direct one spatial_mv_pred_flag" set in the fragment header, thus, it is necessary to The slice is selected in units of time direct mode or spatial direct mode, so it is not possible to switch and use the most suitable direct mode for each large block one by one. [Prior Art Document] [Non-Patent Document] Non-Patent Document 1: MPEG-4 AVC (IS0/IEC 14496-10)/ITU-T H.264 Specifications [Summary of the Invention] [Problems to be Solved by the Invention] Conventional Images The coding apparatus is constructed as described above, and the time direct mode and the spatial direct mode can be switched in units of segments (siice) by referring to the flag "direct_spatial_mv_pred_flag" provided in the slice header. However, since the direct direct mode and the spatial direct mode cannot be switched in units of large blocks, the most suitable direct mode for a large block belonging to a certain segment is, for example, a spatial direct mode 'but corresponding to The direct mode of the segment is determined to be the time direct mode, and the time direct mode must be used for the large block, and the most suitable direct mode cannot be selected. In such a case, since the most suitable direct mode cannot be selected, it is necessary to encode an unnecessary motion vector, and there is a problem that the amount of coding increases. The present invention has been made in order to solve the above problems, and an object thereof is to obtain an image encoding device, a video decoding device, a video encoding method, and a video decoding method capable of selecting an optimum direct mode in units of predetermined blocks in 322763 7 201143455. . [Means for Solving the Problem] The video encoding device of the present invention is provided with a direct vector generating means for generating a spatial direct vector of a spatial direct mode from a motion vector of an encoded block located around a coding target block, and The time vector of the encoded direct face is generated in time near the coding target block to generate a temporal direct vector of the temporal direct mode; the evaluation value calculation means calculates the spatial direct using the spatial direct vector generated by the direct vector generation means The evaluation value of the mode, and using the time direct vector to calculate the evaluation value of the temporal direct mode; and the direct vector selection means, the evaluation value of the spatial direct mode calculated by the evaluation value calculation means and the evaluation value of the temporal direct mode For comparison, one of the spatial direct vector or the temporal direct vector is selected, and the predicted image generating means performs the motion compensated prediction process using the direct vector selected by the direct vector selecting means to generate the predicted image. [Effect of the Invention] According to the present invention, since the direct vector generation means is provided, a spatial direct vector of the spatial direct mode is generated from the motion vector of the coded block located around the coding target block, and from the time The motion vector of the coded facet near the coding target block is used to generate a temporal direct vector of the temporal direct mode; the evaluation value calculation means calculates the spatial direct mode using the spatial direct vector generated by the direct vector generation means 322763 201143455 The evaluation value is sufficient to use the time direct vector to calculate the time direct mode = estimate; R and the direct vector selection means to calculate the value and time of the spatial direct mode calculated by calculating the hand & The ratio of the flat mode of the direct mode & the selection of the spatial direct vector or the temporal direct vector and the predicted image generation means implement the motion compensation prediction process using the direct vector selected by the direct vector selection means to generate the prediction The block that is reserved by the image department selects the most suitable direct mode for the unit. As a result, the result is that the encoding of unnecessary moving vectors can be avoided, and the effect of increasing the amount of encoding is prevented. [Embodiment] Hereinafter, the present invention will be described in order to explain the present invention in more detail. (Embodiment 1) The video coding apparatus of the first embodiment of the video coding apparatus of the first embodiment is described as an example of the coding method used in the AVC/H.264 specification. In Fig. 1, the motion vector memory 1 is a recording medium such as a RAM for storing a motion vector of a coded large block (or a sub-block (sub_macr〇b 1〇ck) in which a large block is divided). . The motion compensation prediction unit 2 performs a process of selecting a reference image of a frame from a motion compensation prediction reference image stored in one of the frames of the frame memory (frame mem〇ry) 9, and then constructing The large block of the input image (or the sub-block formed by dividing the large block) is the unit performing/shifting 322763 9 201143455 motion compensation prediction processing, and generating the motion vector of the large block (coding block large block) to generate The image is predicted, and an identification number, a motion vector, a predicted image, and the like of the reference image selected for each of the large blocks are output. However, for convenience of explanation, it is assumed that the motion vector is generated in units of large blocks to generate a predicted image. However, when generating the motion vector of the large block to be encoded and generating the predicted video, the motion compensation prediction unit 2 performs the following processing for each of the large blocks constituting the input video: from the coded around the large block a large block motion vector (moving vector stored in the motion vector memory 1) to generate a spatial direct vector of the spatial direct mode, and a motion vector from the encoded picture that is temporally near the large block (stored in The motion vector in the vector memory 1 is moved to generate a temporal direct vector of the temporal direct mode. Then, the motion compensation prediction unit 2 performs a process of calculating an evaluation value of the spatial direct mode using the spatial direct vector, and calculating an evaluation value of the temporal direct mode using the time direct vector. Then, the motion compensation prediction section 2 performs a process of comparing the evaluation value of the spatial direct mode with the evaluation value of the temporal direct mode, and selecting either one of the spatial direct vector or the temporal direct vector. Then, the motion compensation prediction unit 2 performs a process of generating a prediction video by performing motion compensation prediction processing using the selected direct vector. The subtracter 3 performs the following processing: the difference image of the predicted image generated by the motion compensation prediction unit 2 and the input image is given to 10 322763 201143455 •=2= indicates that the difference difference signal of the difference image is output to the code=nuclear type The determination unit 4 performs a process of evaluating the prediction efficiency of the difference signal of the round (four) difference of the subtractor 3, and selecting the prediction efficiency most = two difference signals from among the more than one prediction difference signals of the subtractor 3, and then The prediction difference signal relates to a motion vector and a sub-block pattern used in the generation of the predicted image of the prediction unit 2 (including, for example, the -L: The output code of the inter mode or the direct mode refers to the code of the image as the coding mode information, and the signal is output to the compression unit 5 and the prediction difference compression unit 5 with the highest efficiency is implemented as follows: The difference is calculated by the DCT (four) T (discrete cosine transform) processed by the code pattern determining unit 4 to process the coefficient of the temple, and the (10) system is processed to treat the DCT coefficient (a pressure) The quantization unit is used to quantize the quantization unit 1 to μ. The local decoding unit 6 and the variable length, the subtractor 3, and the coding mode are quantized. The 卩4 and the compression unit 5 are localized. The decoding unit 6 performs the following processing. The reduced data is inversely quantized to obtain the inverse DCT (inverse discrete cosine transform) processing of the DCT output by the DCT ZHAO 邛 邛 5 "the tube ^ the 实 difference of the DCT coefficient real part 4 (four) 阙 difference The following equations are performed: The pre-determination adder 7 performs the following processing: The pre-circle 322763 11 201143455 calculated by the local decoding unit 6: the difference signal and the predicted video image generated by the motion compensation unit 2: =r By adding 'to generate a local decoding shadow solution: the chopper (1_filter) 8 performs the following processing: for the addition, the local image signal complemented by the coding distortion contained in the locally decoded video signal The F-wool image is output to the frame memory 9 » ^ memory 9 is used to store a recording medium such as a RAM of the loop chopper image. The τ and output degree encoding unit 1G performs the following processing: From _ section 5 number) into the 弁, 弋 move inward, ginseng The identification bit string 'Ά. entRDPy' of the test image encodes 'and generates a round π reamx_data indicating the entropy coding result'' and then flows the bit stream to the length coding unit 1Q to form a variable length coding The figure shows a configuration diagram of the motion compensation unit 2 in the embodiment J of the present invention. The image correction unit 21 receives the representation inter-surface mode = (real: outer 2 , motion vector, and 'quantity input: = ^ ^ direct vector generation unit 22 in the pile of 5 丨丨 local-station information, the implementation of the non-flat code mode for direct mode implementation: for each stone The large block of the horse object ' 322763 12 201143455 generates a spatial direct vector of the spatial direct mode and a temporal direct vector of the temporal direct mode, and outputs any one of the spatial direct vector or the temporal direct vector as a motion vector to the motion compensation process Part 23. The motion compensation processing unit 23 performs a process of performing motion compensation prediction processing using the motion vector output from the motion vector search unit 21 or the direct vector generation unit 22 and the reference image of a frame stored in the frame memory 9. The process of generating a predicted image. Further, the motion compensation processing unit 23 constitutes a predictive video generating means. Fig. 3 is a view showing the configuration of the direct vector generation unit 22 constituting the motion compensation prediction unit 2. In Fig. 3, the spatial direct vector generation unit 31 performs a process of encoding a large block located around the large block of the coding target from among the motion vectors of the coded large block stored in the motion vector memory 1. The motion vector is read out, and a spatial direct vector of the spatial direct mode is generated from the motion vector. The temporal direct vector generation unit 32 performs a process of shifting the motion vector of the encoded face temporally in the vicinity of the large block of the coding target from among the motion vectors of the coded large block stored in the motion vector memory 1, That is, the motion vector of the large block in the coded picture that is spatially located at the same position as the large block of the coding object is read, and the time direct vector of the temporal direct mode is generated from the motion vector. The spatial direct vector generation unit 31 and the temporal direct vector generation unit 32 constitute a direct vector generation means. The direct vector determination unit 33 performs a process of calculating the evaluation value of the spatial direct mode and the temporal direct vector generated by the temporal direct vector generation unit 32 using the spatial direct vector generated by the spatial direct vector 13 322763 201143455. To calculate the evaluation value of the time direct mode, and then compare the evaluation value of the spatial direct mode with the evaluation value of the temporal direct mode, and select either one of the spatial direct vector or the temporal direct vector. Fig. 4 is a view showing the configuration of the direct vector determining unit 33 constituting the direct vector generating unit 22. In the fourth diagram, the motion compensation unit 41 performs a process of generating a forward prediction video and a backward prediction video in the spatial direct mode using the spatial direct vector generated by the spatial direct vector generation unit 31, and the use time direct vector generation unit 32. The generated time direct vector is used to generate a forward predicted image and a backward predicted image in a temporal direct mode. The similarity calculation unit 42 performs a process of calculating the similarity between the forward predicted image and the backward predicted image in the spatial direct mode as the evaluation value of the spatial direct mode, and calculating the forward predicted image of the temporal direct mode and the backward predicted image. Degree, which is used as an evaluation value of the time direct mode. The motion compensation unit 41 and the similarity calculation unit 42 constitute an evaluation value calculation means. The direct vector selection unit 43 performs a process of performing the similarity between the forward predicted image and the backward predicted image in the spatial direct mode calculated by the similarity calculation unit 42 and the similarity between the forward predicted image and the backward predicted image in the temporal direct mode. In contrast, a direct vector of 14 322763 201143455 ίΠΐ with a higher degree of similarity between the forward-predicted image and the rear-predicted image is selected in the spatial direct vector or the temporal direct vector. The direct vector selecting unit 43 constitutes an example of a coding method employed by the video decoding device of Fig. 5 of the video decoding device in the present embodiment. H 264 specification
W中’移動向量記紐51係用來儲存已解碼大區 等J錄媒:°丨大£塊而成的次區塊)的移動向量之例如RAM 像:=:::==第1㈣ 位元串U編碼枓)的輸人,對於該 立几串机進賴解碼以解碼出壓縮資料及編碼模式資訊 號)U式/次區塊型式、移動向量、參考影像的識別編 料輸^至顏縣解碼部53,將該 二Ϊ 移動補償預測部54。可變長度解碼部 52係構成可變長度解嫣手段。 預測誤差解碼部53實施以下處理1可變長度解瑪部 52所輸出之壓縮資料予以逆量子化而求出㈣係數 對於該DCT係數實施逆Da處理,以算出表示差異影像之 預測誤差訊號(與第1圖之編碼模式狀部4所輸出的預、、則 =::預測誤差訊號)。預測誤差解碼部53係構 移動補償預測部54實施以下處理 _之一_上的參考影像之中,將== 322763 15 201143455 部52所輸出之識別號碼所表示的參考影 * :從可變長度解碼部52所輸出之大區塊型二出塊= 表示使用的是晝面間模式之料,使用從可 1 52所輸出之移動向量及上述參考影 又,*、。 測處理,藉此而生成預測影像。來實知移動補償預 另一方面,在從可變長度料部Μ所⑽之大 ^次旦區塊型絲錢㈣是直接模式之情況,則 的影像編碼裝置中的移動補償預測部2 · 處理:生成空間直接向量及時間直 ,’實施以下 間直接向量或時間直接向量的任°1然後選擇該空 接向量、及識別號碼所表示的參抑僅^用所選擇的直 測處理’藉此而生成預測影像。’f靶移動補償預 加法器55實施以下處理:將 成之預晴、及娜繼解㈣彳部54所生 訊號所表示之差異影料則目加,而彳輪出之預測誤差 第1圖之影像編碼裝£ +的加法 τ料影像(與 像相當者)之_影像訊號。° $之局部解媽影 匕路遽波器56實施以下處理: 的解碼影像訊號中所含的編碼失真進—補^器55所生成 失真經補償後之解碼影像訊號 二員,然後將編碼 影像予以儲存至圖框記憶體像當作參考 外部。 及將該解碼影像輪出至 二加=及迴路遽波器56係構成影像相加仲 匡3己憶體57係用來儲存從棘遽波器56所輪^ 322763 201143455 參考影像的RAM等記錄媒體。 第6圖係顯示本發明實施形態1中 移動補償預測部54之構成圖。 &像解碼裝置的 第6圖中,直接向量生成部61在 52所輸出之大區塊型式/次區塊型 4長又解碼部 模式之情況,實施以下處理:針對每—固下使用的疋直接 生成空間直接模式的空間直接向量碼對象大區塊, 間直接向量,並將該空間直接向量或時間=:式的時 方當作移動向量而輸出至移動補償處β里的任— 移動補償處理部62實施以下處理二。 體57之一圖框以上的參考影像之中,·破儲存於圖框記憶 52所輸出之識別號碼所表示的參考$將從可變長度解碼部 大區塊型式/次區塊型式係表示衫像予以讀出,然後在 況,使用從可變長度解碼部52彳用的是晝面間模式之情 考影像,來實施移動補償預測声出之移動向量及上述參 區塊型式/次區塊型式表示使、:理曰而生成預測影像,在大 直接向量生成部61所輸出的疋,接模式之情況,使用 實施移動補伽測處理而麵量及切參考影像,來 62係構成預測影像生解段。㈣影像。移動補償處理部 第7圖係顯示構成移 部61的構成圖。 員預測部54之直接向量生成 第7圖_,空間直接向息 儲存於移動向量記憶體51里生成部Η實施以下處理:從 令’將位於解碼對象大區=已解碼大區塊的移動向量之 龙的周園之已解媽大區塊的移動 322763 17 201143455 向量予以讀出,從該移動向量來生成空間直接模式的空間 直接向量。 時間直接向量生成部72實施以下處理:從儲存於移動 向量記憶體51之已解碼大區塊的移動向量之中,將為在時 間上處於解碼對象大區塊附近之已解碼晝面的移動向量, 且為該已解碼晝面中之在空間上位於與解碼對象大區塊相 同位置之大區塊的移動向量予以讀出,從該移動向量來生 成時間直接模式的時間直接向量。 空間直接向量生成部71及時間直接向量生成部72係 構成直接向量生成手段。 直接向量判定部73實施以下處理:使用由空間直接向 量生成部71所生成的空間直接向量,來算出空間直接模式 的評估值,並且使用由時間直接向量生成部72所生成的時 間直接向量,來算出時間直接模式的評估值,然後進行該 空間直接模式的評估值、與時間直接模式的評估值之比 較,而選擇該空間直接向量或時間直接向量的任一方。 第8圖係顯示構成直接向量生成部61之直接向量判定 部73之構成圖。 第8圖中,移動補償部81實施以下處理:使用由空間 直接向量生成部71所生成的空間直接向量來生成空間直 接模式的前方預測影像及後方預測影像,以及使用由時間 直接向量生成部72所生成的時間直接向量來生成時間直 接模式的前方預測影像及後方預測影像。 類似度算出部82實施以下處理:算出空間直接模式的 18 322763 201143455 前方預測影像與後方預測影像的類似度以 接模式的言情值,以及算出時間直接模式的前方預 與後方預測影像的類似度,以之作為時間直接模式的評^ 值。 手段移動補償部81及類似度算㈣82_成評估值算出 直接向量選擇部83實施以下虛,田.、 82所算出:空間直接模式中的前方預測: = =In W, the 'moving vector note 51 is used to store the J-recording medium such as the decoded large area: the sub-block formed by the block larger than the block. For example, the RAM image: =:::== 1st (four) bit The input of the meta-string U code 枓), for the vertical string machine to decode to decode the compressed data and coding mode information number) U-type / sub-block type, motion vector, reference image recognition and processing The Yanxian decoding unit 53 performs the second motion compensation prediction unit 54. The variable length decoding unit 52 constitutes a variable length decoding means. The prediction error decoding unit 53 performs inverse quantization on the compressed data output from the variable length demodulation unit 52 in the following processing 1 to obtain (4) coefficients, and performs inverse Da processing on the DCT coefficients to calculate a prediction error signal indicating the difference image (and The preamble outputted by the coding mode portion 4 of Fig. 1 =:: prediction error signal). The prediction error decoding unit 53 performs a reference image indicated by the identification number outputted by the == 322763 15 201143455 unit 52 among the reference pictures on the following processing_the following: * from the variable length The large block type two output block outputted by the decoding unit 52 indicates that the material used in the inter-plane mode is used, and the motion vector output from the 256 and the reference picture, *, are used. The measurement process is performed to generate a predicted image. On the other hand, in the case where the large-length block type money (4) of the variable length material unit (10) is in the direct mode, the motion compensation prediction unit 2 in the video encoding apparatus is provided. Processing: Generate spatial direct vector and time straight, 'implement the following direct vector or time direct vector 任°1 and then select the null vector, and the identification indicated by the identification number only use the selected direct measurement processing This produces a predicted image. The 'f target movement compensation pre-adder 55 performs the following processing: the difference image indicated by the signal generated by the pre-clearing and Na's solution (4) 彳 54 is added, and the prediction error of the 彳 wheel is shown in the first figure. The image code is loaded with the image of the additive image of the + (the image is equivalent). ° The local solution of the mother's shadow wave chopper 56 implements the following processing: the coded distortion included in the decoded image signal is generated by the complemented by the decoded image signal, and then the coded image is decoded. Store it in the frame memory image as a reference external. And the decoded image is rotated out to two plus= and the loop chopper 56 is formed into an image plus a 匡 匡 己 己 己 己 己 己 己 57 57 57 57 57 57 57 322 322 322 322 322 322 322 322 322 322 322 322 322 322 322 322 322 322 322 322 322 322 322 322 media. Fig. 6 is a view showing the configuration of the motion compensation prediction unit 54 in the first embodiment of the present invention. In the sixth image of the image decoding apparatus, the direct vector generation unit 61 performs the following processing in the case of the large block type/sub-block type four-length decoding unit mode outputted by 52: for each solid use疋 Directly generate spatial direct vector space direct vector code object large block, direct vector, and output the space direct vector or time =: as the motion vector to the mobile compensation β - mobile The compensation processing unit 62 performs the following process two. Among the reference images above the frame of the body 57, the reference $ indicated by the identification number outputted in the frame memory 52 will be expressed from the variable length decoding unit large block type/sub-block type system. For example, the motion vector of the motion compensated prediction sound and the reference block type/sub-block are implemented using the emotion image of the inter-plane mode using the variable length decoding unit 52. The pattern indicates that the predicted image is generated by the large direct vector generation unit 61. When the pattern is output from the large direct vector generation unit 61, the surface amount and the reference image are subjected to the motion compensation gamma processing to form the predicted image. Deconstruction section. (4) Imagery. Movement compensation processing unit Fig. 7 is a view showing a configuration of the configuration shifting unit 61. The direct vector of the member prediction unit 54 generates a picture _, and the space is directly stored in the motion vector memory 51. The generation unit performs the following processing: from the order of 'the decoding object large area=the decoded large block's motion vector The Dragon's Zhou Yuanzhi has solved the movement of the big block 322763 17 201143455 The vector is read out from the moving vector to generate a spatial direct vector of the spatial direct mode. The temporal direct vector generation unit 72 performs a process of shifting the motion vector of the decoded face that is temporally in the vicinity of the large block of the decoding object from among the motion vectors stored in the decoded large block of the motion vector memory 51. And reading a motion vector of the large block in the decoded decoding plane that is spatially located at the same position as the decoding target large block, and generating a temporal direct vector of the temporal direct mode from the motion vector. The spatial direct vector generation unit 71 and the temporal direct vector generation unit 72 constitute a direct vector generation means. The direct vector determination unit 73 performs a process of calculating the evaluation value of the spatial direct mode using the spatial direct vector generated by the spatial direct vector generation unit 71, and using the temporal direct vector generated by the temporal direct vector generation unit 72. The evaluation value of the time direct mode is calculated, and then the evaluation value of the spatial direct mode and the evaluation value of the direct time mode are compared, and either one of the spatial direct vector or the temporal direct vector is selected. Fig. 8 is a view showing the configuration of the direct vector determining unit 73 constituting the direct vector generating unit 61. In the eighth diagram, the motion compensation unit 81 performs a process of generating a forward prediction video and a backward prediction video in the spatial direct mode using the spatial direct vector generated by the spatial direct vector generation unit 71, and using the temporal direct vector generation unit 72. The generated temporal direct vector generates a forward predicted image and a backward predicted image in a temporal direct mode. The similarity calculation unit 82 performs a process of calculating the similarity between the forward predicted image and the backward predicted image in the spatial direct mode to calculate the similarity between the forward mode and the backward predicted image, and the similarity between the forward predicted image and the backward predicted image in the temporal direct mode. Use it as the evaluation of the time direct mode. Means movement compensation unit 81 and similarity calculation (4) 82_ into evaluation value calculation The direct vector selection unit 83 calculates the following virtual, Tian., 82: forward prediction in the spatial direct mode: ==
像的類似度、以及時間直接模式中 M 預測影像的類似度之比較,而選擇^酬影像與後方 接向量之中前方預,與後,: = = ίΠί的直接向里。直接向量選擇部83係構成直接向量 第1圖中,雖然將影像編碼裝 移動補償預測部2、減法器3、蝙 #成要素亦即 5、局部解瑪部6、加法器7、^^^定部4、_部 崎’愿波器8及可變i 碼部丨。設想為分別由專用的硬體(例如:安裝 導體積體電路、或者單晶片微電腦(〇ne C叫 半 Γ,ΓΓ可所構成,但影像編碼裝置若由電腦所構 i法=形態:將記述有移動補償預測部2、 加法器/編式判㈣4、壓縮部5、局部解碼部6、 之程·^㈣波器8及可變長度編碼部1G的處理内容 儲存於讀3 ==記^ ’再由^腦的CPU來執行 322763 19 201143455 處理==本發明實施形態1中的影像編碼裝置的 第5圖_ ’賴將影像解碼裝置 可變長度解碼部解碼部^ f成要素,亦即 5[加法器55及迴㈣波器%設乂移_償預澈部 =如:安裝有m;之半導體積體電路;^由專用的硬體 荨)所構成,但爹僮鉉成壯 或者早晶片微電腦 以下形態:將記^了 ^置若由電腦所構成,則可形成為 53、移動補償預、、=I 解碼部52、獅誤差解瑪部 理内容之程;==加法器55及迴路滤咖The degree of similarity of the image, and the comparison of the similarity of the M predicted images in the direct mode of time, and the selection of the front image of the image and the rear vector, and the following: = = ίΠί direct inward. The direct vector selection unit 83 constitutes a direct vector in the first drawing, and the video encoding device is subjected to a motion compensation prediction unit 2, a subtractor 3, a bat# component, that is, a local gamma unit 6, an adder 7, and a ^^^ Fixed part 4, _Miyazaki's wave device 8 and variable i code part. It is assumed that the hardware is dedicated (for example, a guided volume circuit or a single-chip microcomputer), but the image coding device is constructed by a computer. The processing contents of the motion compensation prediction unit 2, the adder/pattern determination (4) 4, the compression unit 5, the local decoding unit 6, the process (4) wave device 8 and the variable length coding unit 1G are stored in the read 3 == record ^ 'By the CPU of the brain, 322763 19 201143455 Process==Fig. 5 of the video encoding device according to the first embodiment of the present invention _ 'The video decoding device variable length decoding unit decoding unit is an element, that is, 5[Adder 55 and back (four) wave device % set to shift _ compensation pre-clear part = such as: installed m; semiconductor integrated circuit; ^ by a dedicated hardware 荨), but the child is strong or The early wafer microcomputer has the following form: if it is composed of a computer, it can be formed as a process of 53, a motion compensation pre-, a =I decoding unit 52, and a lion error solution; == adder 55 Loop filter
來執行儲存於該記憶體,再由該電腦的CPU 處^本發明實施形態1中的影像解碼裝置的 接著針對動作進行說明。 1圖之影像編碼裝置之處理内容。 號之輪入時員則°卩β2,在有表示輸入影像之動態影像訊 訊號的各圖樞塊(或次區塊)為單位將該動態影像 次區塊2將動態影像訊號分割為大區塊(或 移動補償_用之-圖框以上的 然後以大區射·^ 像中選擇一圖框之參考影像, ^成心二為二執):=,處理’ 預剛影像。 人區塊)的移動向量,進而生成 322763 20 201143455 移動補償預測部2生成編碼對象大區塊(或次區塊)的 移動向量而且生成預測影像,就將該預測影像輸出至減法 器3,以及將該預測影像之生成中用到的移動向量、大區 塊型式/次區塊型式(包含例如,表示在該大區塊(或次區塊) 使用的編碼模式係為畫面間模式或直接模式的何者之資訊) 以及參考影像的識別號碼予以輸出至編碼模式判定部4。 以下,具體地說明移動補償預測部2的處理内容。 不過,此處為了說明之方便,假設其為以大區塊為單 位而生成移動向量進而生成預測影像者。 移動補償預測部2的移動向量搜尋部21在接收到表示 編碼模式為晝面間模式之資訊(從例如外部接收表示使用 的是晝面間模式之資訊)時,就以晝面間模式搜尋最適合的 移動向量,並將該移動向量輸出至移動補償處理部23。 以晝面間模式搜尋最適合的移動向量之處理本身為公 知的技術,故省略其詳細的說明。 移動補償預測部2的直接向量生成部22在接收到表示 編碼模式為直接模式之資訊時,就針對每一個編碼對象大 區塊,生成空間直接模式的空間直接向量、及時間直接模 式的時間直接向量,並將該空間直接向量或時間直接向量 的任一方當作移動向量而輸出至移動補償處理部23。 亦即,直接向量生成部22的空間直接向量生成部31 從儲存於移動向量記憶體1之已編碼大區塊的移動向量之 中,將位於編碼對象大區塊的周圍之已編碼大區塊的移動 向量予以讀出,從該移動向量來生成空間直接模式的空間 21 322763 201143455 直接向量(笛 、弟12圖中之步驟ST1)。 從儲广外,直接向量生成部22的時間直接向量生成部32 中:;移動向量記憶體1之已編碼大區塊的移動向量之 必&:在時間上處於編碼對象大區塊附近之已編碼晝面的 移動向量,、 0 亚且為該已編碼晝面中之在空間上位於與編碼 塊相同位置之大區塊的移動向量予以讀出,從該 向里來生成時間直接模式的時間直接向量(步驟ST2)。 第9圖係顯示以時間直接模式來生成移動向量(時間 直接向量)之方法之示意圖。 〇假μ為例如:晝面Β2之中的大區塊ΜΒ1為編碼對象大 區塊’且要以時間直接模式將大區塊ΜΒ1予以編碼之情況。 此情況,係使用在時間軸上位於晝面Β2的後方之已編 碼晝,之中最靠近晝面Β2之晝面Ρ3的移動向量,且該移 動向量為該晝面Ρ3中之在空間上位於與大區塊腿相同位 置之大區塊ΜΒ2的移動向量mv。 此移動向量MV係參考到晝面ρ〇,且進行大區塊Μβ1 的編碼之際要用到的移動向量MVLO, MVL1,係以以下的式 (3)求出。 MVL0 = MVL\^ T2rT0 Π~Γ3 Τ^ΤΟThe operation of the video decoding device according to the first embodiment of the present invention, which is stored in the memory, is described below. The processing content of the image coding device of Fig. 1. When the wheel is in, the player is °卩β2, and the dynamic image sub-block 2 divides the motion image signal into a large area in units of pivots (or sub-blocks) of the motion picture signal indicating the input image. Block (or mobile compensation _ use - above the frame and then select a frame of the reference image in the large area ^ ^ image, ^ Cheng Xin two for the second): =, processing 'pre-rigid image. a motion vector of the human block, and further a 322763 20 201143455 motion compensation prediction unit 2 generates a motion vector of a coding target large block (or a secondary block) and generates a predicted image, and outputs the predicted image to the subtractor 3, and The motion vector used in the generation of the predicted image, the large block type/sub-block pattern (including, for example, the coding mode used in the large block (or the secondary block) is an inter-picture mode or a direct mode. The information of the reference image and the identification number of the reference image are output to the encoding mode determining unit 4. Hereinafter, the processing content of the motion compensation prediction unit 2 will be specifically described. However, for convenience of explanation, it is assumed that the motion vector is generated in units of large blocks to generate a predicted image. When the motion vector search unit 21 of the motion compensation prediction unit 2 receives the information indicating that the coding mode is the inter-plane mode (for example, receiving information indicating that the inter-plane mode is used from the outside), the motion search mode is the most searched by the inter-plane mode. A suitable motion vector is output to the motion compensation processing section 23. The process of searching for the most suitable motion vector in the inter-plane mode is a well-known technique, and a detailed description thereof will be omitted. When receiving the information indicating that the encoding mode is the direct mode, the direct vector generating unit 22 of the motion compensation predicting unit 2 generates the spatial direct vector of the spatial direct mode and the time of the temporal direct mode for each large block of the encoding target. The vector is output to the motion compensation processing unit 23 as any one of the spatial direct vector or the temporal direct vector as a motion vector. That is, the spatial direct vector generation unit 31 of the direct vector generation unit 22, from among the motion vectors stored in the coded large block of the motion vector memory 1, the coded large block located around the large block of the coding target. The motion vector is read out, and the space of the spatial direct mode is generated from the motion vector. 21 322763 201143455 Direct vector (step ST1 in the flute, 12). From the time of the storage, the direct vector generation unit 32 of the direct vector generation unit 22: • the motion vector of the coded large block of the motion vector memory 1 must be &: in the vicinity of the large block of the coding object in time The motion vector of the encoded face, 0, and the motion vector of the large block in the encoded face that is spatially located at the same position as the coded block is read, from which the time direct mode is generated. Time direct vector (step ST2). Figure 9 is a schematic diagram showing a method of generating a motion vector (time direct vector) in a temporal direct mode. The false μ is, for example, a case where the large block ΜΒ 1 in the face Β 2 is the coding target large block ’ and the large block ΜΒ 1 is to be coded in the time direct mode. In this case, the coded 昼 located at the rear of the facet Β 2 on the time axis is used, and the motion vector closest to the face Ρ 3 of the face Β 2 is located, and the motion vector is spatially located in the face Ρ 3 The motion vector mv of the large block ΜΒ2 at the same position as the large block leg. The motion vector MV is referred to the pupil plane ρ〇, and the motion vectors MVLO, MVL1 to be used when encoding the large block Μβ1 are obtained by the following equation (3). MVL0 = MVL\^ T2rT0 Π~Γ3 Τ^ΤΟ
xMVXMV
MV 3 時間直接向篁生成部32算出移動向量MVL〇,MVU, 就將該移動向量MVLO, MVL1作為時間直接模式的時間直接 向量而輸出至直接向量判定部33。 322763 22 201143455 時間直接向量生成部32 旦 法,雖可如上述採用如第9 ,曰α里的生成方 於此,亦可採用其他的方法:不之11. 264方式,但不限 直接=圖係顯示以空間直接模式來生成移動向量卜門 直接向量)之方法之示意圖。 里(二間 :二圖中,current MB表示編碼對象大區塊。 此時’假設編石馬對象大 塊A的移動向量為MV : =邊的已編竭大區 已編^量為啊,編碼對象大區塊之右上方的 示,&塊C的移動向量為MVc,則如以下的式(4)所 專移動向里MVa,MVb,MVc的中央值(median)求 Mv^尤可算出編碼對象大區塊的移動向量MV。 m e d 1 a Π (Mv a、MVb、MVc) (4) 在空間直接模式中,雖然要針對前方及後方分別求出 移動向量,但不論是哪一方都可用上述的方法來求出。 I間直接向量生成部31如上述算出前方及後方的移 動向量MV ’就將該前方及後方的移動向量MV作為空間直 接模式的空間直接向量 而輸出至直接向量判定部33。 空間直接向量生成部31中之空間直接向量的生成方 〆’雖可如上述採用如第1〇圖所示之Η. 264方式,但不 限於此’亦可採用其他的方法。 直接向量生成部22的直接向量判定部33在空間直接 =量生成部31生成空間直接向量時,就使用該空間直接向 里來算出空間直接模式的評估值。 23 322763 201143455 此外,直接向量判定部33在時間直接向量生成部32 生成時間直接向量時,就使用該時間直接向量來算出時間 直接模式的評估值。 直接向量判定部33接著進行該空間直接模式的評估 值與時間直接模式的評估值之比較,且從該空間直接向量 或時間直接向量之中,選擇藉由如以下記載的判定手段而 判定出之直接模式的直接向量並將之輸出至移動補償處理 部23。 以下,具體地說明直接向量判定部33的處理内容。 直接向量判定部33的移動補償部41在空間直接向量 生成部31生成空間直接向量MVLO, MVL1時,就如第11圖 所示,使用該空間直接向量MVL0來生成空間直接模式的前 方預測影像fspatial,且使用該空間直接向量MVL1來生成空 間直接模式的後方預測影像gSPatial(步驟ST3)。 此外,移動補償部41在時間直接向量生成部32生成 時間直接向量(亦即前方及後方的移動向量MV)時,就如第 11圖所示,使用該前方的移動向量MV來生成時間直接模 式的前方預測影像f^poral,且使用該後方的移動向量MV來 生成時間直接模式的後方預測影像gte^ral(步驟ST4)。 直接向量判定部33的類似度算出部42,在移動補償 部41生成空間直接模式的前方預測影像fSPatial及後方預測 影像gspatial時,就算出該前方預測影像f spatial與後方預測影 像gspatial的類似度並以之作為空間直接模式的評估值 SADspatial(步驟 ST5) 0 24 322763 201143455 SADSpatial — If spatial δ spatial I (5) 此外,類似度算出部42在移動補償部41生成時間直 接模式的前方預測影像f丨啊。…及後方預測影像gtemp(3ra丨時, 就算出該前方預測影像f temp〇ral與後方預測影像gtemp()ral的類 似度並以之作為時間直接模式的評估值SADte^ra!(步驟 ST6)。 SADlenlporal — Iftcmporal Stemporal I (6) 前方預測影像與後方預測影像間的差愈大,兩個影像 間的類似度就變低(表示兩個影像的差的絕對值和之評估 值SAD變大),時間性的相關就變低。反之,前方預測影像 與後方預測影像間的差愈小,兩個影像間的類似度就變高 (表示兩個影像的差的絕對值和之評估值SAD變小),時間 性的相關就變高。 直接向量判定部33的直接向量選擇部43,在類似度 算出部42算出空間直接模式的評估值SADspatial及時間直接 模式的評估值SADte^radf,就進行該評估值SADSpatial與評 估值SADtemporal之比較’來比較空間直接模式中的前方預測 影像f spatial與後方預測影像gspatial的類似度、與時間直接模 式中的前方預測影像fteDlpc)ral&後方預測影像gtelnpc)ral的類似 度(步驟ST7)。 直接向量選擇部43,在空間直接模式中的前方預測影 像f SPatial與後方預測影像gspatial的類似度比時間直接模式 中的前方預測影像 f temporal 及後方預測影像gtemp〇raI的類似度 高之情況(3人0_131$540_—),選擇空間直接向量生成部 25 322763 201143455 3i所生成之空間直接向量,並將該空間直接向量作為移動 向量而輸出至移動補償處理部23(步驟ST8)。 /另一方面,在時間直接模式中的前方預測影像f_ai 及後方預鄉像gteinp(3I_al的類似度tb空間直接模式中的前方 預測影像以1與後方賴影像gs_義似度高之情況 (SADSPatial>SADtempQral),選擇時間直接向量生成部犯所生成 之時間直接向量’並將該時間直接向量作為移動向量而輸 出至移動補償處理部23(步驟ST9)。 移動補彳員處理部23在編碼模式並非直接模式時(步驟 stio),若接收到從移動向量搜尋部21傳來的移動向量, 則使用4移動向篁及儲存於圖框記憶體9中的_圖框的參 考影像來實施移動補償預測處理,而生成賴影像 ST11) 〇 另一方面,在編碼模式為直接模式時(步驟ST10),若 接,到從直接向量生成部22傳來的移動向量(直接向量選 擇部/3所選擇的直接向量),則使用該移動向量及儲存於 圖杧η己It體9中的—圖框的參考影像,來實施移動補償預 測處理,而生成預測影像(步驟ST12)。 一移動補償處理部23之移動補償預測處理為公知的技 術’故省略其詳細的說明。 此處,顯示的雖然是類似度算出部42算出差的絕對值 寿來作為3平估值SAD,直接向量選擇部⑸進行評估值⑽ :1匕較之形態’但亦可為類似度算出部4 2算出前方預測影 Ί方預測影像間之差的平方和卿來作為評估值,直 322763 26 201143455 接向量選擇部43進行差的平方和SSE的比較之形態。使用 SSE雖會使處理量增加,但可更正確地算出類似度。 減法器3在移動補償預測部2生成預測影像時,就算 出該預測影像與輸入影像的差異影像,並將表示該差異影 像之預測差異訊號輸出至編碼模式判定部4(步驟ST13)。 編碼模式判定部4每次接收到從減法器3傳來的預測 差異訊號,都對於該預測差異訊號的預測效率進行評估, 而在減法器3所輸出的至少一個以上的預測差異訊號之 中,選出預測效率最高的預測差異訊號。 編碼模式判定部4中之評估預測差異訊號的預測效率 之處理本身為公知的技術,故省略其詳細的說明。 編碼模式判定部4選出效率最高的預測差異訊號,就 將編碼模式資訊〔其中包含:與該預測差異訊號有關之曾 在移動補償預測部2的預測影像的生成中使用到的移動向 量、大區塊型式/次區塊型式(包含例如,表示在該大區塊 使用的編碼模式係為畫面間模式或直接模式的哪一者之資 訊)、以及參考影像的識別號碼〕輸出至可變長度編碼部 10。 此外,編碼模式判定部4將預測效率最高之預測差異 訊號輸出至壓縮部5(步驟ST14)。 編碼模式判定部4只要是在編碼模式為畫面間模式的 情況,就將曾在預測影像的生成中使用到的移動向量包含 在編碼模式資訊中,然後將包含該移動向量之編碼模式資 訊輸出至可變長度編碼部10,但在編碼模式為直接模式的 27 322763 201143455 情況’則並不將曾在預測影像的生成 包含在編碼模式資訊中,而將不包含今用至】的移動向量 式資訊輸出至可變長度編碼部1〇。 移動向量之編碼模 壓縮部5接收到從編碼模式判定部4 訊號,就對於該預測差異訊號實施DCT處王來的預測差異The MV 3 time directly calculates the motion vector MVL 〇, MVU from the 篁 generating unit 32, and outputs the motion vector MVLO, MVL1 as the temporal direct vector of the temporal direct mode to the direct vector determining unit 33. 322763 22 201143455 The time direct vector generation unit 32 can be used as described above in the ninth method, and other methods can be used: no. 11. 264 mode, but not limited to direct = map A schematic diagram showing a method of generating a mobile vector gate direct vector in a spatial direct mode. In the two (two: two pictures, current MB represents the large block of the encoding object. At this point, 'Assume that the moving vector of the large block A of the stone object is MV : = the edited area of the side has been compiled, ah, The upper right side of the coding object block, and the motion vector of the block C is MVc, then the median value of the media MVa, MVb, MVc (M) is as follows in the following equation (4). Calculate the motion vector MV of the large block of the encoding target. med 1 a Π (Mv a, MVb, MVc) (4) In the spatial direct mode, the motion vector is obtained for the front and the rear, respectively, but either side The inter-I direct vector generation unit 31 calculates the forward and backward motion vectors MV' as the spatial direct vectors of the spatial direct mode and outputs them to the direct vector determination as described above. Part 33. The generation method of the spatial direct vector in the spatial direct vector generation unit 31 may be as described above with reference to the 264 method as shown in Fig. 1, but is not limited thereto. Other methods may be employed. Direct vector determination unit 3 of vector generation unit 22 3. When the spatial direct-quantity generation unit 31 generates the spatial direct vector, the spatial direct mode evaluation value is calculated directly from the space. 23 322763 201143455 Further, the direct vector determination unit 33 generates the time in the temporal direct vector generation unit 32. When the direct vector is used, the time direct vector is used to calculate the evaluation value of the temporal direct mode. The direct vector determining unit 33 then performs the comparison of the evaluation value of the spatial direct mode with the evaluation value of the temporal direct mode, and directly vector or from the space. Among the time direct vectors, the direct vector of the direct mode is determined by the determination means described below and output to the motion compensation processing unit 23. Hereinafter, the processing content of the direct vector determining unit 33 will be specifically described. When the spatial direct vector generation unit 31 generates the spatial direct vectors MVLO and MVL1, the spatial direct vector generation unit 31 generates the forward prediction image fspatial of the spatial direct mode using the spatial direct vector MVL0 as shown in FIG. And use the spatial direct vector MVL1 to generate the back of the spatial direct mode The image gSPatial (step ST3). When the temporal direct vector generation unit 32 generates the temporal direct vector (i.e., the front and rear motion vectors MV), the motion compensation unit 41 uses the forward motion as shown in FIG. The vector MV generates the forward predicted image f^poral in the temporal direct mode, and generates the backward predicted image gte^ral in the temporal direct mode using the rear motion vector MV (step ST4). The similarity calculating unit of the direct vector determining unit 33 42. When the motion compensation unit 41 generates the forward prediction video fSPatial and the backward prediction image gspatial in the spatial direct mode, the similarity between the forward prediction video f spatial and the backward prediction image gspatial is calculated and used as the evaluation value of the spatial direct mode SADspatial (Step ST5) 0 24 322763 201143455 SADSpatial — If spatial δ spatial I (5) Further, the similarity calculation unit 42 generates a forward predicted image f时间 in the temporal direct mode in the motion compensation unit 41. ... and the rear predicted image gtemp (3ra丨, the similarity between the forward predicted image f temp〇ral and the backward predicted image gtemp() ral is calculated and used as the evaluation value SADte^ra! of the time direct mode (step ST6) SADlenlporal — Iftcmporal Stemporal I (6) The greater the difference between the forward predicted image and the backward predicted image, the lower the similarity between the two images (indicating that the absolute value of the difference between the two images and the evaluation value SAD become larger) The temporal correlation becomes lower. Conversely, the smaller the difference between the forward predicted image and the backward predicted image, the higher the similarity between the two images becomes higher (indicating the absolute value of the difference between the two images and the evaluation value SAD becomes The direct vector selection unit 43 of the direct vector determination unit 33 calculates the evaluation value SADspatial of the spatial direct mode and the evaluation value SADte^radf of the temporal direct mode by the similarity calculation unit 42. Comparing the evaluation value SADSpatial with the evaluation value SADtemporal' to compare the similarity between the forward prediction image f spatial and the backward prediction image gspatial in the spatial direct mode, and the direct mode of time The degree of similarity of the forward predicted image fteDlpc)ral&back predicted image gtelnpc)ral in the equation (step ST7). The direct vector selection unit 43 has a higher degree of similarity between the forward predicted image f SPatial and the backward predicted image gspatial in the spatial direct mode than the forward predicted image f temporal and the backward predicted image gtemp 〇raI in the temporal direct mode ( The three persons 0_131$540_-) select the spatial direct vector generated by the spatial direct vector generating unit 25 322763 201143455 3i, and output the spatial direct vector as the motion vector to the motion compensation processing unit 23 (step ST8). On the other hand, in the time direct mode, the forward predicted image f_ai and the rear predicted image like gteinp (the similarity of the 3I_al similarity in the tb spatial direct mode is higher in the forward predicted image with 1 and the rearward image gs_ ( SADSPatial>SADtempQral) selects the time direct vector generated by the temporal direct vector generation unit and outputs the temporal direct vector to the motion compensation processing unit 23 as a motion vector (step ST9). The mobile complement processing unit 23 encodes When the mode is not the direct mode (step stio), if the motion vector transmitted from the motion vector search unit 21 is received, the movement is performed using the reference image of the _ frame stored in the frame memory 9 using 4 movements. On the other hand, when the coding mode is the direct mode (step ST10), the motion vector transmitted from the direct vector generation unit 22 (direct vector selection unit/3) is generated. The selected direct vector is used to implement the motion compensation prediction process using the motion vector and the reference image stored in the frame of the image Test image (step ST12). The motion compensation prediction process of the motion compensation processing unit 23 is a well-known technique, and detailed description thereof will be omitted. Here, the similarity calculation unit 42 calculates the absolute value of the difference as the three-level estimate SAD, and the direct vector selection unit (5) performs the evaluation value (10): 1 匕 compared to the form 'but the similarity calculation unit may be used. 4 2 Calculate the sum of the squares of the differences between the predicted images of the forward predictive influences as the evaluation values, and 322763 26 201143455 The vector selection unit 43 compares the squares of the differences and the SSE. Although the use of SSE increases the throughput, the similarity can be calculated more correctly. When the motion compensation prediction unit 2 generates the predicted video, the subtracter 3 calculates the difference video between the predicted video and the input video, and outputs the predicted difference signal indicating the difference image to the encoding mode determining unit 4 (step ST13). Each time the coding mode determination unit 4 receives the prediction difference signal transmitted from the subtractor 3, the prediction efficiency of the prediction difference signal is evaluated, and among the at least one prediction difference signal output by the subtractor 3, The predicted difference signal with the highest prediction efficiency is selected. The process of evaluating the prediction efficiency of the prediction difference signal in the coding mode determination unit 4 is a well-known technique, and detailed description thereof will be omitted. The coding mode determination unit 4 selects the most significant prediction difference signal, and the coding mode information includes the motion vector and the region used in the generation of the prediction image of the motion compensation prediction unit 2 related to the prediction difference signal. The block type/sub-block type (including, for example, information indicating whether the coding mode used in the large block is the inter-picture mode or the direct mode), and the identification number of the reference picture are output to the variable length coding. Department 10. Further, the coding mode determination unit 4 outputs the prediction difference signal having the highest prediction efficiency to the compression unit 5 (step ST14). The coding mode determination unit 4 includes the motion vector used in the generation of the predicted video in the coding mode information, and outputs the coding mode information including the motion vector to the case where the coding mode is the inter mode. The variable length coding unit 10, but the case where the coding mode is the direct mode, 27 322763 201143455, does not include the generation of the predicted image in the coding mode information, but does not include the mobile vector information of the current use. Output to the variable length coding unit 1〇. The coding mode compression unit 5 of the motion vector receives the signal from the coding mode determining unit 4, and performs the prediction difference of the DCT at the prediction difference signal.
係數,而且將該DCT係數予以/ κ处理,以算出DCT 至丁化〔步驟 壓縮部5將為量子化後DCT係數的 =)。 部解碼部6及可變長度編碼部10 ^ 為料輪出至局The coefficient, and the DCT coefficient is /κ processed to calculate the DCT to the singulation [the step compression unit 5 will be the quantized DCT coefficient =). Part decoding unit 6 and variable length coding unit 10 ^
當局部解碼部Θ接收到從壓縮部5 則將該壓縮資料予以逆量子化來求出d來的壓縮資料, 係數實施逆DCT處理,以算出與 倾,再對該DCT 的預測差異訊號相當的預測誤差訊號。、"判定部4所輪出 加法器7在局部解碼部6解碼出 =誤差訊號、與表示由移動補償_部、=時’將 剃〜像之預測訊號予以相加,來生 p 2所生成的預 局部解碼影像訊號。 不局部解碼影像之 迴路濾波器8為了之後的 :輸出的料解鄕像訊號㈣Μ於加法器7 =將_償編碼失真後之局部 &失真進行補償, 德喝影像當作參切料 =^_表示的局 可變長度編竭部^ ,存至圖樞錢體9。 料’就對於簡縮資料 Π缩部5傳來的壓縮 喝模式資訊〔大區塊「動補償預測部2所輸出之蝙 式為晝面間模式的情況)二區塊型式、移動向量(編石馬模 /)、參考影像的識別編號〕進行墒編 322763 28 201143455 碼,而生成表示該編碼結果 _ 流予以輪出(步驟sti6)。 位元串流,然後將該位元串 接著’說明第5圖之影 當可變長度解碼部52解蝎裝置的處理内容。 所輸出之位元串流的輪入時!從第1圖之景>像編碼裝置 碼以解碼出壓縮資 料及編t則對於該位元串流進行熵解 塊型式、移動向量(蝙碼模式1式資訊〔大區塊型式/次區 影像的識別編號〕,然後將'分、4面間模式的情況)、參考 碼部53,將襲碼模式資訊二壓㈣料輸以預測誤差解 13圖中之步驟π? 1)。 § ;出至移動補你預測部54(第 當預測誤差解碼部53接 來的壓縮資料時,就將 J從可變長度解碼部52傳When the local decoding unit Θ receives the compressed data from the compression unit 5, the compressed data is inverse quantized to obtain d, and the coefficient is subjected to inverse DCT processing to calculate the tilt and the predicted difference signal of the DCT. Prediction error signal. And the determination unit 4 rotates the adder 7 to decode the = error signal in the local decoding unit 6, and adds the predicted signal of the shaved image to the image indicating that the motion compensation_part, ==, to generate the p 2 The generated pre-local decoded video signal. The loop filter 8 that does not locally decode the image is used for the following: the output material is decoded by the image signal (4), and the adder 7 = compensates for the local &litude distortion after the compensation of the code, and the image of the drink is used as the reference material =^ The variable length length of the _ indicated by the _ is stored in the figure. The material is 'compressed drinking mode information transmitted to the constricted data contraction section 5 (the block type "the bat type output by the dynamic compensation prediction unit 2 is the inter-plane mode"), the second block type, the motion vector (the stone) Mamo/), the identification number of the reference image] is encoded 322763 28 201143455 code, and the generation indicates that the encoding result _ stream is rotated (step sti6). The bit stream is streamed, and then the bit string is followed by 'description number The shadow of the figure 5 is used by the variable length decoding unit 52 to decode the processing contents of the device. When the bit stream of the output bit is rounded up! From the scene of the first picture > the encoding device code to decode the compressed data and edit Then, the entropy deblocking pattern and the motion vector are performed on the bit stream (the bat code mode 1 type information [large block type/sub-region image identification number], and then the 'minute, 4-face mode mode'), reference The code portion 53 transmits the attack code mode information (four) material to the step π? 1) in the prediction error solution 13; §; to the mobile complement prediction unit 54 (the compression received by the first prediction error decoding unit 53) When the data is obtained, J is transmitted from the variable length decoding unit 52.
係數,然後對於該DCT係、數:4予以❹子化而求出DCT 示差異影像之關縣喊(與HDGT處理,藉以算出表 所輸出的_差異職相當之預圖模式判定部4 當移動補償預測部54^=1^步驟咖)。 來的失t 使叹勻從可憂長度解碼部52傳 一:〜象的識別編號時,則從儲存於圖框記憶體Μ: 上的參考影像之中,將該識別號碼所表 衫像予以讀出。 巧 此外,當移動補償預測部54接收到從可變長度解喝部 52傳來的大區塊型式μ區塊型柄,财相大區塊型 式/次區塊型式,判別第1圖之影像編碼裝置係使用晝面間 模式來作為編媽模式,還是使用直接模式來作為編碼模式 (步驟 ST23)。 、 322763 29 201143455Then, the DCT of the DCT system and the number of 4 are obtained, and the DCT shows the difference image of the Guanxian shout (with the HDGT processing, the pre-image mode determination unit 4 that calculates the _ difference of the output of the table is calculated. The compensation prediction unit 54^=1^step coffee). When the sigh is lost, the sigh is transmitted from the sizable length decoding unit 52: the identification number of the image is displayed from the reference image stored in the frame memory Μ: read out. In addition, when the motion compensation prediction unit 54 receives the large block type μ block type handle transmitted from the variable length debunking unit 52, the financial phase large block type/subblock type, and the image of the first picture is discriminated. The encoding device uses the inter-panel mode as the editing mode or the direct mode as the encoding mode (step ST23). , 322763 29 201143455
移動補償預測部54在笛1 -面間模式來作為編碼模式 所輸出之移動向量及上述 處理,而生成預測影像G 另一方面,在第1圖之影像編碼裝置 來作為編碼模式之情況, 移動補償預測部2 -樣,生成空間直接向量及時間直接^ 量,然後選擇該空間直接向量或時間直接向量的任_方°The motion compensation predicting unit 54 generates the predicted video G as the motion vector outputted in the encoding mode and the above-described processing in the flute 1-interface mode. On the other hand, when the video encoding apparatus of the first image is used as the encoding mode, the motion compensation unit 54 moves. The compensation prediction unit 2 generates a spatial direct vector and a time direct quantity, and then selects the spatial direct vector or the time direct vector.
實施移動補償預測處理而生成預測影像(步驟ST25)。 以下’具體地說明移動補償預測部54的處理内容。 移動補償預測部54的直接向量生成部61在可變長产 解碼部52所輸出之大區塊型式/次區塊型式表示使用的= 直接模式時,針對每一個解碼對象大區塊生成空間直接模 式的空間直接向量、及時間直接模式的時間直接向量,、 將該空間直接向量或時間直接向量的任一方當作移動向二 而輸出至移動補償處理部62。 ° ,從儲存於移自向量記憶體51之已解碼大區塊的移動 …亦即,直接向量生成部61的空間直接向量生成部?1 量之中:讀出位於解碼對象大區塊周圍之已解喝大區塊^ 移動向#依據该移動向量而生成空間直接模式古 接向量。 工间罝 但’空間直接向量生成部71中之空間直接向 方法,叙後q m & a 置的生成 万興第3圖之空間直接向量生成部31中之空間直接向 322763 30 201143455 量的生成方法一樣,故將其詳細的說明予以省略。 直接向量生成部61的時間直接向量生成部7 2係從儲 存於移動向量記憶體51之已解碼大區塊的移動向量之 中,將在時間上處於解碼對象大區塊附近之已解碼晝面的 移動向量,並且為於該已解碼晝面中在空間上位於與解碼 對象大區塊相同位置之大區塊的移動向量予以讀出,從該 移動向量來生成時間直接模式的時間直接向量。 但,時間直接向量生成部72中之時間直接向量的生成 方法與第3圖之時間直接向量生成部32中之時間直接向量 的生成方法一樣,故將其詳細的說明子以省略。 直接向量生成部61的直接向量判定部73在空間直接 向量生成部71生成空間直接向量時,使用該空間直接向 量,來算出空間直接模式的評估值。 此外,直接向量判定部73在時間直接向量生成部72 生成時間直接向量時,使用該時間直接向量來算出時間直 接模式的評估值。 然後,直接向量判定部73進行該空間直接模式的評估 值、與時間直接模式的評估值之比較,而選擇該空間直接 向量或時間直接向量之中較適當的直接模式的直接向量並 將之輸出至移動補償處理部62。 以下,具體地說明直接向量判定部73的處理内容。 直接向量判定部73的移動補償部81在空間直接向量 生成部71生成空間直接向量MVLO, MVL1時,就如第11圖 所示,使用該空間直接向量MVL0來生成空間直接模式的前 31 322763 201143455 方預測影像fspatial ’且使用該空間直接向量MVL1來生成空 間直接模式的後方預測影像gspatiai。 此外,移動補償部81在當時間直接向量生成部72生 成為前方及後方的移動向量MV之時間直接向量時,就如第 11圖所示,使用該前方的移動向量MV來生成時間直接模 式的前方預測影像fteinp(3ral ’且使用後方的移動向量MV來生 成時間直接模式的後方預測影像 gtemporal ° 直接向量判定部73的類似度算出部82,在移動補償 部81生成空間直接模式的前方預測影像fspatial及後方預測 影像gspatial時,就與第4圖之類似度算出部42 —樣,算出 該前方預測影像f spatial 與後方預測影像gspatial的類似度並 以之作為空間直接模式的評估值SADspatUl。 此外,類似度算出部82在移動補償部81生成時間直 接模式的前方預測影像 f temporal 及後方預測影像gtemporal時’ 就與第4圖之類似度算出部42 —樣,算出該前方預測影像 f teap〇ral與後方預測影像gtemp(3ral的類似度並以之作為時間直 接模式的評估值SADtenp(>fal。 前方預測影像與後方預測影像間的差愈大,兩個影像 間的類似度就變低(表示兩個影像的差的絕對值和之評估 值SAD 1_大)時間性的相關就變低。反之,前方預測影像 ”後一方預測W像間的差愈小,兩個影像間的類似度就變高 (表示兩個;^像的差的絕對值和之評估值⑽變小),時間 性的相關就變高。 直接向置判定部73的直接向量選擇部83 ,在類似度 322763 32 201143455 算出部82算出空間直接模式的評估值SADspatial、及時間直 接模式的評估值SADtemporal時’就與第4圖之直接向量選擇 部43 —樣,藉由比較該評估值SADspatial與評估值 SADtemporal ’而比較空間直接模式中的前方預測影像;f spatial 與後方預測影像gspatial的類似度、與時間直接模式中的前 方預測影像f temporal及後方預測影像gtemp()ral的類似度。 當在空間直接模式中的前方預測影像fspatial與後方預 測影像gspatial的類似度、比時間直接模式中的前方預測影 像f tenoral及後方預測影像gtempcral的類似度高時(SADspatial $ SADtemporal),直接向量選擇部83係選擇空間直接向量生成 部71所生成之空間直接向量,並將該空間直接向量作為移 動向量而輸出至移動補償處理部62。 另一方面’當在時間直接模式中的前方預測影像 ftempc>rai及後方預測影像gtemp()ral的類似度、比空間直接模式 中的刖方預測影像f spatial與後方預測影像gspatiai的類似度 高時(SADspatiai>SADte,ral) ’則選擇時間直接向量生成部72 所生成之時間直接向量,並將該時間直接向量作為移動向 量而輸出至移動補償處理部62。 移動補償處理部62在可變長度解碼部52所輸出之大 區塊型式/次區塊型式表示使用献晝面賴式時,則使用 可變長度解碼部52所輸出的移動向量、及儲存於圖框記憶 體9中的-圖框的參考影像(可變長度解碼部52所輸出的 識別號瑪所表示的參考影像)來實施移動補償預測處理,而 生成預測影像。 322763 33 201143455 Λ欠巴另塊在可變長度解碼部52所輸出之大區塊型式 〇〇 式表不使用的是直接模式時,使用直接向量生成 ㈣向量、續存㈣框記㈣9中的一圖 ==?長度解碼部52所輸出的識別號碼所表 Γ)來實施移動補償預測處理,而生成預測影 技術又故:==:::補償預測處理為公知的 和來S評irsrtr算出部82算出差的絕對值 ==測影像間的差的平方和8 值。 使用與—貞似度算出部4M目同之評估 號所表示之差異影像^以相加 u _誤差訊 於從第1圖之影像編 生成表不解碼影像(相當 影像)之解碼影像訊號(步驟^:器7所輸出之局部解碼 迴路濾波器56在加法止、 對於該解碼影像訊號t所含的編碼衫像訊號時,就 補償編碼失真後之解$ 1失真進行補償’然後將 考影像予以儲存至圖』:::所表示的解碼影像當作參 體57,収㈣料影像輪出 322763 34 201143455 至外部(步驟ST27)。 如透過以上說明所能暸解的,根據本實施形態1而構 成為設有以下元件之形態:直接向量生成部22,依據位於 編碼對象大區塊周圍之已編碼大區塊的移動向量而生成空 間直接模式的空間直接向量,且依據在時間上處於編碼對 象大區塊附近之已編碼晝面的移動向量而生成時間直接模 式的時間直接向量;類似度算出部42,使用直接向量生成 部22所生成的空間直接向量而算出空間直接模式的評估 值,以及使用該時間直接向量而算出時間直接模式的評估 值;以及直接向量選擇部43,比較由類似度算出部42所 算出的空間直接模式的評估值、與時間直接模式的評估 值,而選擇空間直接向量或時間直接向量的任一方,且移 動補償處理部23係使用由直接向量選擇部43所選擇的直 接向量而實施移動補償預測處理,而生成預測影像,所以 能夠以大區塊為單位選擇最適合的直接模式,因而產生能 夠獲得可避免掉不必要的移動向量之編碼,而防止編碼量 增加的影像編碼裝置之效果。 此外,藉由根據本實施形態1而構成為設有以下元件 之形態:從位於解碼對象大區塊的周圍之已解碼的大區塊 的移動向量來生成空間直接模式的空間直接向量,以及從 在時間上處於解碼對象大區塊的附近之已解碼晝面的移動 向量來生成時間直接模式的時間直接向量;類似度算出部 82,係使用由直接向量生成部61所生成的空間直接向量而 算出空間直接模式的評估值,且使用該時間直接向量而算 35 322763 201143455 出時間直接模式的評估值;以及直接向量選擇部83,比_ 由類似度算出部82所算出的空間直接模式的評估值、與^ ,直接模式的評估值,而選擇空間直接向量或時間直^向 量的任一方’且移動補償處理部62係使用由直接向量選 部83所選擇的直接向量來實施移動補償預測處理,而 預測影像,所以可產生能夠獲得可將以大區塊為單位選 最適合的直接模式之方式的編碼資料 解 裝置的效果。 办像解碼 實施形態2. 在上述的實施形態1中,雖將類似度算出部4 疋成.异出空間直接模式的前方糊影像^㈤與後二 :像g_al的類似度並以之作為空間直接模式 SADsPatia1,^外算㈣間直接模式的前方預測 古值 與後方預測影像g—的類似度並以之作為時間= 的汗估值SAD—之形態,然而,亦可將之設^成算工 :於編碼對象(解碼對象)大區塊周圍之已編碼大區塊= ::::區塊)的移動向量的分散值 =間直接模式的評估值,且另外算出在相 乍為 象大區塊(已解碼大區塊) ;、、扁碼對 中,在空間上位於與編竭』:二:面(已解碼晝面) 向量的分散值·::===的移動 評估值之縣,W⑽增罐/挺式的 亦即’類似度算出部4U2係如第14圖⑷二 322763 36 201143455 出位於編碼對象(解碼對象)大區塊周圍之已編碼大區塊 (已解碼大區塊)的移動向量的分散值σ ( spat i a 1)來作為 空間直接模式的評估值(參照下面的式(7)),以此方式來取 代算出空間直接模式的前方預測影像fspatial與後方預測影 像gspatial的類似度並以之作為空間直接模式的評估值 SADspatial 之^作法0 另外,類似度算出部42, 82係如第14圖(b)所示,算 出在時間上處於編碼對象大區塊(已解碼大區塊)附近之已 編碼晝面(已解碼晝面)中,在空間上位於與編碼對象大區 塊(已解碼大區塊)相同位置之大區塊周圍之已編碼大區塊 (已解碼大區塊)的移動向量的分散值σ (temporal)來作為 時間直接模式的評估值(參照下面的式(7)),以此方式來取 代算出時間直接模式的前方預測影像f bporai與後方預測影 像gtemporal的類似度並以之作為時間直接模式的評估值 SADtemporal 之^作法。 (w)=去 Σ 一 兩-)2,# = 4 (7)The motion compensation prediction process is performed to generate a predicted image (step ST25). The processing contents of the motion compensation prediction unit 54 will be specifically described below. The direct vector generation unit 61 of the motion compensation prediction unit 54 generates a spatial direct space for each decoding target large block when the large block type/sub-block pattern output by the variable-length decoding unit 52 indicates the used direct mode. The spatial direct vector of the pattern and the temporal direct vector of the temporal direct mode are output to the motion compensation processing unit 62 as either of the spatial direct vector or the temporal direct vector. °, from the movement of the decoded large block stored in the vector memory 51, that is, the spatial direct vector generation unit of the direct vector generation unit 61? 1 Quantity: Read out the large block of decontamination located around the large block of the decoding object ^ Move to # Generate a spatial direct mode ancient vector according to the motion vector. However, the space in the 'spatial direct vector generation unit 71 is directly directed to the method, and the space in the spatial direct vector generation unit 31 of the fourth generation of the Wanxing diagram is generated directly to the 322763 30 201143455. The method is the same, so the detailed description will be omitted. The temporal direct vector generation unit 72 of the direct vector generation unit 61 is from the motion vector stored in the decoded large block of the motion vector memory 51, and is decoded in the vicinity of the decoding target large block in time. a motion vector, and for reading a motion vector of a large block spatially located at the same position as the decoding object large block in the decoded plane, a time direct vector of the temporal direct mode is generated from the motion vector. However, the method of generating the temporal direct vector in the temporal direct vector generating unit 72 is the same as the method of generating the temporal direct vector in the temporal direct vector generating unit 32 of Fig. 3, and therefore detailed description thereof will be omitted. When the spatial direct vector generation unit 71 generates the spatial direct vector, the direct vector determination unit 73 of the direct vector generation unit 61 calculates the evaluation value of the spatial direct mode using the spatial direct vector. Further, when the time direct vector generation unit 72 generates the time direct vector, the direct vector determination unit 73 calculates the evaluation value of the time direct mode using the time direct vector. Then, the direct vector determining unit 73 compares the evaluation value of the spatial direct mode with the evaluation value of the temporal direct mode, and selects a direct vector of the appropriate direct mode among the spatial direct vector or the temporal direct vector and outputs it. To the motion compensation processing unit 62. Hereinafter, the processing content of the direct vector determining unit 73 will be specifically described. When the spatial direct vector generation unit 71 generates the spatial direct vectors MVLO and MVL1, the spatial direct vector generation unit 71 generates the first 31 of the spatial direct mode using the spatial direct vector MVL0 as shown in FIG. 11 . The square predicts the image fspatial 'and uses the spatial direct vector MVL1 to generate the spatially direct mode rear prediction image gspatiai. Further, when the temporal direct vector generation unit 72 generates the temporal direct vector of the motion vector MV of the front and the rear, the motion compensation unit 81 generates the temporal direct mode using the forward motion vector MV as shown in FIG. The forward-predicted video fteinp (3ral' and the rear-predicted video using the rear motion vector MV to generate the temporal direct mode gtemporal ° The similarity calculating unit 82 of the direct vector determining unit 73 generates the forward-predicted video of the spatial direct mode in the motion compensating unit 81. In the fspatial and the backward prediction image gspatial, the degree of similarity between the forward predicted image f spatial and the backward predicted image gspatial is calculated and used as the evaluation value SADspatU1 of the spatial direct mode, similarly to the similarity calculating unit 42 of Fig. 4 . When the motion compensation unit 81 generates the forward predicted video f temporal and the backward predicted video gtemporal in the temporal direct mode, the similarity calculating unit 82 calculates the forward predicted video f teap〇 as in the similarity calculating unit 42 of the fourth figure. Ral and rear prediction image gtemp (3ral similarity and use it as time direct mode Valuation SADtenp(>fal. The larger the difference between the forward predicted image and the backward predicted image, the lower the similarity between the two images (indicating the absolute value of the difference between the two images and the evaluation value SAD 1_ is large) The temporal correlation is lower. Conversely, the smaller the difference between the W images predicted by the latter prediction image, the higher the similarity between the two images (representing the two; the absolute value of the difference between the image and the evaluation) When the value (10) is small, the temporal correlation is increased. The direct vector selection unit 83 of the direct orientation determination unit 73 calculates the evaluation value SADspatial of the spatial direct mode and the temporal direct mode in the similarity 322763 32 201143455 calculation unit 82. When evaluating the value SADtemporal, it is compared with the direct vector selection unit 43 of Fig. 4, and the forward predicted image in the spatial direct mode is compared by comparing the evaluation value SADspatial with the evaluation value SADtemporal '; f spatial and the rear predicted image gspatial Similarity, similarity to the forward predicted image f temporal and the backward predicted image gtemp() ral in the temporal direct mode. When the forward predicted image fspatial and after in the spatial direct mode When the similarity degree of the predicted image gspatial is higher than the similarity between the forward predicted image f tenoral and the backward predicted image gtempcral in the temporal direct mode (SADspatial $ SADtemporal), the direct vector selecting unit 83 selects the generated by the spatial direct vector generating unit 71. The spatial direct vector is output to the motion compensation processing unit 62 as a motion vector. On the other hand, 'the similarity between the forward predicted image ftempc> rai and the backward predicted image gtemp() ral in the temporal direct mode is higher than the similarity between the temporal predicted image f spatial and the rear predicted image gspatiai in the spatial direct mode. When (SADspatiai > SADte, ral) ', the time direct vector generated by the time direct vector generating unit 72 is selected, and the time direct vector is output as the motion vector to the motion compensation processing unit 62. The motion compensation processing unit 62 uses the motion vector output from the variable length decoding unit 52 when the large block type/secondary block pattern output from the variable length decoding unit 52 indicates that the use of the presentation format is used. The reference image of the frame in the frame memory 9 (the reference image indicated by the identification number output by the variable length decoding unit 52) performs motion compensation prediction processing to generate a predicted image. 322763 33 201143455 大 巴 另 在 在 在 在 在 在 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变 可变Fig. ==? The identification number output by the length decoding unit 52 is performed), and the motion compensation prediction process is performed to generate a prediction shadow technique. The ==::: compensation prediction process is a well-known sum S evaluation irsrtr calculation unit 82 Calculate the absolute value of the difference == The square of the difference between the measured images and the 8 value. Using the difference image indicated by the evaluation number of the similarity calculation unit 4M, the decoded image signal of the undecoded image (corresponding image) is generated by adding the image from the image of the first image by adding the u_error. ^: The local decoding loop filter 56 outputted by the device 7 compensates for the distortion of the decoded distortion image after the encoding of the encoded image signal contained in the decoded image signal t, and then compensates the image. The decoded image stored in the picture:::: is taken as the parameter 57, and the received (four) material image is rotated 322763 34 201143455 to the outside (step ST27). As can be understood from the above description, it is constructed according to the first embodiment. In the form of the following components: the direct vector generation unit 22 generates a spatial direct vector of the spatial direct mode according to the motion vector of the coded large block located around the large block of the coding object, and is based on the time of the coding object. A time direct vector of the temporal direct mode is generated by the motion vector of the coded face near the block; the similarity calculation unit 42 uses the space generated by the direct vector generation unit 22 The evaluation value of the spatial direct mode is calculated by the direct vector, and the evaluation value of the temporal direct mode is calculated using the direct vector of the time; and the direct vector selection unit 43 compares the evaluation value of the spatial direct mode calculated by the similarity calculation unit 42 and With respect to the evaluation value of the time direct mode, one of the spatial direct vector or the temporal direct vector is selected, and the motion compensation processing unit 23 performs motion compensation prediction processing using the direct vector selected by the direct vector selecting unit 43 to generate a prediction. Since the image is imaged, it is possible to select the most suitable direct mode in units of large blocks, thereby producing an image encoding device capable of obtaining an encoding which can avoid unnecessary moving vectors and preventing an increase in the amount of encoding. According to the first embodiment, the configuration is such that a spatial direct vector of the spatial direct mode is generated from the motion vector of the decoded large block located around the large block of the decoding target, and the decoding target is temporally decoded. When the motion vector of the decoded face near the large block is generated The temporal direct vector of the direct mode; the similarity calculating unit 82 calculates the evaluation value of the spatial direct mode using the spatial direct vector generated by the direct vector generating unit 61, and uses the time direct vector to calculate 35 322763 201143455 The evaluation value of the mode; and the direct vector selection unit 83 selects the spatial direct vector or the time straight vector by the evaluation value of the spatial direct mode calculated by the similarity calculation unit 82 and the evaluation value of the direct mode. The motion compensation processing unit 62 performs the motion compensation prediction process using the direct vector selected by the direct vector selection unit 83 to predict the video, so that it is possible to obtain the most suitable one that can be selected in units of large blocks. The effect of the coded data solution device in the direct mode mode. Embodiment 2 In the first embodiment described above, the similarity calculation unit 4 converts the similarity between the front paste image (5) and the second image: g_al in the direct space mode and uses it as a space. The direct mode SADsPatia1, ^ external calculation (four) direct mode front prediction ancient value and the rear prediction image g - similarity and used as the time = the sweat valuation SAD - the form, however, can also be set to Work: the dispersion value of the motion vector of the coded large block =::: block around the large block of the coding object (decoding object) = the evaluation value of the direct mode, and additionally calculate the image in the opposite direction Block (decoded large block);,, flat code alignment, spatially located and edited: 2: face (decoded face) The dispersion value of the vector ·::=== In the prefecture, the W(10) canning/stiffing type, that is, the similarity calculating unit 4U2 is as shown in Fig. 14 (4) two 322763 36 201143455. The encoded large block located around the large block of the encoding target (decoding target) (decoded large area) The dispersion value σ (spat ia 1) of the motion vector of the block is used as the evaluation value of the spatial direct mode Referring to the following formula (7)), in this way, the similarity between the forward predicted image fspatial and the backward predicted image gspatial of the spatial direct mode is calculated and used as the evaluation value of the spatial direct mode SADspatial. As shown in FIG. 14(b), the calculation units 42 and 82 calculate the spatially encoded side (decoded face) in the vicinity of the large block to be encoded (the decoded large block). The dispersion value σ (temporal) of the motion vector of the coded large block (decoded large block) around the large block at the same position as the coding target large block (decoded large block) is used as the temporal direct mode The evaluation value (refer to the following equation (7)) is used in place of the similarity between the forward predicted image f bporai and the backward predicted image gtemporal of the temporal direct mode and used as the evaluation value SADtemporal of the temporal direct mode. (w)=Go Σ One Two-)2,#= 4 (7)
^ ieR 其中,MVm,i表示周圍的移動向量m表示周圍的移 動向量之平均。 另外,m係表示spatial或temporal之記號。 直接向量選擇部43,83係進行移動向量的分散值σ (spatial)與移動向量的分散值σ (temporal)之比較,且在 移動向量的分散值σ (spat ial)比移動向量的分散值σ (temporal)大時,判斷為空間直接模式的移動向量(空間直 37 322763 201143455 接向量)的可靠性較低,而選擇時間直接模式的移動向量 (時間直接向量)。 另一方面,在當移動向量的分散值σ (temporal)比移 動向量的分散值σ (spatial)大時,則判斷為時間直接模式 的移動向量(時間直接向量)的可靠性較低,而選擇空間直 接模式的移動向量(空間直接向量)。 實施形態3. 在上述的實施形態1中雖設定成.當編碼模式為直接 模式時,係以大區塊為單位生成空間直接向量及時間直接 向量,並選擇空間直接向量或時間直接向量,然後使用選 擇的直接向量來生成預測影像之形態,然而,亦可設定成 例如:只有在片段標頭(slice header)中所含的直接模式 切換旗標 “direct_spatial_mv_pred_flag” 表示為「無意 義」(例如“0”)時,才實施與上述實施形態1 一樣之預測 影像生成處理,而在該直接模式切換旗標表示「有意義」(例 如“Γ或“2”)時,則選擇該直接模式切換旗標所指示的 直接模式之直接向量(例如當旗標=1時,選擇空間直接模 式的空間直接向量,當旗標=2時,選擇時間直接模式的時 間直接向量)之形態。 以下,具體地說明本實施形態3的處理内容。 此處,為了說明之方便,假設:若片段標頭中所含的 直接模式切換旗標為“0”,就實施與上述實施形態1 一樣 之預測影像生成處理(以大區塊為單位,選擇空間直接向量 或時間直接向量)。 38 322763 201143455 若直接模式切換旗標為“Γ,則對於該片段中的所有 大區塊都選擇空間直接模式的空間直接向量。 若直接模式切換旗標為“2”,則對於該片段中的所有 大區塊都選擇時間直接模式的時間直接向量。 本實施形態3,在直接模式切換旗標為“Γ或“2” 之情況,係以片段為一個單位,切換為選擇空間直接向量 或是選擇時間直接向量,但不限於此,亦可設定成例如: 以晝面(picture)為一個單位或以序列(sequence)為一個 單位,切換為選擇空間直接向量或是選擇時間直接向量之 形態。 而且,本實施形態3雖以直接模式切換旗標表示三種 狀態(“0”,“Γ,“2”)者進行說明,但不限於此,亦可 設定成例如:在直接模式切換旗標只表示0N(有意義)或 OFF(無意義)時,另外再輸入額外的旗標(例如組態 (profile)資訊、強制設定旗標(constraint_set_flag) 等)之形態。 亦即,在直接模式切換旗標為OFF之情況,實施與上 述實施形態1一樣之預測影像生成處理。 另一方面,當直接模式切換旗標為0N時,則例如:當 額外旗標為“0” ,則選擇空間直接模式的空間直接向量, 當額外旗標為“Γ ,則選擇時間直接模式的時間直接向 量。 第15圖係顯示本發明的實施形態3中的影像編碼裝置 之構成圖,圖中,與第1圖相同之符號係表示相同或相當 39 322763 201143455 的部份’故將其說明予以省略。 移動補償預測部11在片段標頭中所含的直接模式切 換旗"^為G k’實施與第1圖之移動補償預測部2-樣 之處理。 ’ 移動補償預測部n在片段標頭中所含的直接模式切 換旗標為θ‘Ί” 4,實施以下處理:生成空間直接模式的空 間直接向里,並使用該空間直接向量來實施移動補償預測 處理而生成預測影像。 另外,移動補償預測部11在片段標頭中所含的直接模 式切換旗標為“2”時,實施以下處理:生成時間直接模式 的時間直接向1,並使用該時間直接向量來實施移動補償 預測處理而生成預測影像。 移動補偵預測部11與第1圖之移動補償預測部2 一 樣,由移動向量搜尋部21、直接向量生成部22及移動補 償處理部23所構成(參照第2圖)。 可變長度編碼部12實施以下處理:對於壓縮部5所輸 出之壓縮資料、移動補償預測部U所輸出之編碼模式資= (大區塊型式/次區塊型式、移動向量、參考影像的識別編 =)、及直接模式切換旗標進行熵編碼,而生成表示該編蝎 結果之位元串流(bit stream)(編碼資料),然後將該位元 串机予以輸iB可變長度編碼部I〗係構成可變長度編碼手 第16圖係顯示構戍移動補償預u之直 成部22之構成圖,圖中,… 王 與第3圖相同之符號係表示相同 322763 40 201143455 或相當的部份,故將其說明予以省略。 直接向量判定部34在片段標頭中所含的直接模式切 換旗標為“0”時,與第3圖之直接向量判定部33 —樣地 實施以下處理:使用空間直接向量生成部31所生成的空間 直接向量來算出空間直接模式的評估值,以及使用時間直 接向量生成部32所生成的時間直接向量來算出時間直接 模式的評估值,然後比較該空間直接模式的評估值、與時 間直接模式的評估值之比較,而選擇該空間直接向量或時 間直接向量的任一方。 直接向量判定部34在片段標頭中所含的直接模式切 換旗標為“Γ時,實施以下處理:選擇空間直接向量生成 部31所生成的空間直接向量,並將該空間直接向量作為移 動向量而輸出至移動補償處理部23。 另外,直接向量判定部34在片段標頭中所含的直接模 式切換旗標為“2”時,實施以下處理:選擇時間直接向量 生成部32所生成的時間直接向量,並將該時間直接向量作 為移動向量而輸出至移動補償處理部23。 第17圖係顯示構成直接向量生成部22之直接向量判 定部34之構成圖。 第17圖中,移動補償部44在片段標頭中所含的直接 模式切換旗標為“〇”時,與第4圖之移動補償部41 一樣 地實施以下處理:使用空間直接向量生成部31所生成的空 間直接向量來生成空間直接模式的前方預測影像及後方預 測影像,以及使用時間直接向量生成部32所生成的時間直 41 322763 201143455 接向量來生成時間直接模式的前方預測影像及後方預測影 像。 另外,移動補償部44實施以下處理:在直接模式切換 旗標為“Γ時,將空間直接向量生成部31所生成的空間 直接向量輸出至類似度算出部45,在該直接模式切換旗標 為“2”時,將時間直接向量生成部32所生成的時間直接 向量輸出至類似度算出部45。 類似度算出部45在直接模式切換旗標為“0”時,與 第4圖之類似度算出部42 —樣地實施以下處理:算出空間 直接模式的前方預測影像與後方預測影像的類似度而以之 作為空間直接模式的評估值,以及算出時間直接模式的前 方預測影像與後方預測影像的類似度而以之作為時間直接 模式的評估值。 另外,類似度算出部45實施以下處理:在直接模式切 換旗標為“Γ時,將移動補償部44所輸出的空間直接向 量輸出至直接向量選擇部46,在直接模式切換旗標為“2” 時,將移動補償部44所輸出的時間直接向量輸出至直接向 量選擇部46。 移動補償部44及類似度算出部45係構成評估值算出 手段。 直接向量選擇部46在直接模式切換旗標為“0”時, 與第4圖之直接向量選擇部43 —樣地實施以下處理:比較 類似度算出部45所算出的空間直接模式中的前方預測影 像與後方預測影像的類似度、以及時間直接模式中的前方 42 322763 201143455 預測影像與後方預測影像的類似度,而選擇* 或時間直接向量之中前方預㈣像與❹接向量 度較高之直接模式的直接向量。 〜像的類似 另外,直接向量選擇部46實施以下處理: 切換旗標為“Γ時,選擇類似度算出部4 接模式 直接向量並將之輸出至移動補償處理部23 ^的空間 換旗標為“2”冑’選擇類似度算出部奶 2式切 接向量並將m移動補償處理部23。=時間直 46係構成直接向量選擇手段。 置選擇部 第18圖係顯示本發明實施形態3中的影 構成圖,中,與第5圖相同之符_表示% 1之 部份,故將其說明予以省略。 门或相當的 1受长度解碼部58實施以下處理:使從 像編碼裝置所輸出之位元串流(編碼資料)輸::影 元串流進行轉碼轉碼出壓縮資料、編顯、二位 塊型式/次區塊型式、移動向量、參考景 "。區 及直接模式切換旗標,然後將該壓縮資料輸出5至^^)、 解碼部53,將該編碼模式資減直接模式切制標 移動補償預測部59。可變長度解碼部58係構成 解碼手段。 移動補償賴部59實施以下處理:從儲存於圖框記憶 體57之-圖框以上的參考影像之十,將可變長度解碼部 58所輸出之識別號碼所表示的參考影像予以讀出,然後^ 可變長度解碼部58所輸出之大區塊型式/次區塊型式表示 322763 43 201143455 面間模式時,使用可變長度解馬部料輪出之 預測影^。上述參考影像來實施移動補償預測處理而生成 移動補償制部59在可變長度解碼部58所輸出之 區塊型式/次區塊型式表示使用的是直接模式,且可變長戶 解碼部58所輸出之直接模式切換旗標為‘τ時,又 圖^影像編碼裝置中的移動補償預測部11 -樣地實施以 直接向量及時間直接向量,然後選擇該 向量或時間直接向量的任-方,再使用所選擇的 °里、及識職碼表示的參考影像來實施移動補償 測處理而生成預測影像。 另外’移動補償_部59在可變長度解 =區塊型式/次區塊型式表示使用的是直接模式 交長度解碼部58所輸出之直接模式切換旗標| “Γ時, 實,以下處理·生成空間直接向量’然後使用該空間直 ==號碼表示的參考影像來實施移動 而生成預測影像。 另外’移動補償預測部59在可變長度解碼部 出之大區塊型式/次區塊型式表示使用的是直接模式,且: 變長度解碼部58所輸出之直接模式切換旗標υ °, 實,以孩理.生成時間直接向量,然後使用該時間直 向量及識職碼表㈣參考f彡絲實祕動 而生成預測影像。 貝頂利處理 且,移動補償預測部59與第6圖之移動補償預測部 322763 201143455 54 —樣,由直接向量生成部61及移動補償處理部62所構 成(參照第6圖)。 第19圖係顯示構成移動補償預測部59之直接向量生 成部61之構成圖,圖中,與第7圖相同之符號係表示相同 或相當的部份,故將其說明予以省略。 直接向量判定部74在可變長度解碼部58所輸出之直 接模式切換旗標為“0”時,與第7圖之直接向量判定部 73 —樣地實施以下處理:使用空間直接向量生成部71所 生成的空間直接向量,來算出空間直接模式的評估值,以 及使用時間直接向量生成部72所生成的時間直接向量,來 算出時間直接模式的評估值,然後比較該空間直接模式的 評估值、與時間直接模式的評估值,而選擇該空間直接向 量或時間直接向量的任一方。 直接向量判定部74在直接模式切換旗標為“Γ時, 實施以下處理:選擇空間直接向量生成部71所生成的空間 直接向量,並將該空間直接向量作為移動向量而輸出至移 動補償處理部62。 另外,直接向量判定部74在直接模式切換旗標為 “2”時,實施以下處理:選擇時間直接向量生成部72所 生成的時間直接向量^並將該時間直接向量作為移動向量 而輸出至移動補償處理部62。 第20圖係顯示構成直接向量生成部61之直接向量判 定部74之構成圖。 第20圖中,移動補償部84在可變長度解碼部58所輸 45 322763 201143455 出之直接模式切換旗標為“0”時,與第8圖之移動補償部 81 —樣,實施以下處理:使用空間直接向量生成部71所 生成的空間直接向量,來生成空間直接模式的前方預測影 像及後方預測影像,以及使用時間直接向量生成部72所生 成的時間直接向量,來生成時間直接模式的前方預測影像 及後方預測影像。 另外,移動補償部84實施以下處理:在直接模式切換 旗標為“Γ時,將空間直接向量生成部71所生成的空間 直接向量輸出至類似度算出部85,在該直接模式切換旗標 為“2”時,將時間直接向量生成部72所生成的時間直接 向量輸出至類似度算出部85。 類似度算出部85在直接模式切換旗標為“0”時,與 第8圖之類似度算出部82 —樣地實施以下處理:算出空間 直接模式的前方預測影像與後方預測影像的類似度且以之 作為空間直接模式的評估值,以及算出時間直接模式的前 方預測影像與後方預測影像的類似度且以之作為時間直接 模式的評估值。 另外,類似度算出部85實施以下處理:在直接模式切 換旗標為“Γ時,將移動補償部84所輸出之空間直接向 量輸出至直接向量選擇部86,在直接模式切換旗標為“2” 時,將移動補償部84所輸出之時間直接向量輸出至直接向 量選擇部86。 移動補償部84及類似度算出部85係構成評估值算出 手段。 46 322763 201143455 直接向量選擇部86在直接模式切換旗標為“〇,,時, 與第8圖之直接向量選擇部83 —樣地實施以下處理 類似度算出部85所算出之空間直接模式中的前方預測¥ 像與後方賴影像的驗度、以及相直減式中的 預測,像與後方預測影像的類似度之比較,而選擇空間直 =向量或時間直接向量之中,前方預測影像與後方預測影 像的類似度較面之直接模式的直接向量。 另外,直接向量選擇部86實施以下處理:在 切換旗標為“Γ,時,選擇類似度算出 、二 ::?為J’,:’選擇類似度算_所輪出 接向置並將之輸出至移動補償處理部62。 86係構成直接向量選擇手段。 接白里擇。p 移』=碼裝置的各構成要素,亦即 部5、局部解碼部6、=::編, 編碼部12為分別由專用的硬體(:巧器8及可變長度 體積體電路、或者單s # # .女裴有CPU之半導 置若由電腦所構成:丨片可=等)所構成,但影像編瑪裝 補償預测部11、減法 以下形態··將記述有移動 局部解竭部6、加法°。編碼模式判定部4、麗縮部5、 12的處理内容之程式儲=路遽波器8及可變長度編碼部 的哪來_存於該的記億體,再由該電腦 ΰ中雖假设影像解喝裝置的各構成要素,亦即 322763 47 201143455 ^變長度解碼部58、預測誤差解碼部53、移 9、加法器55及迴路濾波器56為分 預測部 如:安裝有CPU之半導體積體電路、或者單=^更體(例 所構成’但影像解碼裝置若由電腦ϋ電腦等) 下形態:將記述有可變長度解碼部二^^ ieR where MVm,i represents the surrounding motion vector m representing the average of the surrounding motion vectors. In addition, m is a symbol indicating spatial or temporal. The direct vector selecting sections 43, 83 perform a comparison between the dispersion value σ (spatial) of the motion vector and the dispersion value σ (temporal) of the motion vector, and the dispersion value σ (spatial) of the motion vector is larger than the dispersion value σ of the motion vector. When (temporal) is large, it is judged that the motion vector of the spatial direct mode (space straight 37 322763 201143455) is less reliable, and the motion vector of the direct direct mode (time direct vector) is selected. On the other hand, when the dispersion value σ (temporal) of the motion vector is larger than the dispersion value σ (spatial) of the motion vector, it is determined that the reliability of the motion vector (time direct vector) of the temporal direct mode is low, and the selection is low. The motion vector (spatial direct vector) of the spatial direct mode. Embodiment 3. In the first embodiment described above, when the coding mode is the direct mode, a spatial direct vector and a temporal direct vector are generated in units of large blocks, and a spatial direct vector or a temporal direct vector is selected, and then The selected direct vector is used to generate the predicted image. However, it can be set to, for example, only the direct mode switching flag "direct_spatial_mv_pred_flag" contained in the slice header is expressed as "meaningless" (for example, " When 0"), the same prediction image generation processing as in the first embodiment is performed, and when the direct mode switching flag indicates "meaningful" (for example, "Γ or "2"), the direct mode switching flag is selected. The direct vector of the indicated direct mode (for example, when the flag = 1, the spatial direct vector of the spatial direct mode is selected, and when the flag = 2, the time direct vector of the direct direct mode is selected). The processing content of the third embodiment. Here, for convenience of explanation, it is assumed that if the fragment header contains The mode switching flag is "0", and the same prediction image generation process as that of the above-described first embodiment is performed (the spatial direct vector or the temporal direct vector is selected in units of large blocks). 38 322763 201143455 If the direct mode switching flag is “Γ, then a spatial direct vector of spatial direct mode is selected for all large blocks in the segment. If the direct mode switch flag is "2", the time direct vector of the time direct mode is selected for all large blocks in the segment. In the third embodiment, in the case where the direct mode switching flag is "Γ or "2", the segment is set as one unit, and the direct vector is selected as the selection space or the time direct vector is selected, but it is not limited thereto, and may be set to For example, the picture is taken as a unit or the sequence is a unit, and the mode is switched to the selection of the direct vector of the space or the form of the direct vector of the time. Furthermore, in the third embodiment, the direct mode switching flag is used. The three states ("0", "Γ, "2") are described, but are not limited thereto, and may be set, for example, when the direct mode switching flag indicates only 0N (meaningful) or OFF (meaningless), Then enter the form of additional flags (such as profile information, mandatory flags (constraint_set_flag), etc.). That is, in the case where the direct mode switching flag is OFF, the same predicted image generation processing as in the first embodiment described above is performed. On the other hand, when the direct mode switching flag is 0N, for example, when the extra flag is "0", the spatial direct vector of the spatial direct mode is selected, and when the extra flag is "Γ, the time direct mode is selected. Fig. 15 is a view showing a configuration of a video encoding apparatus according to a third embodiment of the present invention. In the figure, the same reference numerals as in the first embodiment denote the same or equivalent parts of 39 322763 201143455. The motion compensation prediction unit 11 performs a process similar to the motion compensation prediction unit 2 of Fig. 1 in the direct mode switching flag "^ is Gk' included in the slice header. 'The motion compensation prediction unit n is The direct mode switching flag included in the fragment header is θ'Ί" 4, and the following processing is performed: the spatial direct mode space is generated directly inward, and the spatial direct vector is used to implement the motion compensation prediction process to generate a predicted image. Further, when the direct mode switching flag included in the slice header is "2", the motion compensation prediction unit 11 performs a process of generating the time direct mode time directly to 1, and implementing the motion compensation using the time direct vector. The prediction process is generated by the prediction process. The motion compensation prediction unit 11 is composed of the motion vector search unit 21, the direct vector generation unit 22, and the motion compensation processing unit 23 as in the motion compensation prediction unit 2 of Fig. 1 (see Fig. 2). The variable length coding unit 12 performs the following processing: the compressed data output from the compression unit 5 and the coding mode output by the motion compensation prediction unit U = (large block type/sub-block type, motion vector, reference picture recognition) Encoding =), and the direct mode switching flag is entropy encoded, and a bit stream (encoded data) indicating the result of the compilation is generated, and then the bit string machine is input to the iB variable length coding unit. I constituting a variable length coding hand Fig. 16 is a configuration diagram showing a straight portion 22 of a configuration of a motion compensation pre-u. In the figure, the same symbols as in the third diagram represent the same 322763 40 201143455 or equivalent. Partially, the description will be omitted. When the direct mode switching flag included in the slice header is "0", the direct vector determining unit 34 performs the following processing in the same manner as the direct vector determining unit 33 of the third drawing: the generated by the spatial direct vector generating unit 31 The spatial direct vector calculates the evaluation value of the spatial direct mode, and uses the temporal direct vector generated by the temporal direct vector generating unit 32 to calculate the evaluation value of the temporal direct mode, and then compares the evaluation value of the spatial direct mode with the temporal direct mode. To compare the evaluation values, select either the direct vector of the space or the direct vector of the time. When the direct mode switching flag included in the slice header is "Γ", the direct vector determining unit 34 performs the following processing: the spatial direct vector generated by the spatial direct vector generating unit 31 is selected, and the spatial direct vector is used as the motion vector. In addition, when the direct mode switching flag included in the slice header is "2", the direct vector determining unit 34 performs a process of selecting the time generated by the temporal direct vector generating unit 32. The direct vector is output to the motion compensation processing unit 23 as a motion vector. Fig. 17 is a view showing a configuration of the direct vector determining unit 34 constituting the direct vector generating unit 22. In Fig. 17, the motion compensating unit When the direct mode switching flag included in the slice header is "〇", the following processing is performed in the same manner as the motion compensation unit 41 of Fig. 4: the spatial direct vector generated by the spatial direct vector generating unit 31 is used to generate The forward predicted image and the backward predicted image of the spatial direct mode and the time generated by the use time direct vector generating unit 32 are straight 41 3227 63 201143455 The vector is used to generate the forward predicted image and the backward predicted image in the temporal direct mode. Further, the motion compensating unit 44 performs a process of generating the space generated by the spatial direct vector generating unit 31 when the direct mode switching flag is "Γ". The direct vector is output to the similarity calculation unit 45, and when the direct mode switching flag is "2", the temporal direct vector generated by the temporal direct vector generation unit 32 is output to the similarity calculation unit 45. When the direct mode switching flag is "0", the similarity calculating unit 45 performs the following processing similarly to the similarity calculating unit 42 of the fourth drawing: calculating the similarity between the forward predicted image and the backward predicted image in the spatial direct mode. It is used as an evaluation value of the spatial direct mode, and the similarity between the forward predicted image and the backward predicted image of the temporal direct mode is calculated as the evaluation value of the temporal direct mode. Further, the similarity calculation unit 45 performs a process of outputting the spatial direct vector output from the motion compensation unit 44 to the direct vector selection unit 46 when the direct mode switching flag is "Γ", and switching the flag to "2" in the direct mode. The time direct vector output from the motion compensation unit 44 is output to the direct vector selection unit 46. The motion compensation unit 44 and the similarity calculation unit 45 constitute an evaluation value calculation means. The direct vector selection unit 46 switches the flag in the direct mode. When it is "0", the following process is performed in the same manner as the direct vector selection unit 43 of Fig. 4: the similarity and time of the forward predicted image and the backward predicted image in the spatial direct mode calculated by the similarity calculation unit 45 are compared. In the direct mode, the front 42 322763 201143455 predicts the similarity between the image and the rear predicted image, and selects the direct vector of the direct mode in front of the * or time direct vector and the direct vector with higher splicing vector degree. The direct vector selection unit 46 performs the following processing: When the switching flag is "Γ", the similarity calculation unit 4 is selected to be connected to the mode direct vector. The outputted to the movement compensation processing unit 23 of the space for a flag ^ selected similarity calculating unit connected milk cut vector 2 and m of formula movement compensation processing unit 23 is "2" helmet ". = time straight 46 series constitute a direct vector selection means. Fig. 18 is a view showing a configuration of a shadow in the third embodiment of the present invention. The same symbol as in Fig. 5 indicates a part of %1, and therefore the description thereof will be omitted. The gate or equivalent 1-bit length decoding unit 58 performs a process of causing the bit stream (encoded data) output from the image encoding device to be transcoded and transcoded to output compressed data, edited, and Bit block type/sub-block type, motion vector, reference scene ". The area and direct mode switching flags are then outputted to the decoding unit 53 by the compressed data, and the coding mode subtraction direct mode is switched to the standard motion compensation prediction unit 59. The variable length decoding unit 58 constitutes a decoding means. The movement compensation processing unit 59 performs a process of reading out the reference image indicated by the identification number output from the variable length decoding unit 58 from the reference image stored above the frame of the frame memory 57, and then ^ The large block type/sub-block type output by the variable length decoding unit 58 indicates 322763 43 201143455 In the inter-face mode, the predicted image of the variable length solution is used. The motion compensation prediction unit 59 performs the motion compensation prediction processing, and the block type/sub-block pattern output from the variable length decoding unit 58 is used in the direct mode, and the variable-length decoding unit 58 is used. When the direct mode switching flag of the output is 'τ, the motion compensation prediction unit in the video encoding apparatus performs the direct vector and the time direct vector, and then selects any vector of the vector or time direct vector. The predicted image is generated by performing motion compensation measurement processing using the selected reference image and the reference image indicated by the identification code. Further, the 'movement compensation unit' 59 uses the direct mode switching flag output by the direct mode intersection length decoding unit 58 in the variable length solution=block type/subblock type pattern| The spatial direct vector is generated. Then, the reference image represented by the space direct == number is used to perform the motion to generate the predicted image. Further, the motion compensation prediction unit 59 displays the large block type/sub-block type in the variable length decoding unit. The direct mode is used, and: the direct mode switching flag outputted by the variable length decoding unit 58 υ °, real, to the child. Generate a time direct vector, and then use the time straight vector and the job code table (4) reference f彡The predicted image is generated by the motion of the wire. The motion compensation prediction unit 59 is composed of the direct vector generation unit 61 and the motion compensation processing unit 62 as in the motion compensation prediction unit 322763 201143455 54 of the sixth figure ( Fig. 6 is a view showing a configuration of a direct vector generation unit 61 constituting the motion compensation prediction unit 59. In the figure, the same reference numerals as in Fig. 7 indicate the same or equivalent. The direct vector determination unit 74 performs the same as the direct vector determination unit 73 of Fig. 7 when the direct mode switching flag outputted by the variable length decoding unit 58 is "0". The following processing: calculating the evaluation value of the spatial direct mode using the spatial direct vector generated by the spatial direct vector generating unit 71, and calculating the evaluation value of the temporal direct mode using the temporal direct vector generated by the temporal direct vector generating unit 72, Then, the evaluation value of the spatial direct mode and the evaluation value of the temporal direct mode are compared, and either one of the spatial direct vector or the temporal direct vector is selected. The direct vector determining unit 74 implements the following when the direct mode switching flag is “Γ” Processing: The spatial direct vector generated by the spatial direct vector generation unit 71 is selected, and the spatial direct vector is output as a motion vector to the motion compensation processing unit 62. Further, when the direct mode switching flag is "2", the direct vector determining unit 74 performs a process of selecting the time direct vector generated by the temporal direct vector generating unit 72 and outputting the time direct vector as a motion vector to the mobile. The compensation processing unit 62. Fig. 20 is a view showing the configuration of the direct vector determining unit 74 constituting the direct vector generating unit 61. In the 20th figure, when the direct mode switching flag of the variable length decoding unit 58 is 45 322763 201143455, the motion compensation unit 84 performs the following processing as in the motion compensation unit 81 of Fig. 8: The forward direct video and the backward predicted video of the spatial direct mode are generated using the spatial direct vector generated by the spatial direct vector generating unit 71, and the temporal direct vector generated by the temporal direct vector generating unit 72 is used to generate the front of the temporal direct mode. Predicted images and rear predicted images. Further, the motion compensating unit 84 performs a process of outputting the spatial direct vector generated by the spatial direct vector generating unit 71 to the similarity calculating unit 85 when the direct mode switching flag is "Γ", and the direct mode switching flag is In the case of "2", the time direct vector generated by the time direct vector generating unit 72 is output to the similarity calculating unit 85. The similarity calculating unit 85 similarly to the eighth figure when the direct mode switching flag is "0". The calculation unit 82 performs the following processing: calculating the similarity between the forward predicted image and the backward predicted image in the spatial direct mode as the evaluation value of the spatial direct mode, and calculating the forward predicted image and the backward predicted image in the temporal direct mode. The similarity degree is used as the evaluation value of the time direct mode. The similarity calculation unit 85 performs a process of outputting the spatial direct vector output from the motion compensation unit 84 to the direct vector when the direct mode switching flag is "Γ". The selection unit 86 outputs the direct vector of the time outputted by the motion compensation unit 84 to the straight when the direct mode switching flag is "2" The amount of selection unit 86 is connected. The motion compensation unit 84 and the similarity calculation unit 85 constitute an evaluation value calculation means. 46 322763 201143455 The direct vector selection unit 86 performs the following processing in the spatial direct mode calculated by the similarity calculation unit 85 in the same manner as the direct vector selection unit 83 of the eighth mode when the direct mode switching flag is "〇". The forward prediction ¥ image and the rear Lai image, and the prediction in the direct subtraction, like the comparison with the backward prediction image, and select the space straight = vector or time direct vector, the front prediction image and the rear In addition, the direct vector selection unit 86 performs the following processing: when the switching flag is "Γ, the similarity is calculated, and the second::? is J',:' The similarity calculation_selection is turned on and output to the motion compensation processing unit 62. The 86 series constitutes a direct vector selection means. Pick up the choice. Each of the components of the p-transformation code device, that is, the unit 5, the local decoding unit 6, and the code-making unit 12 are respectively dedicated hardware (the device 8 and the variable-length volume circuit, or Single s # # . The 裴 裴 CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU CPU Exhaustion 6, addition °. The code mode determination unit 4 and the program contents of the refinement units 5 and 12 are stored in the channel unit, and the variable length coding unit is stored in the unit. The components of the image defrosting device, that is, 322763 47 201143455 ^ variable length decoding unit 58, prediction error decoding unit 53, shift 9, adder 55, and loop filter 56 are sub-prediction units such as a semiconductor product with a CPU installed. Body circuit, or single = ^ more body (for example, the video decoding device is a computer, computer, etc.). The following form: The variable length decoding unit will be described.
53、移動補償預測部59、加法器55及迴、^差解石馬部 ^容之程式儲存於該電腦的記憶體,再^56的處 來執仃儲存於該記憶體的程式。 °Λ的CPU 接著針對動作進行說明。 明第陳影像編碼農置的處理内容。 及可變長度料部12之外㈣ 冑__部11 因,弟1圖之影像編碼穿番4 同因此只針對移動補償預測部11及可變長产置相 的處理内容進行說明。 隻長度編碼部12 補傷預卿11,在輸人有表示輪人影像 ::::框=塊(或次區塊)為單位將該動態 移動補償預測部u將動態影像訊於 次區塊)單位後,則從儲存於圖框記憶體W之塊(或 的移動補償預測料參考影像之中選擇-圖框之^上 行移動補償預測處理,藉此針對各個色成分執 塊)的移動向量進而生編简象大區以或次區 移動補償預測部11生成編简“區塊U次區塊) 322763 48 201143455 的移動向量且生成預測影像後’則將該預測影 法器3,以及將該預測影像之生成中用到的移動至減 區塊型式/次區塊型式(包含例如,表示在該大區、大 塊)使用的編碣模式係為晝面間模式或直接模式、(或次區 之資訊)以及參考影像的識別號碼予以輪中、X的哪一者 定部4。 出至編碼模式判 以下,具體地說明移動補償預測部丨丨的處理扣 不過,此處為了說明之方便,假設其為以大品各 位而生成移動向量進而生成預測影像者。 區塊為單 移動補彳員預測部11的移動向量搜尋部21 示編碼模式為晝面間模式之資訊(從例如°=收到表 ^的是晝面間模式之資訊)時,就以畫面間=表示 移動補償預測❹的直接向量t貝處理部23。 示編碼模式為直接模式之資訊時,。/在接收到表 大區塊,生成空間直接模式 母一個蝙碼對象 ,時間直接向量,並將該空f:=、及時間直接 I的任-方當作移動向量而。I或時間直接向 亦即’直接向量生成部2 動補償處理部23。 ^上述實施形態】—樣,接向量生成部31 編碼大區塊的移動向量之中=移動向量記憶體!之已 之已編碼大區塊的移動向量以;扁碼對象大區塊周圍 成空間直接模式的空間直接向量從該移動向量來生 另外,直接向量生成部 成❿的時間直接向量生成部32 322763 49 201143455 與上述實施形態1一樣,從儲存於移動向量記憶體1之已 編碼大區塊的移動向量之中’將在時間上處於編碼對象大 區塊附近之已編碼晝面的移動向量,且為該已編碼晝面中 之在空間上位於與編碼對象大區塊相同位置之大區塊的移 動向量予以讀出,從該移動向量來生成時間直接模式的時 間直接向量。 直接向量生成部22的直接向量判定部34在片段標頭 中所含的直接模式切換旗標為“0”之情況,就在空間直接 向量生成部31生成空間直接向量時,使用該空間直接向量 來算出空間直接模式的評估值,以及在時間直接向量生成 部32生成時間直接向量時,使用該時間直接向量來算出時 間直接模式的評估值。 然後,直接向量判定部34與第3圖之直接向量判定部 33 —樣,比較該空間直接模式的評估值與時間直接模式的 評估值,而選擇該空間直接向量或時間直接向量,再將選 出的直接向量作為移動向量而輸出至移動補償處理部23。 直接向量判定部34在直接模式切換旗標為“Γ時, 選擇空間直接向量生成部31所生成的空間直接向量,並將 該空間直接向量作為移動向量而輸出至移動補償處理部 23。 另外,直接向量判定部34在直接模式切換旗標為 “2”時,選擇時間直接向量生成部32所生成的時間直接 向量’並將該時間直接向量作為移動向量而輸出至移動補 償處理部23。 50 322763 201143455 關於直接模式切換旗標之設定,可考慮例如:若影像 整體為移動拍攝(panning)之類的輸入影像,則將直接模式 切換旗標設定為表示選擇的是時間直接模式之“2”,若為 只是在晝面内的動向改變之輸入影像,則將直接模式切換 旗標設定為表示選擇的是空間直接模式之“1” 。 以下,具體地說明直接向量判定部34的處理内容。 直接向量判定部34的移動補償部44在直接模式切換 旗標為“0”時,與第4圖之移動補償部41 一樣,使用空 間直接向量生成部31所生成的空間直接向量來生成空間 直接模式的前方預測影像及後方預測影像,以及使用時間 直接向量生成部32所生成的時間直接向量來生成時間直 接模式的前方預測影像及後方預測影像。 另外,移動補償部44在直接模式切換旗標為“Γ 時,將空間直接向量生成部31所生成的空間直接向量輸出 至類似度算出部45,在直接模式切換旗標為“2”時,將 由時間直接向量生成部32所生成的時間直接向量輸出至 類似度算出部45。 直接向量判定部34的類似度算出部45在直接模式切 換旗標為“0”時,與第4圖之類似度算出部42 —樣,算 出空間直接模式的前方預測影像與後方預測影像的類似度 而以之作為空間直接模式的評估值,以及算出時間直接模 式的前方預測影像與後方預測影像的類似度而以之作為時 間直接模式的評估值。 另外,類似度算出部45在直接模式切換旗標為“Γ 51 322763 201143455 時’將移動補償部44所輸出之空間直接向量輸出至直接向 量選擇部46,在直接模式切換旗標為“2”時,將移動補 償部44所輪出之時間直接向量輸出至直接向量選擇部46。 直接向量判定部34的直接向量選擇部46在直接模式 切換旗標為“0”時,與第4圖之直接向量選擇部43 一樣% 進行類似度算出部45所算出的空間直接模式中的前方預 測影像與後方預測影像的類似度、以及時間直接模式中的 前方預測影像與後方預測影像的類似度之比較,而選擇办 間直接向量或時間直接向量之中,前方預測影像與後方預 測影像的類似度較高之直接模式的直接向量,並將該直接 向量作為移動向量而輸出至移動補償處理部23。 “ ”另外,直接向量選擇部46在直接模式切換旗標為 “1”時’接收從類似度算出部45所輸出之空間直接向量 的輸入,並將該空間直接向量作為移動向量而輸出至移= 補償處理部23。 按向重达擇部P且恢镆式切換旗標為 之情況,將類似度算出部45所輸出之時間直接向量 並將該時間直接向量作為移動向量而輸出至 移動補償處理部23。 料,:==尸接收到從壓縮部5傳來的壓縮資 區塊型式/次區塊測部U傳來的編碼模式資訊(大 就對於該壓縮資料、編f動向量、參考影像的識別編號), 行熵編碼,而生成表示接模式切換旗標進 遣蝙碼結果之位元串流,然後將該 322763 52 201143455 位元串流予以輪出。 接著,說明坌〗0 ^ ^ ^罘18圖之影像解碼裝置的處理内容。 及可變長产解^像解喝裝置中,除了移動補償_部59 的處理内容進補侦預測部59及可變長度解崎部58 可變長度解瑪部58係於當接收第15圖之 置所輸出之位元由、&从认 '、扁碼裝 人時,對於該位元串流進行網解 型式、移動^ _式資訊〔大區塊型式/:欠區塊 ° 1 (編碼模式為晝面間模式的情況)、參與 =的識別編號〕及直接模式切換旗標,然後將該壓縮資料 3出至預測誤差解碼部53,將該編碼模式資訊及直接 切換旗標輸&轉動補償預卿59。 “ 移動補償預測部59接收到從可變長度解碼部58傳來 的參考影像的識別編號’就從儲存於圖框記憶體57之一圖 框以上的參考影像之中,將該識別號碼所表示的參考影傻 予以讀出。 此外’移動補償預測部59接收到從可變長度解碼部 來的大區塊型式/次區塊型式,就參考該大區塊型式/ 人區塊型式,判別第15圖之影像編碼裝置係使用晝面間模 式來作為編碼模式或使用直接模式來作為編碼模式。 查移動補償預測部59在第15圖之影像編碼裝置係使用 息面間模式來作為編碼模式時,使用可變長度解碼部58所 輪出之移動向量及上述參考影像來實施移動補償預測處理 322763 53 201143455 而生成預測影像。 另一方面’移動補償預測部59在第15圖之与 裝置係使用直接模絲作為編碼模式之情況,若;县, 解碼:P 58所輸出之直接模式切換旗標$ “〇”,則:第G 圖之影像編碼裝置中的移動補償預測部11 一樣地生成处 門直接向量及時間直接向量,然後選擇該空間直 接向量的任—方,再使用所選擇的直接向量及識^ :馬表不的參考影像來實施移動補償_處理而生成預測 衫像。 抄動補償預測部59在第15圖之影像編碼裝置係使用 ^接模式來作為編碼模式時,若從可變長度料部58所輪 =之直接模式切換旗標為“Γ,則生成空間直接向量,= 後使用該空間直接向量及識別號碼表示的參考影像來實施 移動補償預測處理而生成預測影像。 ,另外’移動補償預測部59在第15圖之影像編碼裝置 係使用直接模式來作為編碼模式時,若從可變長度解竭部 58所輪出之直接模式切換旗標為“2” ,則生成時間直接 白1 然後使用該時間直接向量及識別號碼表示的參考影 像來實施移動補償預測處理而生成預測影像。 以下’具體地說明移動補償預測部59的處理内容。 移動補償預測部59的直接向量生成部61在可變長度 解馬4 58所輸出之大區塊型式/次區塊型式表示使用的是 直接模式時’與上述實施形態1一樣地針對每一個解碼對 象大區塊生成空間直接模式的空間直接向量、及時間直接 54 322763 201143455 f式的時間直接向量,並將該空間直接向量或時間直接向 量的任一方當作移動向量而輸出至移動補償處理部62。 亦即’直接向量生成部61的空間直接向量生成部71 係與上述實施形態丨一樣地從儲存於移動向量記憶體51之 已解碼大區塊的移動向量之中,將位⑨料對象大區塊周 ^之已解竭大區塊的移動向量予以讀出,而依據該移動向 量來生成空間直接模式的空間直接向量。 直接向量生成部61的時間直接向量生成部72係與上 述實施形態1-樣地從儲存於移動向量記憶體51之已解碼 大區塊的移動向量之中’將在時間上處於解碼對象大區塊 附近之已解碼晝面的移動向量,且為該已解碼晝面中之在 ^間仏於與解碼對象大區塊相同位置之大區塊的移動向 里予以讀出,而依據鮮動向量來生成時間直接模式的時 間直接向量》 直接向量生成部61的直接向量判定部74在可變長度 解碼部58所輸出之直接模式切換旗標為“〇”時,在空間 f接向量生成部71生成空間直接向量,使用該空間直接向 1來异出空間直接模式的評估值,以及在時間直接向量 生成部72生成時間直接向量,使用該時間直接向量,來算 出時間直接模式的評估值。 然後,直接向量判定部74與第7圖之直接向量判定部 73樣地進行該空間直接模式的評估值與時間直接模式的 砰估值之比較,而選擇該空間直接向量或時間直接向量, 並將所選擇的直接向量作為移動向量而輸出至移動補償處 322763 55 201143455 理部62。 直接向量判定部74在直接模式切換旗標為“Γ之情 況,選擇空間直接向量生成部71所生成的空間直接向量, 並將該空間直接向量作為移動向量而輸出至移動補償處理 部62。 另外,直接向量判定部74在直接模式切換旗標為 “2”之情況,選擇時間直接向量生成部72所生成的時間 直接向量,並將該時間直接向量作為移動向量而輸出至移 動補償處理部62。 以下,具體地說明直接向量判定部74的處理内容。 直接向量判定部74的移動補償部84在可變長度解碼 部58所輸出之直接模式切換旗標為“0”時,與第8圖之 移動補償部81 —樣地使用空間直接向量生成部71所生成 的空間直接向量來生成空間直接模式的前方預測影像及後 方預測影像,以及使用時間直接向量生成部72所生成的時 間直接向量來生成時間直接模式的前方預測影像及後方預 測影像。 另外,移動補償部84在直接模式切換旗標為“Γ 時,將空間直接向量生成部71所生成的空間直接向量輸出 至類似度算出部85,在直接模式切換旗標為“2”時,將 時間直接向量生成部72所生成的時間直接向量輸出至類 似度算出部85。 類似度算出部85在直接模式切換旗標為“0”時,與 第8圖之類似度算出部82—樣地算出空間直接模式的前方 56 322763 201143455 預測影像與後方預測影像的類似度且以之作為空間直接模 式的評估值,以及算出時間直接模式的前方預測影像與後 方預測影像的類似度且以之作為時間直接模式的評估值。 另外,類似度算出部85在直接模式切換旗標為“Γ 時,將移動補償部84所輸出之空間直接向量輸出至直接向 量選擇部86,在直接模式切換旗標為“2”時,將移動補 償部84所輸出之時間直接向量輸出至直接向量選擇部86。 直接向量選擇部86在直接模式切換旗標為“0”時, 與第8圖之直接向量選擇部83 —樣地,進行類似度算出部 8 5所算出的空間直接模式中的前方預測影像與後方預測影 像的類似度、以及時間直接模式中的前方預測影像與後方 預測影像的類似度之比較,而選擇空間直接向量或時間直 接向量之中,前方預測影像與後方預測影像的類似度較高 之直接模式的直接向量,並將該直接向量作為移動向量而 輸出至移動補償處理部62。 直接向量選擇部86在直接模式切換旗標為“Γ時, 選擇類似度算出部85所輸出之空間直接向量,並將該空間 直接向量作為移動向量而輸出至移動補償處理部62,在直 接模式切換旗標為“2”時,選擇類似度算出部85所輸出 之時間直接向量,並將該時間直接向量作為移動向量而輸 出至移動補償處理部62。 如透過以上說明所能暸解的,藉由根據本實施形態3 而構成為以下形態:只有在片段標頭中所含的直接模式切 換旗標 “direct_spatial_mv_pred_flag” 表示「無意義」 57 322763 201143455 (例如辦’才實施與上述實施形態^樣之預測影像 生成處理,在该直接模式切換旗標表 “1”々“〇,,、々法.〇 有忍義」(例如 1或2 )之情況,則選擇該直接模式切換旗 的直接模式的直接向量(例如' 曰不 U j如褀钛―1之情況,選擇空間吉 接模式的"直接向量,旗標=2之情況,選擇時間直接模 式的時間直接向量),故可產生_依料單 行 的演算量或可使用的記憶體量而選擇適當模式之效果進订 而且,可吸收影像編碼裝置及影像 的變動,而以適切地處理量進行編碼及解碼。、 理 〔產業上的利用可能性〕 … 本發明適肖於销由避免 防止編碼量增加的黨# 茺妁移動向夏之編碼而 另外,像編碼裝置及料編碼方法。 本發明適用於有將如上述 成的編石馬資料予以解碼的 碼裝置所生 方法 【圖式簡單說明】 圖係顯示本發明實施形態 方法。 〜像解碼裝置及影像解碼 第 構成圖 1中之影像編碼裝置之 ㈣=圖係顯示本發明實施形態11之與傻㈣h 移動__部2之構成圖。 T之⑹像編碼裝置的 部22第之3構^示構成移動補償㈣部2之直接向量生成 圖係顯示構成直接向量生成99 ^ 部33之構成_。 王成。卩22之直接向量判定 322763 58 201143455 第5圖係顯示本發明實施形態1中之影像解碼裝置之 構成圖。 第6圖係顯示本發明實施形態1中之影像解碼裝置的 移動補償預測部54之構成圖。 第7圖係顯示構成移動補償預測部54之直接向量生成 部61之構成圖。 第8圖係顯示構成直接向量生成部61之直接向量判定 部73之構成圖。 第9圖係顯示以時間直接模式來生成移動向量之方法 之示意圖。 第10圖係顯示以空間直接模式來生成移動向量之方 法之示意圖。 第11圖係顯示以前方預測影像與後方預測影像的類 似度為依據之評估值的算出例之說明圖。 第12圖係顯示本發明實施形態1中之影像編碼裝置的 處理内容之流程圖。 第13圖係顯示本發明實施形態1中之影像解碼裝置的 處理内容之流程圖。 第14圖(a)及(b)係顯示使用移動向量的分散值之評 估式之說明圖。 第15圖係顯示本發明實施形態3中之影像編碼裝置之 構成圖。 第16圖係顯示構成移動補償預測部11之直接向量生 成部22之構成圖。 59 322763 201143455 第Π圖係顯示構成直接向量生成部22之直接向量判 疋部34之構成圖。 第18圖係顯示本發明實施形態3中之影像解碼裝置之 構成圖。 第19圖係顯示構成移動補償預測部59之直接向量生 成部61之構成圖。 第20圖係顯示構成直接向量生成部61之直接向量判 疋部74之構成圖。 【主要元件符號說明】 移動向量記憶體 :、11 移動補償預測部 1 減法器(量子化手段) 編碼模式判定部(量子化手段) 1 壓縮部(量子化手段) 1 局部解碼部 '加法器 8 9 10、12 21 22 23 31 32 迴路濾波器 圖框記憶體 可變長度編碼部(可變長度編碼手段) 移動向量搜尋部 直接向量生成部 移動補償處理部(預測影像生成手段) 空間直接向量生成部(直接向量生成手段) 時間直接向量生成部(直接向量生成手段) 322763 60 201143455 33、34 41 ' 44 42、 45 43、 46 51 52、58 53 54、59 55 56 57 61 62 71 72 73、74 81 ' 84 82、 85 83、 86 直接向量判定部 移動補償部(評估值算出手段) 類似度算出部(評估值算出手段) 直接向量選擇部(直接向量選擇手段) 移動向量記憶體 可變長度解碼部(可變長度解碼手段) 預測誤差解碼部(逆量子化手段) 移動補償預測部 加法器(影像相加手段) 迴路濾波器(影像相加手段) 圖框記憶體 直接向量生成部 移動補償處理部(預測影像生成手段) 空間直接向量生成部(直接向量生成手段) 時間直接向量生成部(直接向量生成手段) 直接向量判定部 移動補償部(評估值算出手段) 類似度算出部(評估值算出手段) 直接向量選擇部(直接向量選擇手段) 61 32276353. The motion compensation prediction unit 59, the adder 55, and the program of the back and the solution are stored in the memory of the computer, and then the program stored in the memory is executed. The CPU of °Λ will then explain the action. Mingdi Chen image coding agricultural processing content. In addition to the variable length material portion 12, the processing of the motion compensation prediction unit 11 and the variable length production phase will be described. Only the length coding unit 12 compensates for the pre-clear 11 and the dynamic motion compensation prediction unit u transmits the motion picture to the secondary block in units of the wheel person image::::box=block (or sub-block). After the unit, the motion vector is selected from the block of the memory W of the frame (or the motion compensation prediction reference image of the selection - the uplink motion compensation prediction process of the frame, thereby blocking the block for each color component) Further, the pseudo-image area or the sub-area motion compensation prediction unit 11 generates a motion vector of the block "block U sub-block" 322763 48 201143455 and generates a predicted image, then the prediction picturer 3, and The edit mode used in the generation of the predicted image generation to the reduced block type/sub-block type (including, for example, indicating the large area and the large block) is an inter-panel mode or a direct mode, (or The information of the sub-area) and the identification number of the reference image are used to determine which part of the round and X is fixed. 4. The encoding mode is determined below, and the processing of the motion compensation prediction unit 具体地说 is specifically described. Convenient, suppose it is a big product The motion vector search unit 21 of the single-moving supplement predictor unit 11 displays the information that the coding mode is the inter-plane mode (from, for example, ° = received table ^ is 昼). In the case of the information of the inter-surface mode, the direct vector t-being processing unit 23 indicating the motion compensation prediction 以 is displayed between the pictures. When the coding mode is the information of the direct mode, / when the large block is received, the generation space is directly generated. The pattern parent has a bat code object, a time direct vector, and the empty f:=, and the time-directed I-side is regarded as a motion vector. I or time directly to the direct vector generation unit 2 dynamic compensation processing unit 23. In the above embodiment, the vector generation unit 31 encodes the motion vector of the coded large block of the motion vector memory of the large block; the surrounding area of the large block of the flat code object; The spatial direct vector in the spatial direct mode is generated from the motion vector. The direct vector generation unit 32 322763 49 201143455 is the same as the first embodiment described above, and is stored in the moving direction. Among the motion vectors of the coded large block of the quantity memory 1, 'the motion vector of the coded face that will be temporally near the large block of the coding object, and is spatially located in the coded facet The motion vector of the large block at the same position of the coding target large block is read, and the temporal direct vector of the temporal direct mode is generated from the motion vector. The direct vector decision unit 34 of the direct vector generation unit 22 includes the slice header. When the direct mode switching flag is "0", when the spatial direct vector generating unit 31 generates the spatial direct vector, the spatial direct vector is used to calculate the evaluation value of the spatial direct mode, and is generated by the temporal direct vector generating unit 32. When the time is a direct vector, the time direct vector is used to calculate the evaluation value of the time direct mode. Then, the direct vector determining unit 34 compares the evaluation value of the spatial direct mode with the evaluation value of the temporal direct mode, and selects the spatial direct vector or the temporal direct vector, and selects the same as the direct vector determining unit 33 of the third figure. The direct vector is output to the motion compensation processing unit 23 as a motion vector. When the direct mode switching flag is "Γ", the direct vector determining unit 34 selects the spatial direct vector generated by the spatial direct vector generating unit 31, and outputs the spatial direct vector to the motion compensation processing unit 23 as a motion vector. When the direct mode switching flag is "2", the direct vector determining unit 34 selects the time direct vector generated by the temporal direct vector generating unit 32 and outputs the time direct vector as a motion vector to the motion compensation processing unit 23. 322763 201143455 For the setting of the direct mode switching flag, for example, if the image as a whole is an input image such as panning, the direct mode switching flag is set to indicate that the time direct mode is selected as "2". If the input image is changed only in the movement direction of the face, the direct mode switching flag is set to indicate "1" which is the spatial direct mode selected. Hereinafter, the processing content of the direct vector determining unit 34 will be specifically described. When the direct mode switching flag is "0", the motion compensating unit 44 of the direct vector determining unit 34 and FIG. 4 Similarly to the motion compensating unit 41, the forward direct video and the backward predicted video in the spatial direct mode are generated using the spatial direct vector generated by the spatial direct vector generating unit 31, and the time direct vector generated by the temporal direct vector generating unit 32 is used to generate the time. When the direct mode switching flag is "Γ", the motion compensation unit 44 outputs the spatial direct vector generated by the spatial direct vector generating unit 31 to the similarity calculating unit 45, in the direct mode switching flag. When the direct mode switching flag is "2", the time direct vector generated by the temporal direct vector generating unit 32 is output to the similarity calculating unit 45. When the direct mode switching flag is "0", the similarity calculating unit 45 of the direct vector determining unit 34 calculates the similarity between the forward predicted image and the backward predicted image in the spatial direct mode, similarly to the similarity calculating unit 42 of Fig. 4 . The degree is used as the evaluation value of the spatial direct mode, and the similarity between the forward predicted image and the backward predicted image of the temporal direct mode is calculated as the evaluation value of the temporal direct mode. Further, the similarity calculation unit 45 outputs the spatial direct vector output from the motion compensation unit 44 to the direct vector selection unit 46 when the direct mode switching flag is "Γ 51 322763 201143455", and the direct mode switching flag is "2". The time vector directly rotated by the motion compensating unit 44 is output to the direct vector selecting unit 46. The direct vector selecting unit 46 of the direct vector determining unit 34 when the direct mode switching flag is "0", and the fourth figure The direct vector selection unit 43 compares the similarity between the forward predicted image and the backward predicted image in the spatial direct mode calculated by the similarity calculation unit 45, and the similarity between the forward predicted image and the backward predicted image in the temporal direct mode. A direct vector of a direct mode in which the forward predicted image and the backward predicted image are highly similar is selected, and the direct vector is output as a motion vector to the motion compensation processing unit 23. In addition, the direct vector selection unit 46 receives the received from the similarity calculation unit 45 when the direct mode switching flag is "1". The spatial direct vector input is output, and the spatial direct vector is output as a motion vector to the shift=compensation processing unit 23. The similarity calculation unit is executed by pressing the key selection unit P and restoring the flag. The time direct vector outputted by 45 is output to the motion compensation processing unit 23 as a motion vector. Material::== The corpse receives the compression block type/sub-block measurement transmitted from the compression unit 5. The coding mode information transmitted by the part U (largely for the compressed data, the encoded motion vector, the identification number of the reference image), the entropy coding, and the generation of the bit stream indicating the result of the mode switching flag entering the bat code Then, the 322763 52 201143455 bit stream is rotated. Next, the processing content of the video decoding device of the 坌 0 ^ ^ ^ 罘 18 picture is explained. The processing content complementation prediction unit 59 and the variable length resolution unit 58 of the compensation unit 59 are the variable length deciphering unit 58 that is output when receiving the image of Fig. 15 and & When the code is loaded, for this bit stream Line network solution type, mobile ^ _ type information [large block type /: under block ° 1 (encoding mode is the case of the inter-plane mode), participation = identification number] and direct mode switching flag, and then The compressed data 3 is output to the prediction error decoding unit 53, and the encoding mode information and the direct switching flag are input & rotation compensation pre-clear 59. "The motion compensation prediction unit 59 receives the reference image transmitted from the variable length decoding unit 58. The identification number ' is read from the reference image stored above the frame of one of the frame memories 57, and the reference shadow indicated by the identification number is read. Further, the 'movement compensation prediction unit 59 receives the large block type/sub-block type from the variable length decoding unit, and refers to the large block type/person block type to determine the image coding device used in FIG. The inter-plane mode is used as the coding mode or the direct mode is used as the coding mode. When the video coding apparatus of FIG. 15 uses the inter-plane mode as the coding mode, the motion compensation prediction unit 59 performs the motion compensation prediction process 322763 by using the motion vector and the reference video rotated by the variable length decoding unit 58. 53 201143455 Generates a predicted image. On the other hand, in the case where the motion compensation prediction unit 59 uses the direct mode wire as the coding mode in the device of Fig. 15, if the county, decoding: P 58 outputs the direct mode switching flag $ "〇", then: The motion compensation prediction unit 11 in the image coding apparatus of the Gth map generates the gate direct vector and the temporal direct vector in the same manner, and then selects any of the spatial direct vectors, and then uses the selected direct vector and the identification: The motion compensation_processing is performed on the non-reference image to generate a predicted shirt image. When the video encoding apparatus of Fig. 15 uses the splicing mode as the encoding mode, the smear compensation prediction unit 59 directly generates a space directly from the direct mode switching flag of the variable length material unit 58. The vector, = then uses the spatial direct vector and the reference image indicated by the identification number to perform motion compensation prediction processing to generate a predicted image. Further, the motion compensation prediction unit 59 uses the direct mode as the coding in the video coding apparatus of FIG. In the mode, if the direct mode switching flag rotated from the variable length depletion unit 58 is "2", the generation time is directly white 1 and then the reference vector represented by the direct vector and the identification number is used to implement the motion compensation prediction. The processing is performed to generate a predicted image. The details of the processing by the motion compensation prediction unit 59 will be described below. The direct vector generation unit 61 of the motion compensation prediction unit 59 outputs the large block type/sub-block in the variable length solution 4 58 . When the pattern indicates that the direct mode is used, 'the space direct mode is generated for each decoding target large block as in the above-described first embodiment. The direct vector, and the time direct vector of the direct type 54 322763 201143455 f, and one of the spatial direct vector or the temporal direct vector is output as a motion vector to the motion compensation processing unit 62. That is, the direct vector generation unit 61 The spatial direct vector generation unit 71 is the same as the above-described embodiment, from the motion vector stored in the decoded large block of the motion vector memory 51, and the large block of the bit object has been depleted. The motion vector of the block is read, and a spatial direct vector of the spatial direct mode is generated based on the motion vector. The temporal direct vector generation unit 72 of the direct vector generation unit 61 is stored in the motion vector from the above-described embodiment 1 Among the motion vectors of the decoded large block of the memory 51, 'the motion vector of the decoded face that will be temporally near the large block of the decoding object, and is the same in the decoded facet. The movement of the large block at the same position of the large block of the decoding object is read inward, and the time direct vector of the time direct mode is generated according to the fresh motion vector. When the direct mode switching flag output by the variable length decoding unit 58 is "〇", the direct vector determining unit 74 of the quantity generating unit 61 generates a spatial direct vector in the space f vector generation unit 71, and directly uses the space to the space. The estimated value of the direct space out mode is generated, and the time direct vector generating unit 72 generates a time direct vector, and uses the time direct vector to calculate an evaluation value of the time direct mode. Then, the direct vector determining unit 74 and the seventh figure The direct vector determining unit 73 compares the evaluation value of the spatial direct mode with the 砰 estimate of the temporal direct mode, selects the spatial direct vector or the temporal direct vector, and outputs the selected direct vector as a motion vector to Motion Compensation Department 322763 55 201143455 Department 62. The direct vector determining unit 74 selects the spatial direct vector generated by the spatial direct vector generating unit 71 when the direct mode switching flag is "Γ", and outputs the spatial direct vector as a motion vector to the motion compensation processing unit 62. When the direct mode switching flag is "2", the direct vector determining unit 74 selects the time direct vector generated by the temporal direct vector generating unit 72, and outputs the time direct vector as a motion vector to the motion compensation processing unit 62. Hereinafter, the processing content of the direct vector determining unit 74 will be specifically described. When the direct mode switching flag output by the variable length decoding unit 58 is "0", the motion compensation unit 84 of the direct vector determining unit 74 is the same as Fig. 8 The motion compensation unit 81 similarly generates the forward prediction video and the backward prediction video in the spatial direct mode using the spatial direct vector generated by the spatial direct vector generation unit 71, and uses the temporal direct vector generated by the temporal direct vector generation unit 72. A forward prediction image and a backward prediction image in the time direct mode are generated. Further, the motion compensation unit 8 4, when the direct mode switching flag is "", the spatial direct vector generated by the spatial direct vector generating unit 71 is output to the similarity calculating unit 85, and when the direct mode switching flag is "2", the temporal direct vector is generated. The time direct vector generated by the unit 72 is output to the similarity degree calculation unit 85. When the direct mode switching flag is "0", the similarity calculation unit 85 calculates the similarity between the forward image 562763 201143455 predicted image and the backward predicted image in the spatial direct mode, similarly to the similarity calculating unit 82 of Fig. 8 It is used as an evaluation value of the spatial direct mode, and the similarity between the forward predicted image and the backward predicted image of the temporal direct mode is calculated and used as the evaluation value of the temporal direct mode. Further, when the direct mode switching flag is "", the similarity calculation unit 85 outputs the spatial direct vector output from the motion compensation unit 84 to the direct vector selection unit 86, and when the direct mode switching flag is "2", The time direct vector output from the motion compensation unit 84 is output to the direct vector selection unit 86. The direct vector selection unit 86 performs the same as the direct vector selection unit 83 of Fig. 8 when the direct mode switching flag is "0". The similarity between the forward predicted image and the backward predicted image in the spatial direct mode calculated by the similarity calculation unit 85, and the similarity between the forward predicted image and the backward predicted image in the temporal direct mode, and the spatial direct vector or Among the temporal direct vectors, the direct vector of the direct mode in which the forward predicted image and the backward predicted image are highly similar, and the direct vector is output as a motion vector to the motion compensation processing section 62. The direct vector selecting section 86 is in the direct mode When the switching flag is "Γ", the spatial direct vector output by the similarity calculating unit 85 is selected, and the spatial direct vector is used as The motion vector is output to the motion compensation processing unit 62, and when the direct mode switching flag is "2", the time direct vector output by the similarity calculating unit 85 is selected, and the time direct vector is output as a motion vector to the motion compensation. Processing unit 62. As can be understood from the above description, according to the third embodiment, the direct mode switching flag "direct_spatial_mv_pred_flag" included in the slice header indicates "meaningless" 57 322763 201143455 (for example, 'Implementation of the predicted image generation process of the above-described embodiment, in the case where the direct mode switching flag table "1" 々 "〇,, 々法.〇有忍义" (for example, 1 or 2), Select the direct vector of the direct mode of the direct mode switching flag (for example, ' 曰 not U j such as 褀 titanium - 1 case, select the space ji joint mode " direct vector, flag = 2, select the time direct mode Time direct vector), so it can generate the effect of selecting the appropriate mode according to the calculation amount of the single line or the amount of memory that can be used, and can absorb the variation of the image coding device and the image, and encode with appropriate processing amount. And decoding., [Industrial use possibility] ... The present invention is suitable for the sale of the party to prevent the increase in the amount of coding. The present invention is applicable to a method of generating a code device for decoding a stone-like material as described above. [Simplified description of the drawings] The figure shows a method of an embodiment of the present invention. And the video decoding device (4) in the image encoding device of FIG. 1 is a block diagram showing the configuration of the silly (four) h moving__ portion 2 in the eleventh embodiment of the present invention. The direct vector generation map showing the motion compensation (fourth) section 2 shows the composition of the direct vector generation 99^ section 33. Wang Cheng. The direct vector determination of 卩22 322763 58 201143455 Fig. 5 shows the first embodiment of the present invention. Fig. 6 is a block diagram showing a configuration of a motion compensation prediction unit 54 of the video decoding device according to the first embodiment of the present invention. Fig. 7 is a diagram showing a direct vector generation unit 61 constituting the motion compensation prediction unit 54. Fig. 8 is a view showing a configuration of a direct vector determining unit 73 constituting the direct vector generating unit 61. Fig. 9 is a view showing a method of generating a motion vector in a time direct mode. Fig. 10 is a schematic diagram showing a method of generating a motion vector in a spatial direct mode. Fig. 11 is an explanatory diagram showing an example of calculation of an evaluation value based on the similarity between a forward predicted image and a backward predicted image. The figure shows a flow chart of the processing contents of the video encoding apparatus according to the first embodiment of the present invention. Fig. 13 is a flow chart showing the processing contents of the video decoding apparatus according to the first embodiment of the present invention. Fig. 14(a) and ( b) is an explanatory diagram showing an evaluation formula using the dispersion value of the motion vector. Fig. 15 is a view showing the configuration of the video encoding apparatus in the third embodiment of the present invention. Fig. 16 is a view showing the configuration of the direct vector generation unit 22 constituting the motion compensation prediction unit 11. 59 322763 201143455 The first diagram shows the configuration of the direct vector decision unit 34 constituting the direct vector generation unit 22. Figure 18 is a block diagram showing the configuration of a video decoding device in accordance with a third embodiment of the present invention. Fig. 19 is a view showing the configuration of the direct vector generation unit 61 constituting the motion compensation prediction unit 59. Fig. 20 is a view showing the configuration of the direct vector decision unit 74 constituting the direct vector generating unit 61. [Description of main component symbols] Motion vector memory: 11, Motion compensation prediction unit 1 Subtractor (quantization means) Encoding mode determination unit (quantization means) 1 Compression unit (quantization means) 1 Local decoding section 'Adder 8 9 10, 12 21 22 23 31 32 Loop filter frame memory variable length coding unit (variable length coding means) Motion vector search unit direct vector generation unit motion compensation processing unit (predictive video generation means) Spatial direct vector generation Part (direct vector generation means) Time direct vector generation unit (direct vector generation means) 322763 60 201143455 33, 34 41 ' 44 42, 45 43, 46 51 52, 58 53 54, 59 55 56 57 61 62 71 72 73, 74 81 ' 84 82, 85 83, 86 Direct vector judgment unit motion compensation unit (evaluation value calculation means) Similarity calculation unit (evaluation value calculation means) Direct vector selection unit (direct vector selection means) Moving vector memory variable length Decoding unit (variable length decoding means) prediction error decoding unit (inverse quantization means) motion compensation prediction unit adder (image phase Addition means) Loop filter (image addition means) Frame memory direct vector generation unit motion compensation processing unit (predictive video generation means) Spatial direct vector generation unit (direct vector generation means) Time direct vector generation unit (direct vector generation Means) Direct vector determination unit motion compensation unit (evaluation value calculation means) Similarity calculation unit (evaluation value calculation means) Direct vector selection unit (direct vector selection means) 61 322763
Claims (1)
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP2010029080 | 2010-02-12 | ||
| PCT/JP2011/000419 WO2011099242A1 (en) | 2010-02-12 | 2011-01-26 | Image encoding device, image decoding device, image encoding method, and image decoding method |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| TW201143455A true TW201143455A (en) | 2011-12-01 |
Family
ID=44367531
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| TW100104246A TW201143455A (en) | 2010-02-12 | 2011-02-09 | Image encoding device, image decoding device, image encoding method, image decoding method |
Country Status (3)
| Country | Link |
|---|---|
| JP (1) | JP5442039B2 (en) |
| TW (1) | TW201143455A (en) |
| WO (1) | WO2011099242A1 (en) |
Families Citing this family (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN106713931B (en) | 2010-09-30 | 2019-09-03 | 三菱电机株式会社 | Moving picture encoding device and method thereof, moving picture decoding device and method thereof |
| JP2014183544A (en) * | 2013-03-21 | 2014-09-29 | Fujitsu Ltd | Image processing unit and image processing method |
Family Cites Families (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP3977716B2 (en) * | 2002-09-20 | 2007-09-19 | 株式会社東芝 | Video encoding / decoding method and apparatus |
| JP4429996B2 (en) * | 2005-09-30 | 2010-03-10 | 富士通株式会社 | Moving picture coding program, moving picture coding method, and moving picture coding apparatus |
| JP2007221202A (en) * | 2006-02-14 | 2007-08-30 | Victor Co Of Japan Ltd | Moving picture encoder and moving picture encoding program |
| JP4527677B2 (en) * | 2006-03-15 | 2010-08-18 | 富士通株式会社 | Moving picture coding method, moving picture coding apparatus, moving picture coding program |
| JP2008283490A (en) * | 2007-05-10 | 2008-11-20 | Ntt Docomo Inc | Moving picture encoding apparatus, method and program, and moving picture decoding apparatus, method and program |
| JP2010028221A (en) * | 2008-07-15 | 2010-02-04 | Sony Corp | Motion vector detecting device, motion vector detecting method, image encoding device, and program |
| TWI405469B (en) * | 2009-02-20 | 2013-08-11 | Sony Corp | Image processing apparatus and method |
-
2011
- 2011-01-26 JP JP2011553735A patent/JP5442039B2/en not_active Expired - Fee Related
- 2011-01-26 WO PCT/JP2011/000419 patent/WO2011099242A1/en not_active Ceased
- 2011-02-09 TW TW100104246A patent/TW201143455A/en unknown
Also Published As
| Publication number | Publication date |
|---|---|
| JP5442039B2 (en) | 2014-03-12 |
| WO2011099242A1 (en) | 2011-08-18 |
| JPWO2011099242A1 (en) | 2013-06-13 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US12348754B2 (en) | Image encoding/decoding method and device for performing prediction, and method for transmitting bitstream involving weighted prediction and bi-directional optical flow | |
| Oh et al. | H. 264-based depth map sequence coding using motion information of corresponding texture video | |
| US8542739B2 (en) | Method of estimating disparity vector using camera parameters, apparatus for encoding and decoding multi-view picture using the disparity vector estimation method, and computer-readable recording medium storing a program for executing the method | |
| US8098731B2 (en) | Intraprediction method and apparatus using video symmetry and video encoding and decoding method and apparatus | |
| KR101365575B1 (en) | Method and apparatus for encoding and decoding based on inter prediction | |
| US9118929B2 (en) | Method for performing hybrid multihypothesis prediction during video coding of a coding unit, and associated apparatus | |
| JP5061179B2 (en) | Illumination change compensation motion prediction encoding and decoding method and apparatus | |
| KR101653118B1 (en) | Method for processing one or more videos of a 3d-scene | |
| US12316849B2 (en) | Image encoding/decoding method and device for performing prof, and method for transmitting bitstream | |
| US20070104276A1 (en) | Method and apparatus for encoding multiview video | |
| KR20110009141A (en) | Method and apparatus for template matching prediction in video encoding and decoding | |
| WO2008054176A1 (en) | Method and apparatus for video predictive encoding and method and apparatus for video predictive decoding | |
| US12244859B2 (en) | Method and device for deriving inter-view motion merging candidate | |
| TW201204054A (en) | Techniques for motion estimation | |
| US20130011072A1 (en) | Image processing apparatus and image processing method | |
| CN101248669B (en) | Apparatus and method for encoding and decoding multi-view video | |
| CN101243692B (en) | Method and device for encoding multi-view video | |
| CN105103546A (en) | Method and device for encoding multiple input images, and storage medium for storing programs | |
| EP4510563A1 (en) | Image encoding/decoding method using template matching, method for transmitting bitstream, and recording medium having bitstream stored therein | |
| CN104782128B (en) | Method for three-dimensional or multi-dimensional view video coding and apparatus thereof | |
| TW201143455A (en) | Image encoding device, image decoding device, image encoding method, image decoding method | |
| Suzuki et al. | Block-based reduced resolution inter frame coding with template matching prediction | |
| EP4478707B1 (en) | Image encoding/decoding device for performing bdof, and device for transmitting bitstream | |
| US20170105006A1 (en) | Method and Apparatus for Video Coding Using Master-Slave Prediction Structure | |
| TW201143452A (en) | Image encoding device, image decoding device, image encoding method and image decoding method |