[go: up one dir, main page]

CN103426006A - Self-adaption multi-feature fusion image feature learning method - Google Patents

Self-adaption multi-feature fusion image feature learning method Download PDF

Info

Publication number
CN103426006A
CN103426006A CN2013103420236A CN201310342023A CN103426006A CN 103426006 A CN103426006 A CN 103426006A CN 2013103420236 A CN2013103420236 A CN 2013103420236A CN 201310342023 A CN201310342023 A CN 201310342023A CN 103426006 A CN103426006 A CN 103426006A
Authority
CN
China
Prior art keywords
image
feature
features
self
sigma
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN2013103420236A
Other languages
Chinese (zh)
Inventor
毛金莲
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang Business College
Original Assignee
Zhejiang Business College
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang Business College filed Critical Zhejiang Business College
Priority to CN2013103420236A priority Critical patent/CN103426006A/en
Publication of CN103426006A publication Critical patent/CN103426006A/en
Pending legal-status Critical Current

Links

Images

Landscapes

  • Image Analysis (AREA)

Abstract

The invention discloses a self-adaption multi-feature fusion image feature learning method. The method comprises the steps of first respectively extracting image visual features of various types according to the characteristic that an L1 norm has automatic data sample selection, then respectively adopting data sparse expression constrained by the L1 norm for every visual feature, establishing a directed L1 map, minimizing reestablishment errors of various feature data on a low-dimensional space, introducing feather weight vectors to perform global coordinate alignment operation on all the features, obtaining an objective function of the self-adaption multi-feature fusion image feature learning method, finally solving the objective function, and obtaining the optimal image low-dimensional feature expression by learning on the basis of original multiple image features. The image low-dimensional feature expression obtained by the method serves as input of a nearest neighbor classifier algorithm; compared with other algorithms on a public image data set Core15K and NUS-WIDE-OBJECT, the method improve the classification accuracy rates respectively by 5% and 2%, and accordingly the image features obtained by the method have high expression capability.

Description

The characteristics of image learning method of the many Fusion Features of a kind of self-adaptation
Technical field
The present invention relates to sparse expression, figure study, feature learning, relate in particular to the characteristics of image study of many Fusion Features.
Background technology
In computer vision, pattern-recognition and information retrieval field, owing to inevitably being faced with the semantic gap problem, low-level image feature can't be expressed high-layer semantic information fully exactly; Therefore, in order to express as far as possible all-sidedly and accurately the object that will describe, often can, to the polytype visual signature of this object extraction, from different visual angles, it be expressed.For example: in Images Classification, to coloured image, often can extract its color histogram feature, profile moment characteristics, textural characteristics, sift point feature etc.From traditional to express object based on single type feature different, these characteristics, be distributed in different feature spaces, has its inherent statistical property and real physical significance, and its ability to express (or resolving ability) is the property of there are differences also.In order to take full advantage of the characteristic of different characteristic, accelerating algorithm is processed, and usually needs to carry out many feature learnings according to these many features, obtains optimum low-dimensional feature representation.
Researchers wish by utilizing many feature learnings fully to excavate out the complementary information between different characteristic, thereby obtain a low-dimensional nested result better than the single feature learning of direct utilization.In existing algorithm, the spectrum nested algorithm has very important status.Then, traditional spectrum nested algorithm is all that tentation data is to be distributed in same vector space, thereby can't be applied directly in many feature learnings.The solution existed at present has two: one is that different features first is stitched together, and is then composed nested study.Obvious this way lacks clear and definite physical significance support.Another way is to use the nested study of distributed spectrum (DSE) algorithm, and different features is learnt respectively.Yet many characteristics original in this method are sightless for final learning process, can't fully effectively utilize the complementary information between different characteristic.Under the impact of the many feature learnings universal model proposed people such as Long, the people such as Xia have proposed characteristic spectrum nested algorithm more than, in different characteristic study, obtain simultaneously a low-dimensional, enough level and smooth nested, thereby fully effectively utilize the complementary information between different characteristic.
Yet, because all being based on spectral graph theory, overwhelming majority various visual angles learning algorithm develops, and a core of its algorithm is the adjacency matrix (similar matrix) that utilizes space structure relation between various criteria construction expression datas.The Usefulness Pair of the construction algorithm of adjacency matrix plays vital effect in the performance of whole algorithm.Now widely used two kinds of adjacency matrix construction methods: based on k nearest neighbor and the patterning process based on thermonuclear (heat kernel), be not the patterning process of data adaptive.The former need to set in advance neighboring node number k, and the latter need to arrange nuclear parameter.In actual applications, the setting of these two parameters often by virtue of experience or utilize the mode of cross validation to select best parameter, lacks simple directly method and obtains optimal parameter.Meanwhile, the algorithm of this class based on figure, be again highstrung for the selection of these two parameters sometimes.The quality that parameter is selected directly has influence on the performance of algorithm.Therefore, several existing work noted earlier all is subject to lacking the puzzlement of data adaptive problem.How building the figure with data adaptive, improve the adaptivity of algorithm, is a problem in the urgent need to address.
In the last few years, the L1 norm was subject to extensive concern in sparse expression and compressed sensing research field.Be different from traditional L2 norm constraint, the L1 norm constraint can, so that the solution of problem becomes sparse, have stronger anti-noise ability.According to the people's such as Cheng research, show, L1 figure has three key properties: (1) noise resisting ability is strong; (2) sparse property; (3) neighborhood relationships of data adaptive.
From the above, in order to overcome traditional poor problem of many feature learnings algorithm adaptivity based on figure, the characteristics of image learning method of the many Fusion Features of self-adaptation that this paper proposes is to have according to the L1 norm characteristic that the automaticdata sample is selected, at first extract respectively polytype Image Visual Feature from image, then to each visual signature, adopt respectively the Sparse of L1 norm constraint to express, be built with to L1 and scheme, the good characteristic that L1 figure is had is dissolved in many feature learnings algorithm, and descend the reconstruction error of data on lower dimensional space under the various features of change most, secondly the introduced feature weight vectors is done the world coordinates alignment operation to all features, obtain the objective function of the image feature selection method of the many Fusion Features of self-adaptation, last this objective function of Optimization Solution, can obtain adaptively the low-dimensional feature representation of image optimum from original multiple characteristics of image basis learning.
Summary of the invention
The objective of the invention is to overcome the deficiencies in the prior art, propose the characteristics of image learning method of the many Fusion Features of a kind of self-adaptation.
The characteristics of image learning method of the many Fusion Features of self-adaptation is to comprise the steps:
1) extract respectively the Image Visual Feature of K type from N width image
Figure BDA00003634054200021
Figure BDA00003634054200022
Wherein
Figure BDA00003634054200023
It is the visual signature of v type extracted on i width image;
2) there is the characteristic of automaticdata feature selecting according to the L1 norm, to each visual signature, adopt respectively the Sparse of L1 norm constraint to express:
min | | x i v - X v w i v | | F 2 + λ | | w i v | | 1 1
s . t . w ii v = 0
According to what calculate
Figure BDA00003634054200026
Be built with the g to L1 figure v={ X v, W v, the type characteristics of image X wherein vThe weight on the ,Tu De limit, summit of pie graph
Figure BDA00003634054200027
Descend most the reconstruction error of data on lower dimensional space under the various features of change:
min Y v Σ i = 1 N | | Y i v - Y v w i v | | 2 2 - - - 2
3) introduced feature weight vectors α=[α 1..., α K], all features are done to the world coordinates alignment operation, the characteristics of image learning method objective function that obtains the many Fusion Features of self-adaptation is as follows:
min α , Y Σ v = 1 K ( α v ) r Σ i = 1 N | | Y - Y w i v | | 2 2 3
s . t . YY T = I , Σ v = 1 K α v = 1 , α v ≥ 0 , ∀ v
Wherein, r is evolution number of times parameter;
4) top objective function 3 is optimized and solves, adopt Lagrangian multiplication, obtain computing formula as follows:
α v = ( 1 / tr ( YG v Y T ) 1 / ( r - 1 ) ) Σ v = 1 K ( 1 / tr ( YG v Y T ) 1 / ( r - 1 ) ) - - - 4
Y=select_svd(G,d) 5
Wherein: tr () is the computing of Matrix Calculating mark, G v=(I-W v) (I-W v) T,
Figure BDA00003634054200032
Select_svd (G, d) means to G is done Eigenvalues Decomposition and gets front d the corresponding proper vector of eigenwert of eigenwert minimum, formation matrix, the low-dimensional feature representation of the image optimum that Y is from original multiple characteristics of image basis to learn to obtain.
The beneficial effect that the present invention compared with prior art has is, algorithm has data adaptive, fast operation, and the characteristics of image of learning to obtain has stronger ability to express, can effectively improve successive image sorting algorithm accuracy rate.
The accompanying drawing explanation
Fig. 1 is the part sample image on the Core5K image data set;
Fig. 2 is the part sample image on the NUS-WIDE-OBJECT data set;
Fig. 3 be on the Corel5K image data set each algorithm pattern as the classification accuracy comparing result;
Fig. 4 be on the NUS-WIDE-OBJECT image data set each algorithm pattern as the classification accuracy comparing result;
Fig. 5 is the neighbour's statistics of the constructed L1 figure of this method on the Corel5K image data set;
Fig. 6 is the neighbour's statistics of the constructed L1 figure of this method on the NUS-WIDE-OBJECT image data set.
Embodiment
The characteristics of image learning method of the many Fusion Features of self-adaptation comprises the steps:
1) extract respectively the Image Visual Feature of K type from N width image
Figure BDA00003634054200033
Figure BDA00003634054200034
Wherein
Figure BDA00003634054200035
It is the visual signature of v type extracted on i width image;
2) there is the characteristic of automaticdata feature selecting according to the L1 norm, to each visual signature, adopt respectively the Sparse of L1 norm constraint to express:
min | | x i v - X v w i v | | F 2 + λ | | w i v | | 1 1
s . t . w ii v = 0
According to what calculate Be built with the g to L1 figure v={ X v, W v, the type characteristics of image X wherein vThe weight on the ,Tu De limit, summit of pie graph
Figure BDA00003634054200039
Descend most the reconstruction error of data on lower dimensional space under the various features of change:
min Y v Σ i = 1 N | | Y i v - Y v w i v | | 2 2 - - - 2
3) introduced feature weight vectors α=[α 1..., α K], all features are done to the world coordinates alignment operation, the characteristics of image learning method objective function that obtains the many Fusion Features of self-adaptation is as follows:
min α , Y Σ v = 1 K ( α v ) r Σ i = 1 N | | Y - Y w i v | | 2 2 3
s . t . YY T = I , Σ v = 1 K α v = 1 , α v ≥ 0 , ∀ v
Wherein, r is evolution number of times parameter;
4) top objective function 3 is optimized and solves, adopt Lagrangian multiplication, can obtain computing formula as follows:
α v = ( 1 / tr ( YG v Y T ) 1 / ( r - 1 ) ) Σ v = 1 K ( 1 / tr ( YG v Y T ) 1 / ( r - 1 ) ) - - - 4
Y=select_svd(G,d) 5
Wherein: tr () is the computing of Matrix Calculating mark, G v=(I-W v) (I-W v) T,
Figure BDA00003634054200044
Select_svd (G, d) means to G is done Eigenvalues Decomposition and gets front d the corresponding proper vector of eigenwert of eigenwert minimum, formation matrix, the low-dimensional feature representation of the image optimum that Y is from original multiple characteristics of image basis to learn to obtain.
Will be on two common image test data set Corel5K and NUS-WIDE-OBJECT data set, the impact of the characteristics of image learning algorithm (being designated as AMVL) of the many Fusion Features of contrast test self-adaptation on the Images Classification accuracy rate.The characteristics of image of announcing according to the NUS-WIDE-OBJECT data set, respectively these two concentrated images of data are extracted to following 5 kinds of features: LAB color histogram (64 dimension), hsv color autocorrelogram (144 dimension), edge orientation histogram (73 dimension), wavelet texture (128 dimension) and piecemeal color moment feature (225 dimension) obtain 5 different visual angles features.Test the specifying information of data set used in Table 1.
Table 1. embodiment data set information
Figure BDA00003634054200045
Use the PCA(principal component analysis (PCA)), the classification of SRC(sparse expression), the mono-feature adaptive learning of ASVL() compose from various visual angles nested study with MSE() method of testing as a comparison.Wherein, PCA and SRC are two kinds of unsupervised learning algorithms at present commonly used, and ASVL is single feature learning special case of the AMVL algorithm that proposes of this paper, and MSE is a various visual angles learning algorithm.Adopt nearest neighbor classifier as Images Classification benchmark algorithm, contrast each algorithm performance.
In force, the needed parameter lambda of this method=0.01 is set, r=2, and all adopt parameter traversal mode to obtain optimum parameter setting for other algorithms, then contrast in the situation that acquire the characteristics of image of different dimensions, the performance change situation of each algorithm.
The characteristics of image learning method that Fig. 3 and Fig. 4 have showed the many Fusion Features of self-adaptation that propose and contrast algorithm be the classification accuracy comparing result above image data set at two.Can find out that from the result of Fig. 3 and Fig. 4 the classification accuracy of the resulting characteristics of image of this method on most characteristic dimension all can be much better than other contrast algorithms.In characteristic dimension, be 60 o'clock, Corel5K is upper, and the method AMVL proposed improves 5% than other algorithms, at NUS-WIDE-OBJECT, improves 2%.Fig. 5 and Fig. 6 have showed neighbour's statistics of the L1 figure that on Corel5K and NUS-WIDE-OBJECT image data set this method is constructed.This method can be selected neighbour's number adaptively as can be known from Fig. 5 and Fig. 6.

Claims (1)

1. the characteristics of image learning method of the many Fusion Features of self-adaptation, is characterized in that comprising the steps:
1) extract respectively the Image Visual Feature of K type from N width image
Figure FDA00003634054100011
Wherein
Figure FDA00003634054100013
It is the visual signature of v type extracted on i width image;
2) there is the characteristic of automaticdata feature selecting according to the L1 norm, to each visual signature, adopt respectively the Sparse of L1 norm constraint to express:
min | | x i v - X v w i v | | F 2 + λ | | w i v | | 1 1
s . t . w ii v = 0
According to what calculate
Figure FDA00003634054100016
Be built with the g to L1 figure v={ X v, W v, the type characteristics of image X wherein vThe weight on the ,Tu De limit, summit of pie graph Descend most the reconstruction error of data on lower dimensional space under the various features of change:
min Y v Σ i = 1 N | | Y i v - Y v w i v | | 2 2 - - - 2
3) introduced feature weight vectors α=[α 1..., α K], all features are done to the world coordinates alignment operation, the characteristics of image learning method objective function that obtains the many Fusion Features of self-adaptation is as follows:
min α , Y Σ v = 1 K ( α v ) r Σ i = 1 N | | Y - Y w i v | | 2 2 3
s . t . YY T = I , Σ v = 1 K α v = 1 , α v ≥ 0 , ∀ v
Wherein, r is evolution number of times parameter;
4) top objective function 3 is optimized and solves, adopt Lagrangian multiplication, obtain computing formula as follows:
α v = ( 1 / tr ( YG v Y T ) 1 / ( r - 1 ) ) Σ v = 1 K ( 1 / tr ( YG v Y T ) 1 / ( r - 1 ) ) - - - 4
Y=select_svd(G,d) 5
Wherein: tr () is the computing of Matrix Calculating mark, G v=(I-W v) (I-W v) T,
Figure FDA000036340541000112
Select_svd (G, d) means to G is done Eigenvalues Decomposition and gets front d the corresponding proper vector of eigenwert of eigenwert minimum, formation matrix, the low-dimensional feature representation of the image optimum that Y is from original multiple characteristics of image basis to learn to obtain.
CN2013103420236A 2013-08-07 2013-08-07 Self-adaption multi-feature fusion image feature learning method Pending CN103426006A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN2013103420236A CN103426006A (en) 2013-08-07 2013-08-07 Self-adaption multi-feature fusion image feature learning method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN2013103420236A CN103426006A (en) 2013-08-07 2013-08-07 Self-adaption multi-feature fusion image feature learning method

Publications (1)

Publication Number Publication Date
CN103426006A true CN103426006A (en) 2013-12-04

Family

ID=49650715

Family Applications (1)

Application Number Title Priority Date Filing Date
CN2013103420236A Pending CN103426006A (en) 2013-08-07 2013-08-07 Self-adaption multi-feature fusion image feature learning method

Country Status (1)

Country Link
CN (1) CN103426006A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2016080427A (en) * 2014-10-14 2016-05-16 三菱電機株式会社 Signal processor
CN109191495A (en) * 2018-07-17 2019-01-11 东南大学 Black smoke vehicle detection method based on self-organizing background subtraction model and multiple features fusion

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090220146A1 (en) * 2008-03-01 2009-09-03 Armin Bauer Method and apparatus for characterizing the formation of paper
CN102663453A (en) * 2012-05-03 2012-09-12 西安电子科技大学 Human motion tracking method based on second generation Bandlet transform and top-speed learning machine

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090220146A1 (en) * 2008-03-01 2009-09-03 Armin Bauer Method and apparatus for characterizing the formation of paper
CN102663453A (en) * 2012-05-03 2012-09-12 西安电子科技大学 Human motion tracking method based on second generation Bandlet transform and top-speed learning machine

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
毛金莲: "自适应多视角学习及其在图像分类中的应用", 《计算机应用》, vol. 33, no. 7, 1 July 2013 (2013-07-01), pages 1955 - 1959 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2016080427A (en) * 2014-10-14 2016-05-16 三菱電機株式会社 Signal processor
CN109191495A (en) * 2018-07-17 2019-01-11 东南大学 Black smoke vehicle detection method based on self-organizing background subtraction model and multiple features fusion

Similar Documents

Publication Publication Date Title
CN109447994B (en) Remote Sensing Image Segmentation Method Combining Complete Residual and Feature Fusion
CN114092697B (en) Building facade semantic segmentation method with attention fused with global and local depth features
CN108229338B (en) Video behavior identification method based on deep convolution characteristics
CN103761742B (en) A kind of high-spectrum remote sensing sparse solution mixing method based on homogeneity index
CN101877128A (en) A Segmentation Method of Different Objects in 3D Scene
CN106548153A (en) Video abnormality detection method based on graph structure under multi-scale transform
CN104091169A (en) Behavior identification method based on multi feature fusion
CN107292259A (en) The integrated approach of depth characteristic and traditional characteristic based on AdaRank
CN103617618A (en) SAR image segmentation method based on feature extraction and cluster integration
CN103198479A (en) SAR image segmentation method based on semantic information classification
CN103226825B (en) Based on the method for detecting change of remote sensing image of low-rank sparse model
CN106611421A (en) SAR image segmentation method based on feature learning and sketch line constraint
CN103310230A (en) Method for classifying hyperspectral images on basis of combination of unmixing and adaptive end member extraction
CN105678218A (en) Moving object classification method
CN103606164A (en) SAR image segmentation method based on high-dimensional triple Markov field
CN103914558A (en) Method for mining space-time aggregation patterns of meteorological elements on basis of space-time statistics
CN102800113B (en) Digital image analysis method based on fractal dimension
CN105095913A (en) Remote sensing image classification method and system based on neighbor regular joint sparse representation
CN103136545A (en) High resolution remote sensing image analysis tree automatic extraction method based on space consistency
CN103617413A (en) Method for identifying object in image
CN102663681B (en) Gray scale image segmentation method based on sequencing K-mean algorithm
CN105046268A (en) Polarization SAR image classification method based on Wishart deep network
CN105303546A (en) Affinity propagation clustering image segmentation method based on fuzzy connectedness
CN111738052A (en) Deep learning-based multi-feature fusion hyperspectral remote sensing feature classification method
CN103310229B (en) A kind of multitask machine learning method for image classification and device thereof

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20131204