[go: up one dir, main page]

CN104008095A - Object recognition method based on semantic feature extraction and matching - Google Patents

Object recognition method based on semantic feature extraction and matching Download PDF

Info

Publication number
CN104008095A
CN104008095A CN201210556032.0A CN201210556032A CN104008095A CN 104008095 A CN104008095 A CN 104008095A CN 201210556032 A CN201210556032 A CN 201210556032A CN 104008095 A CN104008095 A CN 104008095A
Authority
CN
China
Prior art keywords
class
training
semantic
points
objects
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201210556032.0A
Other languages
Chinese (zh)
Inventor
艾浩军
艾雄军
艾晓敏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Wuhan San Ji Internet Of Things Science And Technology Ltd
Original Assignee
Wuhan San Ji Internet Of Things Science And Technology Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Wuhan San Ji Internet Of Things Science And Technology Ltd filed Critical Wuhan San Ji Internet Of Things Science And Technology Ltd
Priority to CN201210556032.0A priority Critical patent/CN104008095A/en
Publication of CN104008095A publication Critical patent/CN104008095A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/583Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/5838Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using colour

Landscapes

  • Engineering & Computer Science (AREA)
  • Library & Information Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Image Analysis (AREA)

Abstract

The invention provides an object recognition method based on semantic feature extraction and matching and belongs to the field of information retrieval. The object recognition method based on semantic feature extraction and matching includes semantic feature extraction and semantic feature matching. The semantic feature extraction includes firstly extracting SIFT (Scale Invariant Feature Transform) feature points of training images of a class of objects, then performing spatial clustering on the SIFT feature points through k- means clustering, deciding a plurality of efficient points in every space class through a decision-making mechanism based on kernel function, and finally training the efficient points in every space class through a support vector machine classifier; a visual word with semantic features is trained from every space class, and finally a visual vocabulary describing the semantic features of a class of objects is extracted. The semantic feature matching includes firstly extracting SIFT feature points of an image of an object to be detected as the semantic description of the object to be detected, then using the support vector machine classifier for matching and classifying the semantic description of the object to be detected and visual vocabularies of classes of objects, and finally counting a histogram of the visual vocabulary of the object to be detected for determining the class of the object to be detected.

Description

object identification method based on semantic feature extraction and matching
Technical Field
The invention belongs to the field of information retrieval, and particularly relates to an object identification method based on semantic feature extraction and matching.
Background
The essence of object identification is to establish a computing system capable of identifying the object type of interest in an image, which has wide application requirements in real life and has quite high application value and research significance. In recent years, with the development of pattern classification techniques and the continuous development of artificial intelligence, object recognition techniques based on semantic feature extraction are becoming popular among a large number of scholars. The semantic features of the objects are obtained by extracting the local features of one class of objects and then converting the local features into semantic information describing one class of objects according to a certain processing criterion to form a semantic feature model of the one class of objects, so that feasible and effective object classification and identification effects are realized.
In the current field of object recognition, the Bag of Words algorithm is one of the most representative object recognition algorithms. The algorithm considers that an image is composed of several visual words with semantic information. A plurality of local features in the picture are extracted and converted into visual words, a visual word histogram of the picture is generated according to the relation between the visual words and the visual vocabulary, and the visual word histogram expresses the features of the picture, so that the recognition and classification of objects can be effectively realized.
The visual vocabulary in the Bag of Words algorithm is represented by the cluster centers after k-means clustering of local descriptors. And clustering the characteristic points of all pictures of one type of objects, wherein the number of clusters is the vocabulary of the visual word vocabulary, and the visual words of the vocabulary are the clustering centers of each type. However, only a single feature point in the center of a cluster is used as a description of a class of feature points, and local features are not fully utilized, and semantic information generated after clustering is not fully utilized. A single cluster center loses a large amount of semantic information and is not suitable as an effective visual word.
Disclosure of Invention
The invention aims to solve the problems in the prior art and provides a semantic feature extraction method based on kernel function decision.
The invention is realized by the following technical scheme:
a method for extracting and matching semantic features of an object in an object recognition process, the method comprising:
(1) a semantic feature extraction part: firstly, selecting a plurality of pictures of a class of objects as a training library, and extracting SIFT feature points of all the pictures; performing spatial clustering on all SIFT feature points through a k-means clustering algorithm, and then deciding a plurality of effective points in each spatial category by using a decision mechanism based on a kernel function; training effective points in each space category by using a support vector machine classifier, training a visual word with semantic features in each space category, and finally extracting a visual vocabulary table capable of describing the semantic features of a class of objects; selecting training pictures of multiple objects, and extracting the visual vocabulary of each object to form the visual vocabulary of the multiple objects.
(2) And a semantic feature matching part: firstly, extracting SIFT feature points of a picture of an object to be detected as semantic description of the object to be detected; matching and classifying the semantic description of the object to be detected and the visual vocabulary of the objects of multiple types by using a support vector machine classifier; and (4) counting the visual vocabulary histogram of the object to be detected, and determining the category of the object to be detected.
Wherein the semantic feature extraction part comprises the steps of:
(1) selecting a class of objectsFraming training pictures, and extracting SIFT feature points of each picture;
(2) all the characteristic points of the object of the classAre clustered intoExtracting the clustering center of each space class according to the space class
(3) According to a decision mechanism based on kernel function, forMaking a decision on each feature point in the class to obtain a decision value of each feature point
(4) Setting up support vector machine class training data identification. Selecting a plurality of effective characteristic points in a single space category as training points of the category; selecting a plurality of effective points of all other spatial categories as feature training points of other categories;
(5) aiming at the training data marked in the step (4), learning and training are carried out by utilizing an SVM classifier, and the training is carried out to obtain the object of the same typeA visual word, namely a visual vocabulary of the object;
(6) selecting training pictures of multiple classes of objects, and extracting the visual vocabulary of each class to form the visual vocabulary of the multiple classes of objects.
The calculation formula of the step (3) is as follows:
wherein,representing points to be measuredAnd cluster centerThe euclidean distance of (c).
The category in the step (4) is marked as. In the process of training a visual word, a training sample is setThen, the method for determining the training sample class identifier is as follows:
wherein,the spatial class is represented by a set of spatial classes,representing a range of class decision values.
The semantic feature matching section includes the steps of:
(1) aiming at the picture of the object to be identified, SIFT transformation is carried out on the picture, local feature points are detected, and the picture is extractedSIFT descriptor
(2) Visual vocabulary for local descriptors and multi-class objectsMatching and counting each descriptorCorresponding visual wordThe type of object in which it is located;
(3) statistics ofAnd forming a visual word histogram of the unknown object by the matching number of the test descriptors and the visual vocabulary of each object category, and determining the category of the object to be tested.
The calculation formula of the step (2) is as follows:
compared with the prior art, the invention has the beneficial effects that: in the semantic feature extraction stage, the semantic loss phenomenon caused by taking the original single clustering center as a visual word is replaced, a plurality of effective points in each space category are selected as the visual word through a kernel function decision mechanism, and semantic feature information with stable categories and rich information is extracted.
Drawings
The invention is described in further detail below with reference to the accompanying drawings:
FIG. 1 is a block diagram of the semantic feature extraction step of the present invention.
FIG. 2 is a block diagram of the semantic feature matching step of the present invention.
Detailed Description
1. A semantic feature extraction part: as shown in fig. 1, firstly, selecting a plurality of pictures of a class of objects as a training library, and extracting SIFT feature points of all the pictures; performing spatial clustering on all SIFT feature points through a k-means clustering algorithm, and then deciding a plurality of effective points in each spatial category by using a decision mechanism based on a kernel function; training effective points in each space category by using a support vector machine classifier, training a visual word with semantic features in each space category, and finally extracting a visual vocabulary table capable of describing the semantic features of a class of objects; selecting training pictures of multiple objects, and extracting the visual vocabulary of each object to form the visual vocabulary of the multiple objects.
2. And a semantic feature matching part: as shown in fig. 2, firstly, SIFT feature points of a picture of an object to be detected are extracted as semantic descriptions of the object to be detected; matching and classifying the semantic description of the object to be detected and the visual vocabulary of the objects of multiple types by using a support vector machine classifier; and (4) counting the visual vocabulary histogram of the object to be detected, and determining the category of the object to be detected.
The semantic feature extraction part comprises the following steps:
(1) selecting a class of objectsFraming training pictures, and extracting SIFT feature points of each picture;
(2) all the characteristic points of the object of the classAre clustered intoExtracting the clustering center of each space class according to the space class
(3) According to a decision mechanism based on kernel function, forMaking a decision on each feature point in the class to obtain a decision value of each feature point
(4) Setting up support vector machine class training data identification. Selecting a plurality of effective characteristic points in a single space category as training points of the category; selecting a plurality of effective points of all other spatial categories as feature training points of other categories;
(5) aiming at the training data marked in the step (4), learning and training are carried out by utilizing an SVM classifier, and the training is carried out to obtain the object of the same typeA visual word, namely a visual vocabulary of the object;
(6) selecting training pictures of multiple classes of objects, and extracting the visual vocabulary of each class to form the visual vocabulary of the multiple classes of objects.
The calculation formula of the step (3) is as follows:
wherein,representing points to be measuredAnd cluster centerThe euclidean distance of (c).
The category in the step (4) is marked as. In the process of training a visual word, a training sample is setThen, the method for determining the training sample class identifier is as follows:
wherein,the spatial class is represented by a set of spatial classes,representing a range of class decision values.
The semantic feature matching section includes the steps of:
(1) aiming at the picture of the object to be identified, SIFT transformation is carried out on the picture, local feature points are detected, and the picture is extractedSIFT descriptor
(2) Visual vocabulary for local descriptors and multi-class objectsMatching and counting each descriptorCorresponding visual wordThe type of object in which it is located;
(3) statistics ofAnd forming a visual word histogram of the unknown object by the matching number of the test descriptors and the visual vocabulary of each object category, and determining the category of the object to be tested.
The calculation formula of the step (2) is as follows:

Claims (6)

1. An object recognition method based on semantic feature extraction and matching, the method comprising:
(1) a semantic feature extraction part: firstly, selecting a plurality of pictures of a class of objects as a training library, and extracting SIFT feature points of all the pictures; performing spatial clustering on all SIFT feature points through a k-means clustering algorithm, and then deciding a plurality of effective points in each spatial category by using a decision mechanism based on a kernel function; training effective points in each space category by using a support vector machine classifier, training a visual word with semantic features in each space category, and finally extracting a visual vocabulary table capable of describing the semantic features of a class of objects; selecting training pictures of multiple objects, and extracting a visual vocabulary list of each object to form the visual vocabulary list of the multiple objects;
(2) and a semantic feature matching part: firstly, extracting SIFT feature points of a picture of an object to be detected as semantic description of the object to be detected; matching and classifying the semantic description of the object to be detected and the visual vocabulary of the objects of multiple types by using a support vector machine classifier; and (4) counting the visual vocabulary histogram of the object to be detected, and determining the category of the object to be detected.
2. The method according to claim 1, wherein the semantic feature extraction section comprises the steps of:
(1) selecting a class of objectsFraming training pictures, and extracting SIFT feature points of each picture;
(2) all the characteristic points of the object of the classAre clustered intoExtracting the clustering center of each space class according to the space class
(3) According to a decision mechanism based on kernel function, forMaking a decision on each feature point in the class to obtain a decision value of each feature point
(4) Setting up support vector machine class training data identificationSelecting a plurality of effective characteristic points in a single space category as training points of the category; selecting a plurality of effective points of all other spatial categories as feature training points of other categories;
(5) aiming at the training data marked in the step (4), learning and training are carried out by utilizing an SVM classifier, and the training is carried out to obtain the object of the same typeA visual word, namely a visual vocabulary of the object;
(6) selecting training pictures of multiple classes of objects, and extracting the visual vocabulary of each class to form the visual vocabulary of the multiple classes of objects.
3. The method according to claim 1, wherein the semantic feature matching component comprises the steps of:
(1) aiming at the picture of the object to be identified, SIFT transformation is carried out on the picture, local feature points are detected, and the picture is extractedSIFT descriptor
(2) Visual vocabulary for local descriptors and multi-class objectsMatching and counting each descriptorCorresponding visual wordThe type of object in which it is located;
(3) statistics ofAnd forming a visual word histogram of the unknown object by the matching number of the test descriptors and the visual vocabulary of each object category, and determining the category of the object to be tested.
4. The method of claim 2, wherein the calculation formula of step (3) is as follows:
wherein,representing points to be measuredAnd cluster centerThe euclidean distance of (c).
5. The method of claim 2, wherein the category is identified as in step (4)In the process of training a visual word, a training sample is setThen, the method for determining the training sample class identifier is as follows:
wherein,the spatial class is represented by a set of spatial classes,representing a range of class decision values.
6. The method of claim 3, wherein the calculation formula of step (2) is as follows:
CN201210556032.0A 2013-02-25 2013-02-25 Object recognition method based on semantic feature extraction and matching Pending CN104008095A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201210556032.0A CN104008095A (en) 2013-02-25 2013-02-25 Object recognition method based on semantic feature extraction and matching

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201210556032.0A CN104008095A (en) 2013-02-25 2013-02-25 Object recognition method based on semantic feature extraction and matching

Publications (1)

Publication Number Publication Date
CN104008095A true CN104008095A (en) 2014-08-27

Family

ID=51368754

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201210556032.0A Pending CN104008095A (en) 2013-02-25 2013-02-25 Object recognition method based on semantic feature extraction and matching

Country Status (1)

Country Link
CN (1) CN104008095A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105139390A (en) * 2015-08-14 2015-12-09 四川大学 Image processing method for detecting pulmonary tuberculosis focus in chest X-ray DR film
CN105389593A (en) * 2015-11-16 2016-03-09 上海交通大学 Image object recognition method based on SURF
CN107624192A (en) * 2015-05-11 2018-01-23 西门子公司 The system and method for pathology in the surgical guidance and art broken up by endoscopic tissue
CN112612914A (en) * 2020-12-29 2021-04-06 浙江金实乐环境工程有限公司 Image garbage recognition method based on deep learning

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101944183A (en) * 2010-09-02 2011-01-12 北京航空航天大学 Method for identifying object by utilizing SIFT tree
US20110123120A1 (en) * 2008-06-03 2011-05-26 Eth Zurich Method and system for generating a pictorial reference database using geographical information
CN102629328A (en) * 2012-03-12 2012-08-08 北京工业大学 Probabilistic latent semantic model object image recognition method with fusion of significant characteristic of color

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110123120A1 (en) * 2008-06-03 2011-05-26 Eth Zurich Method and system for generating a pictorial reference database using geographical information
CN101944183A (en) * 2010-09-02 2011-01-12 北京航空航天大学 Method for identifying object by utilizing SIFT tree
CN102629328A (en) * 2012-03-12 2012-08-08 北京工业大学 Probabilistic latent semantic model object image recognition method with fusion of significant characteristic of color

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
王红霞 等: "物体分类识别中视觉词汇生成方法研究", 《武汉理土大学学报(信息与管理工程版)》 *

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107624192A (en) * 2015-05-11 2018-01-23 西门子公司 The system and method for pathology in the surgical guidance and art broken up by endoscopic tissue
CN107624192B (en) * 2015-05-11 2020-10-27 西门子公司 System and method for surgical guidance and intraoperative pathology by endoscopic tissue differentiation
US11380084B2 (en) 2015-05-11 2022-07-05 Siemens Aktiengesellschaft System and method for surgical guidance and intra-operative pathology through endo-microscopic tissue differentiation
CN105139390A (en) * 2015-08-14 2015-12-09 四川大学 Image processing method for detecting pulmonary tuberculosis focus in chest X-ray DR film
CN105389593A (en) * 2015-11-16 2016-03-09 上海交通大学 Image object recognition method based on SURF
CN105389593B (en) * 2015-11-16 2019-01-11 上海交通大学 Image object recognition methods based on SURF feature
CN112612914A (en) * 2020-12-29 2021-04-06 浙江金实乐环境工程有限公司 Image garbage recognition method based on deep learning

Similar Documents

Publication Publication Date Title
Zhao et al. Learning mid-level filters for person re-identification
Wu et al. Harvesting discriminative meta objects with deep CNN features for scene classification
Zheng et al. Person re-identification meets image search
CN106156777B (en) Text picture detection method and device
CN103425996B (en) A kind of large-scale image recognition methods of parallel distributed
CN102622607A (en) Remote sensing image classification method based on multi-feature fusion
CN103440508B (en) The Remote Sensing Target recognition methods of view-based access control model word bag model
Zhang et al. Automatic discrimination of text and non-text natural images
CN104915673A (en) Object classification method and system based on bag of visual word model
JP4553300B2 (en) Content identification device
CN106845375A (en) A kind of action identification method based on hierarchical feature learning
CN103279738A (en) Automatic identification method and system for vehicle logo
CN108073940B (en) Method for detecting 3D target example object in unstructured environment
CN102609718B (en) Method for generating vision dictionary set by combining different clustering algorithms
CN104008095A (en) Object recognition method based on semantic feature extraction and matching
CN104732209B (en) A kind of recognition methods of indoor scene and device
Najibi et al. Towards the success rate of one: Real-time unconstrained salient object detection
JP5959446B2 (en) Retrieval device, program, and method for high-speed retrieval by expressing contents as a set of binary feature vectors
Liu et al. Video retrieval based on object discovery
Gupta et al. The semantic multinomial representation of images obtained using dynamic kernel based pseudo-concept SVMs
Golge et al. FAME: face association through model evolution
Zhi-Jie Image classification method based on visual saliency and bag of words model
Li et al. Image categorization based on visual saliency and Bag-of-Words model
Dong et al. Superpixel appearance and motion descriptors for action recognition
Mansur et al. Improving recognition through object sub-categorization

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20140827

WD01 Invention patent application deemed withdrawn after publication