[go: up one dir, main page]

CN108415955B - Interest point database establishing method based on eye movement fixation point moving track - Google Patents

Interest point database establishing method based on eye movement fixation point moving track Download PDF

Info

Publication number
CN108415955B
CN108415955B CN201810116386.0A CN201810116386A CN108415955B CN 108415955 B CN108415955 B CN 108415955B CN 201810116386 A CN201810116386 A CN 201810116386A CN 108415955 B CN108415955 B CN 108415955B
Authority
CN
China
Prior art keywords
point
interest
model
points
eye
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201810116386.0A
Other languages
Chinese (zh)
Other versions
CN108415955A (en
Inventor
张桦
戴美想
戴国骏
周文晖
王彧
张悦
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hangzhou Dianzi University
Original Assignee
Hangzhou Dianzi University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hangzhou Dianzi University filed Critical Hangzhou Dianzi University
Priority to CN201810116386.0A priority Critical patent/CN108415955B/en
Publication of CN108415955A publication Critical patent/CN108415955A/en
Application granted granted Critical
Publication of CN108415955B publication Critical patent/CN108415955B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/51Indexing; Data structures therefor; Storage structures
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/013Eye tracking input arrangements
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Geometry (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Computer Graphics (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Processing Or Creating Images (AREA)

Abstract

本发明公开了一种基于眼动注视点移动轨迹的感兴趣点数据库建立方法。本方法将三维模型利用投影技术投影成二维图片,并将图片做成视频,放置于眼动仪让志愿者观看,进而收集眼动注视点的数据,然后将视频按帧分割,提取图片中的眼动注视点,借助二维到三维的映射技术得到眼动注视点点在三维模型上的坐标点。最后使用建立Ground Truth的算法将得到的眼动注视点数据建立成感兴趣点数据库。本发明更加适用于对模型的重建,因为利用眼动仪可以将模型上人眼最关注区域的部分和不关注区域的部分进行划分。通过感兴趣点标准库对人类感兴趣的地方可以加强重建的精度,相反对于不感兴趣的区域相对减少重建精度,从而减少模型重建的工作量和存储量。

Figure 201810116386

The invention discloses a method for establishing a database of interest points based on the movement trajectory of eye movement gaze points. In this method, the three-dimensional model is projected into a two-dimensional picture by projection technology, and the picture is made into a video, which is placed on the eye tracker for volunteers to watch, and then the data of the eye-movement fixation point is collected. The eye movement fixation point is obtained, and the coordinate point of the eye movement fixation point on the 3D model is obtained by means of 2D to 3D mapping technology. Finally, the algorithm of establishing Ground Truth is used to build the obtained eye movement gaze point data into a database of interest points. The present invention is more suitable for the reconstruction of the model, because the eye tracker can be used to divide the part of the most concerned area of the human eye and the part of the unconcerned area on the model. Through the standard library of points of interest, the accuracy of reconstruction can be enhanced for the places of interest to humans. On the contrary, the reconstruction accuracy is relatively reduced for areas that are not of interest, thereby reducing the workload and storage of model reconstruction.

Figure 201810116386

Description

Interest point database establishing method based on eye movement fixation point moving track
Technical Field
The invention relates to a method for establishing a point-of-interest database of a three-dimensional model by collecting eye movement fixation point movement tracks when a human eye observes the three-dimensional model by using an eye movement instrument, analyzing and processing the tracks.
Background
The point-of-interest database, which is currently used as a three-dimensional model with a large reference amount, is a text "Evaluation of 3D interest point detection technologies via human-generated ground route" published in Vis Comut by Helin Dutagaci et al in 2012. In the experiment, 24 three-dimensional models are displayed on a window of a website, 23 experimenters can obtain views from different angles of the models through keys arranged on the window, and the experimenters can mark the interested areas of the models through a mouse after observing the views. And collecting data marked by the experimenter, eliminating pseudo interest points through an algorithm, and integrating an interest point data set of the three-dimensional model.
The points of interest obtained by this way of the database are not objective. Because the experimenter observes the data for a long time, the selected corner points are strongly subjective through brain thinking. This is in contrast to the point of interest where the model comes into sight of the human eye, drawing the attention of the human vision the first time. The invention establishes the interest point database through the eye movement fixation point moving track, and can truly reflect the interest points when the human observes the model.
Disclosure of Invention
In view of the fact that the current three-dimensional model interest point data set cannot reflect the reality of a human being when observing the model. The invention provides a method for establishing an interest point database based on eye movement fixation point movement tracks. The interest point database established by the method can reflect the real condition of human eyes when observing the model, and the data is more real and reliable.
In order to achieve the purpose, the invention is realized by the following technical scheme, which specifically comprises the following 4 steps:
and (1) collecting the three-dimensional model and making the three-dimensional model into a video material required by the experiment.
And (2) placing the video on an eye movement instrument for an experimenter to watch, obtaining data of the eye movement fixation point, and synthesizing the video with the eye movement fixation point through corresponding software.
And (3) generating a three-dimensional model with the eye movement fixation point by using the video with the eye movement fixation point through a motion point extraction and three-dimensional mapping algorithm.
And (4) analyzing the eye movement fixation points of the experimenters, sorting the eye movement fixation points of all the experimenters to obtain a three-dimensional model interest point set, discarding some inappropriate and abnormal data, merging the interest points and establishing an interest point database.
The method for making the video material in the step (1) is as follows:
the 24 stanford three-dimensional model libraries and models in the SHREC2007 model database were selected, and these three-dimensional models are widely used in the standard libraries for three-dimensional model research. And saving two groups of data for each three-dimensional model by using MATLAB, wherein one group is formed by sequentially rotating the three-dimensional models around an X axis at intervals of 60 degrees: three-dimensional models which rotate 0 degree, 60 degree, 120 degree, 180 degree, 240 degree, 300 degree and 360 degree respectively; one group is that the three-dimensional model rotates around the Y axis at intervals of 60 degrees in sequence: three-dimensional models which rotate 0 degree, 60 degree, 120 degree, 180 degree, 240 degree, 300 degree and 360 degree respectively; the Z-axis direction is selected as a viewpoint, and the model at 12 angles is projected on the XOY plane, so that 12 two-dimensional projection pictures are obtained by one model.
A formula for three-dimensional model rotation and a formula for projection onto the XOY plane.
Rotation around the X axis:
z′=zcosθ-xsinθ
x′=xcosθ+zsinθ
y′=y
rotation around the Y axis:
z′=zcosθ-ysinθ
x′=x
y′=ycosθ+zsinθ
formula for parallel projection:
x′=x
y′=y
z′=0
and transforming a picture every 1.5 seconds by using a video editor Movie Maker, synthesizing 12 two-dimensional pictures of each model into a short video, inserting a blank picture of ten seconds between every two models as rest time, and synthesizing 6 models into a long video. A total of 4 long videos were made from 24 models.
The precondition for the experimenter to carry out the experiment in the step (2) is as follows:
a) the display device of the image is placed on the left side, and the experimenter sits right in front of the display device of the image and keeps the distance at 70cm, while the experimenter keeps the eyes in accordance with the height of the screen, and can look up the center of the screen.
b) The operating personnel is on the right side, and with computer control video broadcast on display device, the operating personnel is kept apart with the light shield baffle with the experimenter.
c) The experiment interference to the experimenter caused by other light sources is prevented by isolating baffles around the experimenter and isolating curtains around the laboratory.
d) The sound of the environment of the laboratory is kept not more than 30dB, an ideal quiet environment is created, and the experiment interference of other sound sources to an experimenter is prevented.
The specific experiment in the step (2) is as follows:
the eye position of the test person was first adjusted using the iViewX software. After the pupil Image of the experimenter appears in the Eye Image frame, an operator needs to adjust the relative position of the screen and the experimenter, so that the pupil Image on the screen can be stably displayed in a centered mode. Wherein, the slight movement of the head of the experimenter can not influence the projection, and the loss of the image caused by blinking can be quickly recovered.
The movement locus of the eye movement fixation point when the experimenter watches the video is collected by using the expert Center software. The sight of the eyes of an experimenter needs to be calibrated, after calibration is finished, calibration feedback, namely deviation in the X, Y direction, can occur, when the deviation X and the deviation Y are smaller than 1.0, an experiment can be carried out, and then model video playing can occur;
and finally, synthesizing the material video and the eye movement fixation point tracking track by using BeGaze analysis software to obtain the model video with the experimenter eye movement fixation point track.
And (3) cutting the model video with the experimenter eye movement fixation point track acquired by the eye tracker according to frames, acquiring eye movement fixation point coordinates on each picture by a moving point extraction algorithm, and converting two-dimensional coordinates of a moving point on the picture into three-dimensional coordinates on a space by a three-dimensional mapping algorithm.
Firstly, extracting coordinates of a two-dimensional eye movement fixation point, and utilizing FFmpeg built software to cut a synthesized video into pictures according to frames to obtain two-dimensional pictures; and acquiring the color of the eye movement fixation point in the picture, wherein the movement track of the eye movement fixation point is embodied in a mode that the eye movement fixation point moves on the model, and the eye movement fixation point is orange through software setting in the last step. We call it the moving point color here and set the color tolerance value; then setting a two-dimensional array, initializing the two-dimensional array to set the value to be 1, comparing the two-dimensional array with a two-dimensional picture, and if the two-dimensional array is compared with the moving point color, obtaining three RGB values which are all in a tolerance range and meet the condition; otherwise, the point is not the eye movement point we are looking for. If no point satisfies the condition, the point of interest flag is set to 0, otherwise to 1. And (3) putting the points meeting the conditions into a point array, respectively recording the maximum and minimum row and column of the points meeting the conditions, taking the median of the maximum and minimum row and column coordinates, namely the coordinate (row, rank) of the interest point, and outputting the coordinate point of the interest point.
And then obtaining a three-dimensional model with an eye movement fixation point through a three-dimensional mapping algorithm. Because the two-dimensional picture of the model is obtained by projection on the three-dimensional XOY plane, we can see from the nature of the projection that the two-dimensional and three-dimensional coordinate points are linked to each other. The three-dimensional coordinates (x, y, z) are determined by first projecting the model onto the XOY plane so that x, y coordinate values are determined and the z coordinate is taken of the vertex of the model surface closest to the viewpoint and the z coordinate value of the vertex on the three-dimensional model is the z coordinate value of the eye gaze point on the three-dimensional model, and if the vertex is not present on the model surface, selecting the coordinate value of the model vertex at a distance within a threshold range as the coordinate value of the eye gaze point. Further, since the model is rotated around the coordinate axis when preparing the picture material, it is now required to be rotated in the reverse direction by a corresponding angle. And finally, obtaining the data of the interest points collected by the eye tracker.
And (4) correspondingly analyzing the experimental data result of the experimenter, sorting all the extracted eye movement fixation points to obtain a database of an interest point set, discarding some inappropriate and abnormal data, and appropriately merging the interest points, wherein the integration method comprises the following steps:
when constructing an evaluation library that evaluates the point of interest operator, two criteria are selected, one being the radius of the region of interest and the other being the number of action points within the region. The radius of the region of interest is set to σ dM, where dM represents the model diameter, i.e., the maximum euclidean distance between all vertices of the model M, and σ represents a constant coefficient. All interest points with a measurement distance of less than 2 σ dM are divided into the same region, and if the data volume of different experimenters in the region is less than n, the interest points in the region are discarded. Selecting a point from each region as a representative, and using the point as a standard interest point, wherein the standard is as follows: these points of interest that are selected as criteria need to satisfy the minimum sum of their geometric distances from all other points of interest in the area in which they are located. Note that this is also reasonable if the interest points of the two regions overlap. If the distance between two regions is less than 2 σ dM, the representation with the smaller number of points in the two regions is discarded from the interest point set of the evaluation criterion, and the representation with the larger number of points in the regions is taken as the interest point of the criterion. We denote the point of interest criteria library by the parameters n and σ, i.e., GM(n, σ) represents the point of interest data set for a particular model M. The corresponding values of these two parameters determine the point of interest criteria library. When the value of n is taken to be correspondingly higher, more moving points will fall within the interest area as σ increases, which are considered reasonable because not all volunteers select the details of a certain model as interest points, and σ increases accept more local variations of the labeled points. However, as σ is further increased, it doesOften, the defined regions will contain different regions of interest, so that the closely interesting points marked on different structures start to merge. The average number of point of interest evaluation criteria libraries given will vary with n and σ.
The invention has the following beneficial effects:
the method for establishing the interest point database is more suitable for reconstructing the model, because the eye tracker can be used for dividing the most concerned region part and the non-concerned region part of human eyes on the model. The reconstruction accuracy can be enhanced for the interested places of human beings through the interest point standard library, and the reconstruction accuracy is relatively reduced for the uninterested regions, so that the workload and the storage capacity of model reconstruction can be reduced.
Drawings
FIG. 1 is a database design flow diagram.
Fig. 2 is a video timing diagram.
Fig. 3 is an experimental environment diagram.
FIG. 4 is a flow chart of two-dimensional eye movement extraction.
Fig. 5 is a flow chart of data point integration.
Detailed Description
The invention is further described below with reference to the accompanying drawings.
A design flow of a method for establishing a point of interest database based on an eye movement fixation point moving track is shown in figure 1, and specifically comprises the following 4 steps:
step (1) firstly, a three-dimensional database model needs to be collected, and materials needed by an experiment are manufactured. The specific operation is as follows:
the 24 stanford three-dimensional model libraries and models in the SHREC2007 model database were selected, and these three-dimensional models are widely used in the standard libraries for three-dimensional model research. Storing two groups of data for each three-dimensional model by using MATLAB, wherein one group of data is the three-dimensional model which is sequentially rotated by 60 degrees around an X axis and is respectively rotated by 0 degree, 60 degrees, 120 degrees, 180 degrees, 240 degrees, 300 degrees and 360 degrees; one group is three-dimensional models of which the three-dimensional models rotate by 60 degrees around the Y axis in sequence, and the three-dimensional models rotate by 0 degree, 60 degrees, 120 degrees, 180 degrees, 240 degrees, 300 degrees and 360 degrees respectively; the Z-axis direction is selected as a viewpoint, and the model at 12 angles is projected on the XOY plane, so that 12 two-dimensional projection pictures are obtained by one model. The following are the equations for the rotation of the three-dimensional model and the equations projected onto the XOY plane.
Rotation around the X axis:
z′=zcosθ-xsinθ
x′=xcosθ+zsinθ
y′=y
rotation around the Y axis:
z′=zcosθ-ysinθ
x′=x
y′=ycosθ+zsinθ
formula for parallel projection:
x′=x
y′=y
z′=0
and transforming a picture every 1.5 seconds by using a video editor Movie Maker, synthesizing 12 two-dimensional pictures of each model into a short video, inserting a blank picture of ten seconds between every two models as rest time, and synthesizing 6 models into a long video. The 24 models were made into 4 long videos in total, and the timing diagram is shown in fig. 2.
And (2) putting the video on an eye tracker for an experimenter to watch, and acquiring experimental data by using the eye tracker and corresponding software, as shown in fig. 3.
The specific experimental conditions were as follows:
a) the display device of the image is placed on the left side, and the experimenter sits right in front of the display device of the image and keeps the distance at 70cm, while the experimenter keeps the eyes in accordance with the height of the screen, and can look up the center of the screen.
b) The operating personnel is on the right side, and with computer control video broadcast on display device, the operating personnel is kept apart with the light shield baffle with the experimenter.
c) The experiment interference to the experimenter caused by other light sources is prevented by isolating baffles around the experimenter and isolating curtains around the laboratory.
d) The sound of the environment of the laboratory is kept not more than 30dB, an ideal quiet environment is created, and the experiment interference of other sound sources to an experimenter is prevented.
The eye position of the test person was adjusted using the iViewX software. After the pupil Image of the experimenter appears in the Eye Image frame, an operator needs to adjust the relative position of the screen and the experimenter, so that the pupil Image on the screen can be stably displayed in a centered mode. Wherein, the slight movement of the head of the experimenter can not influence the projection, and the loss of the image caused by blinking can be quickly recovered. The movement locus of the eye movement fixation point when the experimenter watches the video is collected by using the expert Center software. The eye sight of an experimenter needs to be calibrated, after calibration is finished, calibration feedback, namely deviation in the X, Y direction, can occur, when the deviation X and Y are smaller than 1.0, an experiment can be carried out, and a model video is played on a screen; and finally, synthesizing the material video and the eye movement fixation point tracking track by using BeGaze analysis software to obtain the video with the eye movement fixation point track of the experimenter.
And (3) cutting the video with the experimenter eye movement fixation point track acquired by the eye movement instrument according to frames, extracting eye movement fixation point coordinates on each picture, and converting two-dimensional coordinates of the moving point on the picture into three-dimensional coordinates on a space through mapping. The specific operation comprises the following two steps:
1. coordinate extraction of two-dimensional eye movement fixation point
The flow of extracting the eye movement fixation point is shown in fig. 4. Firstly, utilizing FFmpeg built software to cut a synthesized video into pictures according to frames to obtain two-dimensional pictures; the color of the moving point in the picture is obtained, the moving track of the eye movement fixation point is embodied in the mode that the moving point moves on the model, and the moving point is orange through software setting in the last step. We call it the moving point color here and set the color tolerance value; then setting a two-dimensional array, initializing the two-dimensional array to set the value to be 1, comparing the two-dimensional array with a two-dimensional picture, and if the two-dimensional array is compared with the moving point color, obtaining three RGB values which are all in a tolerance range and meet the condition; otherwise, the point is not the eye movement point we are looking for. If no point satisfies the condition, the point of interest flag is set to 0, otherwise to 1. And (3) putting the points meeting the conditions into a point array, respectively recording the maximum and minimum row and column of the points meeting the conditions, taking the median of the maximum and minimum row and column coordinates, namely the coordinate (row, rank) of the interest point, and outputting the coordinate point of the interest point.
2. Three-dimensional mapping
The two-dimensional picture of the model is obtained by projection on a three-dimensional XOY plane, and by the nature of the projection, we can see that two-dimensional and three-dimensional coordinate points are related to each other. The three-dimensional coordinates (x, y, z) are determined by first projecting the model onto the XOY plane so that x, y coordinate values are determined and the z coordinate is taken of the vertex of the model surface closest to the viewpoint and the z coordinate value of the vertex on the three-dimensional model is the z coordinate value of the eye gaze point on the three-dimensional model, and if the vertex is not present on the model surface, selecting the coordinate value of the model vertex at a distance within a threshold range as the coordinate value of the eye gaze point. Further, since the model is rotated around the coordinate axis when preparing the picture material, it is now required to be rotated in the reverse direction by a corresponding angle. And finally, obtaining a data model of the interest point collected by the eye tracker.
And (4) correspondingly analyzing the experimental data result of the experimenter, sorting all the extracted eye movement fixation points to obtain a database of an interest point set, discarding some inappropriate and abnormal data, and appropriately combining the interest points, wherein the integration flow of the data is shown in fig. 5, so that the interest point database based on the eye movement fixation point movement track can be obtained. The specific operation is as follows:
when constructing an evaluation library that evaluates the point of interest operator, two criteria are selected, one being the radius of the region of interest and the other being the number of action points within the region. The radius of the region of interest is set to σ dM, where dM represents the model diameter, i.e., the maximum euclidean distance between all vertices of the model M, and σ represents a constant coefficient. All interest points with a measurement distance of less than 2 σ dM are divided into the same region, and if the data volume of different experimenters in the region is less than n, the interest points in the region are discarded. Selecting a point from each region as a representative by using a criterion, the criterion being a point of interest of the criterionComprises the following steps: these points of interest that are selected as criteria need to satisfy the minimum sum of their geometric distances from all other points of interest in the area in which they are located. Note that this is also reasonable if the interest points of the two regions overlap. If the distance between two regions is less than 2 σ dM, the representation with the smaller number of points in the two regions is discarded from the interest point set of the evaluation criterion, and the representation with the larger number of points in the regions is taken as the interest point of the criterion. We denote the point of interest criteria library by the parameters n and σ, i.e., GM(n, σ) represents the point of interest data set for a particular model M. The corresponding values of these two parameters determine the point of interest criteria library. When the value of n is taken to be correspondingly higher, more moving points will fall within the interest area as σ increases, which are considered reasonable because not all volunteers select the details of a certain model as interest points, and σ increases accept more local variations of the labeled points. However, as σ is further increased, the regions it defines tend to contain different regions of interest, so the closely interesting points marked on different structures begin to merge. The average number of point of interest evaluation criteria libraries given will vary with n and σ.

Claims (6)

1.一种基于眼动注视点移动轨迹的感兴趣点数据库建立方法,其特征在于包括如下步骤:1. a method for establishing a point-of-interest database based on eye movement gaze point movement track, is characterized in that comprising the steps: 步骤(1)收集三维模型并制作成实验所需的视频素材;Step (1) collect the three-dimensional model and make it into the video material required for the experiment; 将三维模型在多个角度上的二维投影图片合成所述视频素材;Synthesizing the video material from two-dimensional projection pictures of the three-dimensional model at multiple angles; 步骤(2)将视频放在眼动仪上让实验者观看,获得眼动注视点的数据,并通过相应软件合成附带眼动注视点的视频;Step (2) put the video on the eye tracker for the experimenter to watch, obtain the data of the eye movement fixation point, and synthesize the video with the eye movement fixation point through the corresponding software; 步骤(3)将附带眼动注视点的视频通过动点提取和三维映射算法,生成附带眼动注视点的三维模型;Step (3) generates a three-dimensional model with eye-movement fixation point by moving point extraction and three-dimensional mapping algorithm with the video with eye-movement fixation point; 将附带眼动注视点的视频剪切,通过动点提取算法获得每张图片上的眼动注视点坐标,再通过三维映射算法将动点在图片上的二维坐标转成空间上的三维坐标;Cut the video with eye movement fixation point, obtain the eye movement fixation point coordinates on each picture through the moving point extraction algorithm, and then convert the two-dimensional coordinates of the moving point on the picture into the three-dimensional coordinates in space through the three-dimensional mapping algorithm ; 步骤(4)对实验者的眼动注视点做分析,对所有实验者的眼动注视点进行整理,得到三维模型感兴趣点集合,舍弃集合中异常的数据,然后对感兴趣点进行合并,建立感兴趣点数据库。Step (4) analyze the eye movement fixation points of the experimenter, sort out the eye movement fixation points of all the experimenters, obtain a set of interest points of the three-dimensional model, discard abnormal data in the collection, and then merge the interest points, Build a database of points of interest. 2.根据权利要求1所述的一种基于眼动注视点移动轨迹的感兴趣点数据库建立方法,其特征在于步骤(1)制作视频素材的方法如下:2. a kind of interest point database establishment method based on eye movement gazing point movement trajectory according to claim 1, it is characterized in that the method for step (1) making video material is as follows: 选择24个斯坦福三维模型库和SHREC2007模型数据库中模型,这些三维模型被广泛用于三维模型研究的标准库;利用MATLAB为每个三维模型保存两组数据,一组是三维模型围绕X轴依次间隔旋转60度:分别是旋转0度,60度,120度,180度,240度,300度,360度的三维模型;一组是三维模型围绕Y轴依次间隔旋转60度:分别是旋转0度,60度,120度,180度,240度,300度,360度的三维模型;选择Z轴方向作为视点,将12个角度上的模型投影在XOY平面,由此一个模型将得到12张二维的投影图片;Select 24 models in Stanford 3D model library and SHREC2007 model database, these 3D models are widely used in the standard library of 3D model research; use MATLAB to save two sets of data for each 3D model, one is the 3D model spaced around the X axis Rotate 60 degrees: 3D models are rotated by 0 degrees, 60 degrees, 120 degrees, 180 degrees, 240 degrees, 300 degrees, and 360 degrees respectively; a group of 3D models are rotated 60 degrees around the Y axis in turn: they are rotated by 0 degrees. , 60 degrees, 120 degrees, 180 degrees, 240 degrees, 300 degrees, 360 degrees three-dimensional models; select the Z-axis direction as the viewpoint, project the models at 12 angles on the XOY plane, and a model will get 12 two-dimensional images projection image; 三维模型旋转的公式以及投影于XOY平面的公式;The formula for the rotation of the 3D model and the formula for the projection on the XOY plane; 围绕X轴旋转:Rotate around the X axis: z′=zcosθ-xsinθz′=zcosθ-xsinθ x′=xcosθ+zsinθx′=xcosθ+zsinθ y′=yy′=y 围绕Y轴旋转:Rotate around the Y axis: z′=zcosθ-ysinθz′=zcosθ-ysinθ x′=xx'=x y′=ycosθ+zsinθy′=ycosθ+zsinθ 平行投影的公式:The formula for parallel projection: x′=xx'=x y′=yy′=y z′=0z′=0 利用视频编辑器Movie Maker,每隔1.5秒变换一张图片,将每个模型的12张二维图片合成一段短视频,并在每两个模型间插入十秒钟空白图片作为休息时间,将6个模型合成一个长视频;24个模型共制成4个长视频。Using the video editor Movie Maker, change a picture every 1.5 seconds, synthesize 12 two-dimensional pictures of each model into a short video, and insert a blank picture for ten seconds between every two models as a rest time. Synthesize a long video; 24 models make a total of 4 long videos. 3.根据权利要求2所述的一种基于眼动注视点移动轨迹的感兴趣点数据库建立方法,其特征在于步骤(2)中让实验者进行实验的前提条件如下:3. a kind of interest point database establishment method based on eye movement gaze point movement trajectory according to claim 2, it is characterized in that in step (2), the precondition that experimenter is allowed to carry out experiment is as follows: a)图像的显示设备放置在左侧,实验者坐在图像的显示设备正前方,并且保持距离为70cm,同时实验者保持眼睛与屏幕高度一致,能够平视屏幕中心;a) The image display device is placed on the left side, and the experimenter sits directly in front of the image display device and maintains a distance of 70cm. At the same time, the experimenter keeps his eyes at the same height as the screen and can look at the center of the screen; b)操作人员在右侧,用电脑控制视频播放在显示设备上,操作人员和实验者之间用隔光挡板隔离;b) The operator is on the right side, and the computer-controlled video is played on the display device, and the operator and the experimenter are isolated by a light barrier; c)实验者的四周需要用隔光挡板隔离,实验室周围使用隔光窗帘,防止其他光源对实验者造成实验干扰;c) The surrounding of the experimenter needs to be isolated by light-shielding baffles, and light-shielding curtains are used around the laboratory to prevent other light sources from causing experimental interference to the experimenter; d)保持实验室环境声音不超过30dB,创造较为理想的安静环境,防止其他声源对实验者造成实验干扰。d) Keep the sound of the laboratory environment no more than 30dB, create an ideal quiet environment, and prevent other sound sources from causing experimental interference to the experimenter. 4.根据权利要求3所述的一种基于眼动注视点移动轨迹的感兴趣点数据库建立方法,其特征在于步骤(2)中具体实验的如下:4. a kind of interest point database establishment method based on eye movement gazing point movement trajectory according to claim 3, is characterized in that in step (2), concrete experiment is as follows: 首先,使用iViewX软件对测试者的眼睛位置进行调整;在Eye Image框出现实验者瞳孔影像后,操作者通过调整屏幕与实验者相对位置,使得屏幕上的瞳孔影像能够居中稳定呈现;First, use the iViewX software to adjust the tester's eye position; after the experimenter's pupil image appears in the Eye Image box, the operator adjusts the relative position of the screen and the experimenter, so that the pupil image on the screen can be centered and presented stably; 然后,利用Experiment Center软件收集实验者观看视频时眼动注视点移动轨迹;收集过程前需要校准实验者眼睛的视线,当校准完后,会出现校准反馈即X、Y方向上的偏差,当偏差X,Y都小于1.0的时候,则可进入实验,之后会出现模型素材视频播放;Then, use the Experiment Center software to collect the movement track of the experimenter's eye movement fixation point when watching the video; before the collection process, the eye of the experimenter needs to be calibrated. After the calibration, there will be calibration feedback, namely the deviation in the X and Y directions. When both X and Y are less than 1.0, you can enter the experiment, and then there will be model material video playback; 最后,利用BeGaze分析软件将模型素材视频和眼动注视点追踪轨迹进行合成,得到附带实验者眼动注视点轨迹的模型视频。Finally, using the BeGaze analysis software to synthesize the model material video and the eye movement fixation track, the model video with the eye movement fixation track of the experimenter is obtained. 5.根据权利要求4所述的一种基于眼动注视点移动轨迹的感兴趣点数据库建立方法,其特征在于步骤(3)先通过将眼动仪获取的附带实验者眼动注视点轨迹的模型视频按帧剪切,通过提取动点算法获得每张图片上的眼动注视点坐标,再通过三维映射算法将动点在图片上的二维坐标转成空间上的三维坐标,具体实现如下:5. a kind of interest point database establishment method based on eye movement fixation point movement trajectory according to claim 4, it is characterized in that step (3) first by the incidental experimenter eye movement fixation point trajectory that eye tracker obtains. The model video is cut by frame, and the eye movement gaze point coordinates on each picture are obtained by extracting the moving point algorithm, and then the three-dimensional mapping algorithm is used to convert the two-dimensional coordinates of the moving point on the picture into three-dimensional coordinates in space. The specific implementation is as follows : 首先,进行二维眼动注视点的坐标提取,利用FFmpeg Build软件将合成的视频按帧截成图片获得二维图片;获取图片中的眼动注视点颜色,眼动注视点移动轨迹通过眼动注视点在模型上移动的形式体现,步骤(2)中通过BeGaze分析软件的设置,将眼动注视点设置为呈橘色,即动点颜色为橘色,并且设置颜色容差值;First, extract the coordinates of the two-dimensional eye-movement fixation point, use FFmpeg Build software to cut the synthesized video into pictures by frame to obtain a two-dimensional picture; The form of the gaze point moving on the model is embodied. In step (2), through the settings of the BeGaze analysis software, the eye movement gaze point is set to be orange, that is, the color of the moving point is orange, and the color tolerance value is set; 接着设置二维数组,并对其进行初始化设值为1;将初始化后的二维数组与二维图片相比较,如果与动点颜色比较,得到的RGB三个值都在颜色容差值范围内,满足条件;否则该点不是要找的眼动点;如果没有点满足条件,则感兴趣点标志置0,反之置1;将满足条件的点放入point数组中,分别需要记录满足条件的点最大最小的行和列,并取最大最小的行和列坐标的中值,即为所求感兴趣点坐标(row,rank),输出感兴趣点坐标点;Then set a two-dimensional array, and initialize it with a value of 1; compare the initialized two-dimensional array with the two-dimensional image, if compared with the moving point color, the three RGB values obtained are all within the color tolerance value range Otherwise, the point is not the eye-tracking point to be found; if there is no point that meets the condition, the point of interest flag is set to 0, otherwise, it is set to 1; put the points that meet the conditions into the point array, and respectively need to record the points that meet the conditions The maximum and minimum rows and columns of the points, and take the median of the maximum and minimum row and column coordinates, that is, the coordinates of the point of interest (row, rank), and output the coordinates of the point of interest; 然后通过三维映射算法获得附带眼动注视点的三维模型;因为模型的二维图片是通过投影在三维的XOY平面上获得的,由投影的性质能够看出二维和三维的坐标点相互之间存在联系:首先使模型投影在XOY平面,因此x,y坐标值是确定的,而z坐标则是取模型表面离视点距离最近的顶点,该顶点在三维模型上的z坐标值就是眼动注视点在三维模型上的z坐标值,如果模型表面上不存在该顶点则选择距离在阈值范围内的模型顶点的坐标值作为眼动注视点的坐标值,如此确定三维坐标(x,y,z);并且因为准备图片素材的时候将模型绕坐标轴进行旋转,所以现在需要将它反向旋转相应的角度;最后获得利用眼动仪收集的感兴趣点的数据。Then, a 3D model with eye-moving fixation points is obtained through a 3D mapping algorithm; because the 2D image of the model is obtained by projecting it on the 3D XOY plane, it can be seen from the nature of the projection that the 2D and 3D coordinate points are between each other There is a connection: first, the model is projected on the XOY plane, so the x and y coordinate values are determined, and the z coordinate is the vertex that is closest to the viewpoint on the model surface. The z coordinate value of this vertex on the 3D model is the eye gaze. The z coordinate value of the point on the 3D model. If the vertex does not exist on the model surface, the coordinate value of the model vertex with a distance within the threshold range is selected as the coordinate value of the gaze point, and the 3D coordinates (x, y, z) are determined in this way. ); and because the model is rotated around the coordinate axis when preparing the picture material, it is now necessary to reversely rotate it by the corresponding angle; finally, the data of the point of interest collected by the eye tracker is obtained. 6.根据权利要求5所述的一种基于眼动注视点移动轨迹的感兴趣点数据库建立方法,其特征在于步骤(4)对实验者的实验数据结果做相应的分析,以此把所有提取的眼动注视点进行整理,得到一个感兴趣点集合的数据库,舍弃异常的数据,并对感兴趣点进行合并,整合方法如下:6. a kind of interest point database establishment method based on eye movement gazing point movement trajectory according to claim 5, it is characterized in that step (4) does corresponding analysis to experimenter's experimental data result, with this all extraction The eye-movement fixation points are sorted out to obtain a database of interest points collection, the abnormal data is discarded, and the interest points are merged. The integration method is as follows: 在构建评估感兴趣点算子的评价库的时候,选择两个标准,一个是感兴趣区域的半径,另一个是在该区域内动点数量;将感兴趣区域的半径设为σdM,其中dM代表模型直径,即模型M的所有顶点之间的最大欧几里得距离,σ代表常系数;将测量距离小于2σdM的所有感兴趣点划分在同一个区域,如果区域中不同的实验者的数据量少于n,则丢弃该区域的感兴趣点;When constructing an evaluation library for evaluating point-of-interest operators, two criteria are selected, one is the radius of the region of interest, and the other is the number of moving points in the region; the radius of the region of interest is set to σdM, where dM Represents the model diameter, that is, the maximum Euclidean distance between all vertices of the model M, and σ represents a constant coefficient; all points of interest with a measurement distance less than 2σdM are divided into the same area, if the data of different experimenters in the area If the amount is less than n, the points of interest in this area are discarded; 从每个区域中以下述的标准选出一个点作为代表,把它作为一个标准的感兴趣点,标准为:被选定为标准的感兴趣点需要满足它与其所在的区域内所有其他的感兴趣点几何距离之和最小;如果两个区域之间的距离小于2σdM,那么两个区域中点数较少的代表将从评价标准的感兴趣点集合中丢弃,将区域中点数较多的代表作为标准的感兴趣点;用参数n和σ表示感兴趣点标准库,即GM(n,σ)表示为特定模型M的感兴趣点数据集;这两个参数对应的值决定了感兴趣点标准库;当n值取得相应较高时,随着σ的增加,将会有更多的动点落在这个兴趣区域内,因为不是所有的实验者都选择某个模型的细节作为感兴趣点,而σ的增加接受更多的标记点的本地局部变化,所以认为σ的变化是合理的;但是随着σ的进一步增加,它所定义的区域往往会包含不同的兴趣区域,因此在不同结构上标记的紧密感兴趣点开始合并;给出的感兴趣点评价标准库的平均数量会随着n和σ的变化而变化的。Select a point from each area as a representative and use it as a standard point of interest. The standard is: the point of interest selected as a standard needs to satisfy all other senses of it and all other areas in the area. The sum of the geometric distances of the interest points is the smallest; if the distance between the two regions is less than 2σdM, the representatives with fewer points in the two regions will be discarded from the set of interest points of the evaluation criteria, and the representatives with more points in the regions will be used as the Standard points of interest; parameters n and σ are used to represent the standard library of points of interest, that is, G M (n,σ) is represented as a data set of points of interest for a specific model M; the corresponding values of these two parameters determine the points of interest Standard library; when the value of n is correspondingly high, as σ increases, more moving points will fall within this region of interest, because not all experimenters choose the details of a model as points of interest , and the increase of σ accepts more local local changes of the marked points, so it is reasonable to think that the change of σ is reasonable; but with the further increase of σ, the area defined by it tends to contain different areas of interest, so it can be found in different structures. Tight points of interest marked above begin to coalesce; the average number of points of interest given for the evaluation criteria library will vary with n and σ.
CN201810116386.0A 2018-02-06 2018-02-06 Interest point database establishing method based on eye movement fixation point moving track Active CN108415955B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810116386.0A CN108415955B (en) 2018-02-06 2018-02-06 Interest point database establishing method based on eye movement fixation point moving track

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810116386.0A CN108415955B (en) 2018-02-06 2018-02-06 Interest point database establishing method based on eye movement fixation point moving track

Publications (2)

Publication Number Publication Date
CN108415955A CN108415955A (en) 2018-08-17
CN108415955B true CN108415955B (en) 2021-12-28

Family

ID=63126834

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810116386.0A Active CN108415955B (en) 2018-02-06 2018-02-06 Interest point database establishing method based on eye movement fixation point moving track

Country Status (1)

Country Link
CN (1) CN108415955B (en)

Families Citing this family (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109255342B (en) * 2018-11-20 2020-07-10 武汉大学 A method and system for extracting regions of interest in images based on two-step clustering of eye-tracking data
CN111382591B (en) * 2018-12-27 2023-09-29 海信集团有限公司 Binocular camera ranging correction method and vehicle-mounted equipment
CN109887583B (en) * 2019-03-11 2020-12-22 数坤(北京)网络科技有限公司 Data acquisition method/system based on doctor behaviors and medical image processing system
CN110032278B (en) * 2019-03-29 2020-07-14 华中科技大学 Method, device and system for pose recognition of objects of interest to human eyes
CN110786868A (en) * 2019-10-18 2020-02-14 天津大学 Non-invasive detection and eye movement analysis method for ASD screening
CN111079740A (en) * 2019-12-02 2020-04-28 咪咕文化科技有限公司 Image quality evaluation method, electronic device, and computer-readable storage medium
CN111510630B (en) * 2020-04-24 2021-09-28 Oppo广东移动通信有限公司 Image processing method, device and storage medium
CN111966225A (en) * 2020-08-31 2020-11-20 东莞市东全智能科技有限公司 Visual data-based multimedia data interaction platform and analysis method
CN114283348A (en) * 2021-04-07 2022-04-05 赢富仪器科技(上海)有限公司 Eye tracker data analysis method, system, device and readable storage medium
CN113158879B (en) * 2021-04-19 2022-06-10 天津大学 3D Gaze Estimation and 3D Eye Movement Model Establishment Based on Matching Features

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102521595A (en) * 2011-12-07 2012-06-27 中南大学 Method for extracting image region of interest based on eye movement data and bottom-layer features
CN205563453U (en) * 2015-12-16 2016-09-07 中国科学院西安光学精密机械研究所 Sight tracking and human eye region-of-interest positioning system
CN107179829A (en) * 2017-05-22 2017-09-19 贵州大学 The method that user individual preference is predicted using eye movement data
CN107506751A (en) * 2017-09-13 2017-12-22 重庆爱威视科技有限公司 Advertisement placement method based on the dynamic control of eye
CN107656613A (en) * 2017-09-08 2018-02-02 国网山东省电力公司电力科学研究院 A kind of man-machine interactive system and its method of work based on the dynamic tracking of eye

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9852543B2 (en) * 2015-03-27 2017-12-26 Snap Inc. Automated three dimensional model generation
US10674139B2 (en) * 2015-06-03 2020-06-02 University Of Connecticut Methods and systems for human action recognition using 3D integral imaging
US9849364B2 (en) * 2016-02-02 2017-12-26 Bao Tran Smart device

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102521595A (en) * 2011-12-07 2012-06-27 中南大学 Method for extracting image region of interest based on eye movement data and bottom-layer features
CN205563453U (en) * 2015-12-16 2016-09-07 中国科学院西安光学精密机械研究所 Sight tracking and human eye region-of-interest positioning system
CN107179829A (en) * 2017-05-22 2017-09-19 贵州大学 The method that user individual preference is predicted using eye movement data
CN107656613A (en) * 2017-09-08 2018-02-02 国网山东省电力公司电力科学研究院 A kind of man-machine interactive system and its method of work based on the dynamic tracking of eye
CN107506751A (en) * 2017-09-13 2017-12-22 重庆爱威视科技有限公司 Advertisement placement method based on the dynamic control of eye

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
"Evaluation of 3D interest point detection techniques via human-generated ground truth";Helin Dutagaci 等;《The Visual Computer》;20120629;第28卷;第901-917页 *
"基于感兴趣区域图像眼动数据库的建立";闫苗苗 等;《西安理工大学学报》;20140331;第30卷(第1期);第85-90页 *
"基于眼动感知信息的彩色图像感兴趣区域提取算法";李双 等;《光学工程》;20110430;第38卷(第4期);第115-123页 *
Wolf Kienzle 等."How to Find Interesting Locations in Video: A Spatiotemporal Interest Point Detector Learned from Human Eye Movements".《Pattern Recognition》.2007, *

Also Published As

Publication number Publication date
CN108415955A (en) 2018-08-17

Similar Documents

Publication Publication Date Title
CN108415955B (en) Interest point database establishing method based on eye movement fixation point moving track
Sitzmann et al. Saliency in VR: How do people explore virtual environments?
Hu et al. Sgaze: A data-driven eye-head coordination model for realtime gaze prediction
US8343067B2 (en) System and method for quantifying and mapping visual salience
US8636361B2 (en) Learning-based visual attention prediction system and method thereof
Kurzhals et al. Space-time visual analytics of eye-tracking data for dynamic stimuli
Ofli et al. Berkeley mhad: A comprehensive multimodal human action database
Coutrot et al. Influence of soundtrack on eye movements during video exploration
CN111427150B (en) Eye movement signal processing method and wearable device for virtual reality head-mounted display
CN109191369A (en) 2D pictures turn method, storage medium and the device of 3D model
CN105931283B (en) A kind of 3-dimensional digital content intelligence production cloud platform based on motion capture big data
Choi et al. “Starkfest” Vision and Clinic Science Special Issue: String Editing Analysis of Human Visual Search
CN112597967A (en) Emotion recognition method and device for immersive virtual environment and multi-modal physiological signals
Panetta et al. Software architecture for automating cognitive science eye-tracking data analysis and object annotation
Cheng et al. Computer-aided autism spectrum disorder diagnosis with behavior signal processing
Perrin et al. EyeTrackUAV2: A large-scale binocular eye-tracking dataset for UAV videos
Krassanakis et al. Detection of moving point symbols on cartographic backgrounds
Malpica et al. Task-dependent visual behavior in immersive environments: A comparative study of free exploration, memory and visual search
Irfan et al. Deepdive: A learning-based approach for virtual camera in immersive contents
Liang A video images-aware knowledge extraction method for intelligent healthcare management of basketball players
Ren et al. Vista: Enhancing long-duration and high-resolution video understanding by video spatiotemporal augmentation
CN113255786B (en) Video quality evaluation method based on electroencephalogram signals and target salient characteristics
CN119600543A (en) A multi-dimensional user data collection and analysis method based on virtual reality scene
EP2112619B1 (en) Video data processing
CN119620855A (en) Trajectory acquisition method, device and system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant