[go: up one dir, main page]

CN106600576A - RGB-D camera-based human head locking method - Google Patents

RGB-D camera-based human head locking method Download PDF

Info

Publication number
CN106600576A
CN106600576A CN201610937575.5A CN201610937575A CN106600576A CN 106600576 A CN106600576 A CN 106600576A CN 201610937575 A CN201610937575 A CN 201610937575A CN 106600576 A CN106600576 A CN 106600576A
Authority
CN
China
Prior art keywords
camera
pixel
local maximum
picture
point
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201610937575.5A
Other languages
Chinese (zh)
Other versions
CN106600576B (en
Inventor
王国强
宋焕生
孙士杰
王韬
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Xiudan Technology Co ltd
Original Assignee
Shanghai Ge Ge Electronic Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Ge Ge Electronic Technology Co Ltd filed Critical Shanghai Ge Ge Electronic Technology Co Ltd
Priority to CN201610937575.5A priority Critical patent/CN106600576B/en
Publication of CN106600576A publication Critical patent/CN106600576A/en
Application granted granted Critical
Publication of CN106600576B publication Critical patent/CN106600576B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20021Dividing image into blocks, subimages or windows
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30242Counting objects in image

Landscapes

  • Engineering & Computer Science (AREA)
  • Quality & Reliability (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Processing (AREA)

Abstract

本发明公开了一种基于RGB‑D相机的人头锁定方法,通过在通道中架设RGB‑D相机,用相机对包含人体目标的通道进行拍摄,获取多幅深度图,并得到深度图对应的俯视图,根据俯视图形成矩形框集合,实现人头的锁定。本发明的方法能够精确的锁定人头。

The invention discloses a head locking method based on an RGB-D camera. By setting up an RGB-D camera in the channel, the camera is used to shoot the channel containing a human body target, obtain multiple depth maps, and obtain a top view corresponding to the depth map , form a set of rectangular frames according to the top view, and realize the locking of the human head. The method of the invention can accurately lock the human head.

Description

Human head locking method based on RGB-D camera
Technical Field
The invention relates to a human head locking method based on an RGB-D camera.
Background
With the development of camera technology, an RGB-D camera appears as a new technology, and in recent years, the RGB-D camera has been widely used as the price thereof is lowered. At present, the RGB-D camera has many implementation principles, such as speckle, TOF and the like, and is gradually and widely applied to various fields, such as three-dimensional reconstruction, image understanding and video monitoring. An advantage of an RGB-D camera is that the distance of the scene to the camera can be directly obtained and then presented to the user in the form of an image (referred to as a depth image or depth image) which is more accurate than the conventional depth image obtained using binoculars. The advantages of an RGB-D camera may provide great convenience for people counting in complex environments.
People counting is one of core contents of video monitoring all the time, and is not well solved for a long time, the main reason is that not only human targets but also other targets exist in scenes, and the targets do not have obvious colors or edge features in some crowded scenes, such as public transportation scenes, so that the targets are often difficult to be segmented by using a traditional RGB camera design algorithm, such as in supermarket channels, except people, bags, carts, purchased articles and the like, false targets (such as bags and carried articles) and human heads do not have obvious features to be distinguished, and the traditional RGB camera is difficult to be accurately locked to people.
Disclosure of Invention
Aiming at the defects in the prior art, the invention aims to provide a human head locking method based on an RGB-D camera, which can accurately lock the human head.
In order to achieve the purpose, the invention adopts the following technical scheme:
a human head locking method based on an RGB-D camera comprises the following steps:
the method comprises the following steps: erecting an RGB-D camera in a channel scene, calibrating the camera, and calculating a parameter matrix of the camera, wherein the channel comprises an A direction and a B direction which are opposite;
step two: continuously shooting a channel containing a human body target by using a camera to obtain N depth maps; obtaining a top view of each depth map; obtaining background image I by using all the obtained top viewsb
Step three: shooting a channel containing a human body target by using a camera to obtain a depth map of m at a certain moment; acquiring a corresponding top view of the depth map; performing background removal operation on a top view to obtain a foreground picture, performing blocking operation on the foreground picture to obtain a blocked picture, searching a local maximum region operation on the blocked picture to obtain a local maximum region set, performing expansion operation on the local maximum region set to obtain an expanded local maximum region set, and performing filtering rectangular frame processing on the expanded local maximum region set to obtain a rectangular frame set S containing a plurality of elementsFmThe purpose of locking the human head is achieved.
Specifically, in the second step and the third step, the top view of each depth map is obtained by the following formula:
len=m*r
where θ is the distance P (x) passing through the depth mapp,yp,zp) Of dotsCorresponding to the included angle between the ray and the ground plane; g (x)G,yG0) is the intersection point of the oblique line passing through the point P and the ground plane; hCIs the camera height; m (0 < m < D) is the depth value of the P point in the depth map, wherein D is the maximum pixel value set by a user; r is the distance in world space corresponding to the unit depth value;
plan view I is obtained using the following formula:
wherein, (u, v) represents a pixel point in the top view I corresponding to the point P on the depth map, and I (u, v) represents a pixel value at the pixel point (u, v);
and aiming at each point in the depth map, obtaining a pixel point in the top view corresponding to the point and a pixel value at the pixel point, wherein all the pixel values form a top view I.
Specifically, in the third step, a background removal operation is performed on the top view to obtain a foreground picture, and an adopted formula is as follows:
wherein,Fthreshold value set for user for extracting foreground, IF(u, v) denotes the foreground Picture IFPixel value, I, at the middle pixel point (u, v)b(u, v) is background plot IbPixel value at pixel point (u, v) position, Im(u, v) shows the plan view ImPixel value at pixel point (u, v).
Specifically, in the third step, a blocking operation is performed on the foreground picture to obtain a blocked picture, and an adopted formula is as follows:
wherein, IF(u, v) is the foreground picture IFPixel value with coordinates (u, v), IB(x, y) is Picture IBThe size of the delineated block is w for the pixel value at the location of the pixel point (x, y)b×wb
Specifically, the step three includes performing an operation of finding a local maximum region for a picture after being partitioned to obtain a local maximum region set, and specifically includes the following steps:
for picture IBThe above pixel point (x, y) is searched for eight pixel points around the pixel point, if the pixel value corresponding to the pixel point is larger than the pixel values corresponding to the eight pixel points, the pixel point is put into a local maximum area set SLIn (1) by SL (i)Denotes SLIs a member of, and SL (i)=(ui,vi,di),(ui,vi) Represents the pixel point, diIs a pixel point (u)i,vi) In picture IBThe pixel value of (1).
Specifically, the expanding the local maximum region operation on the local maximum region set in the third step to obtain an expanded local maximum region set specifically includes the following steps:
for a local maximum region set SLEach element S ofL (i)Looking for SL (i)In the foreground picture IFThe formula adopted by the corresponding pixel position in (1) is as follows:
wherein (x)i,yi) Is SL (i)Corresponding to the foreground picture IFThe position of (1); order SS (i)=(xi,yi,zi),(xi,yi) Denotes SL (i)Corresponding to the foreground picture IFTo obtain a set SS,SS (i)Is a set SSAn element of (1);
for SSEach member S inS (i)=(xi,yi,zi) With SS (i)For the seeds, a seed filling method is utilized, outward expansion is carried out, and the conditions of the expansion are as follows: if IF(xi,yi)-zi|≤EThen use a rectangular frame SE (i)=(ui,vi,Hi,Wi,zi) All pixel points satisfying the condition in the frame selection, wherein (u)i,vi) Is the upper left corner of the rectangular frame, (H)i,Wi) Is the height and width of a rectangular frame, ziAs a result of the original pixel values,Eforming a set S of expanded regions for a specified thresholdE,SE (i)Is a set SEOf (2) is used.
Specifically, the step three includes performing rectangular frame filtering processing on the expanded local maximum region set to obtain a rectangular frame set including multiple elements, and includes the following steps:
using two sets S of filter condition pairsEFiltering the elements in (1):
(1) if the element SE (i)The following conditions are met:then the element is deleted;
(2) if two rectangular frames SE (i)=(ui,vi,Hi,Wi,zi) And SE (j)=(uj,vj,Hj,Wj,zj) Satisfy the following requirementsThen determine SE (i)And SE (j)Coincidence, if coincident, then z is retainediAnd zjA larger rectangular frame;
forming the reserved rectangular frames into a rectangular frame set SFmSet of rectangular frames SFmThe element in (A) is SFm (i)Where m represents time.
Compared with the prior art, the invention has the following technical effects: according to the method, the RGB-D camera is erected in the channel, the channel containing the human body target is shot by the camera, a plurality of depth maps are obtained, the top view corresponding to the depth maps is obtained, and the rectangular frame set is formed according to the top view.
The embodiments of the invention will be explained and explained in further detail with reference to the figures and the detailed description.
Drawings
FIG. 1 is a scene model without a coordinate system;
FIG. 2 is a channel model of the world coordinate system;
FIG. 3 is a schematic diagram of a top view image blocking operation;
FIG. 4 is a schematic illustration of finding a local maximum; wherein, (a) represents the picture area in which the maximum value is sought, (b) represents the process of finding the local maximum value, and (c) represents the final finding of the local maximum value;
FIG. 5 is a schematic view of a camera mounting location;
FIG. 6 is a diagram illustrating the selection of six sets of world coordinates and their corresponding image coordinates;
FIG. 7 is a schematic diagram of a top view from a depth map; the method comprises the following steps of (a) obtaining a background image of a channel scene, (b) obtaining a depth image, (c) obtaining a foreground image through background removing operation, and (d) obtaining a top view;
FIG. 8 is a schematic diagram of a filtered set of rectangular boxes taken from a top view; the method comprises the following steps of (a) representing a blocking operation result graph, (b) representing a local maximum area set, (c) extending a rectangular frame set after a local maximum area, and (d) filtering the rectangular frame set after rectangular frame processing.
Detailed Description
The invention discloses a human head locking method based on an RGB camera, which comprises the following steps:
erecting an RGB-D camera in a channel scene, calibrating the camera, and calculating a parameter matrix P of the camera;
step 1.1: selecting a certain channel as a scene for people counting, referring to fig. 1, installing a camera right above the channel, and enabling a plurality of human body targets to walk on the channel along the direction A or the direction B, wherein the direction A is opposite to the direction B;
step 1.2: and establishing a world coordinate system. Referring to fig. 2, the camera is located on the Z-axis of the world coordinate system, the direction along the channel is the Y-axis direction of the world coordinate system, the direction perpendicular to the channel is the X-axis direction of the world coordinate system, and the position coordinate of the camera in the world coordinate system is (0,0, H), where H is the distance of the camera from the origin of the world coordinate system.
Step 1.3: and calibrating the camera. Using a calibration support, selecting N (N is more than or equal to 6) groups of image coordinates and world coordinates corresponding to the image coordinates:
the parameter matrix P of the camera is calculated using the following formula:
wherein,
step two: continuously shooting a channel containing a human body target by using a camera to obtain N (N is more than or equal to 50) depth maps; obtaining a top view of each depth map; obtaining a background image I from a top viewb
The method for obtaining the top view of each depth map comprises the following steps:
the depth values in the depth map represent points in the world coordinate space, such as the distance len from a point P to the camera, i.e. the length of the hypotenuse of the small right triangle in the map, and we can obtain the following formula according to the geometric relationship of the objects under the world coordinate system:
len=m*r (4)
wherein, theta is an included angle between a corresponding ray passing through the point P on the depth map and the ground plane; g (x)G,yG0) is the intersection point of the oblique line passing through the point P and the ground plane; hCIs the camera height; m (0 < m < D) is the depth value of the P point in the depth map, wherein D is set by a userA maximum pixel value; and r is the distance in world space corresponding to the unit depth value.
After obtaining the coordinates of the point P, zooming and translating the point P to be located at the center of the top view I, then:
wherein, (u, v) represents a pixel point in the top view I corresponding to the point P, and I (u, v) represents a pixel value at the pixel point (u, v), where (r)x,ry) To point P of (x)p,yp) Scaling factor of (d)x,dy) To point P of (x)p,yp) And (4) translation coefficient.
And aiming at each point in the depth map, obtaining a pixel point in the top view corresponding to the point and a pixel value of the pixel point, wherein all the pixel values form a top view I. N top views I can be obtained by adopting the method aiming at N depth mapsi(i=1,...N)。
Wherein, a background image I is obtained by using a top viewbThe formula adopted is as follows;
wherein H is the length of the top view, W is the width of the top view, Ib(x, y) is background picture IbThe pixel value at the position of the pixel point (x, y) can be used to obtain the background image Ib
Step three: shooting a channel containing a human body target by using a camera to obtain a depth map at a certain moment; acquiring a corresponding top view of the depth map; background removal, blocking, local maximum area searching, local maximum area expanding and rectangular frame filtering processing are carried out on the top view, and a rectangular frame set S is obtainedFm(ii) a The method specifically comprises the following steps:
step 3.1: the method comprises the steps that a camera is used for shooting a channel containing a human body target, an RGB-D camera is adopted, and a depth map of a certain moment m (m is 1, 2.);
step 3.2, acquiring a corresponding top view I of the shot depth mapmAnd the adopted method is the same as the method for acquiring the top view in the second step.
Step 3.3, for top view I obtained in step 3.2mPerforming background removal, blocking, local maximum area searching, local maximum area expanding and rectangular frame filtering processing to obtain a rectangular frame set SFmThe specific treatment process is as follows:
removing the background: for top view ImObtaining the foreground picture I by adopting a formula (8)F
Wherein,Fthreshold value set for user for extracting foreground, IF(u, v) denotes the foreground Picture IFPixel value at the middle pixel point (u, v).
And (3) blocking operation: with a size wb×wbBlock pair foreground picture IFBlocking to obtain picture IBThe formula adopted is as follows:
wherein, IF(u, v) is the foreground picture IFPixel value with coordinates (u, v), IB(x, y) is Picture IBPixel value at pixel point (x, y) location.
Finding the local maximum area: for picture IBThe (x, y) of the above point, and the (x, y) of the point around the point is searchedIf the pixel value corresponding to the pixel point is larger than the pixel values corresponding to the eight pixel points, the pixel point is placed into a local maximum area set SLIn, adopt SL (i)Denotes SLAnd S isL (i)=(ui,vi,di),(ui,vi) Represents the pixel point, diIs a pixel point (u)i,vi) In picture IBThe pixel value of (1).
Expanding the local maximum area: for a local maximum region set SLEach element S ofL (i)Looking for SL (i)In the foreground picture IFThe formula adopted by the corresponding pixel position in (1) is as follows:
wherein (x)i,yi) Is SL (i)Corresponding to the foreground picture IFOf (c) is used. Order SS (i)=(xi,yi,zi),(xi,yi) Denotes SL (i)Corresponding to the foreground picture IFThe pixel points of (2) can obtain a set SS,SS (i)Is a set SSOf (2) is used.
For SSEach member S inS (i)=(xi,yi,zi) With SS (i)For the seeds, a seed filling method is utilized, outward expansion is carried out, and the conditions of the expansion are as follows: if IF(xi,yi)-zi|≤EETo set the threshold value to 10, a rectangular frame S is usedE (i)=(ui,vi,Hi,Wi,zi) All pixel points satisfying the condition in the frame selection, wherein (u)i,vi) Is the upper left corner of the rectangular frame, (H)i,Wi) Is a rectangular frameHeight and width of (z)iFor the original pixel values (i.e. the spatial height of the rectangular box), a set S of expanded regions is finally formedE,SE (i)Is a set SEOf (2) is used.
And (3) filtering a rectangular frame: after the extended area is obtained, the overlap area and the abnormal area need to be filtered, and two filtering conditions are used, 1. if the rectangle frame SE (i)The following conditions are met:then not reserved; 2. if two rectangular frames SE (i)=(ui,vi,Hi,Wi,zi) And SE (j)=(uj,vj,Hj,Wj,zj) Satisfy the following requirementsThen determine SE (i)And SE (j)Coincidence, if coincident, then z is retainediAnd zjA larger rectangular frame.
The remaining rectangular frames form a set S of rectangular framesFmSet of rectangular frames SFmThe element in (A) is SFm (i)And completing the human head locking task.
Examples
In the processing process of the embodiment, the sampling frequency is 25 frames/second, the size of the frame image is 320 × 240, and the scene is a front door scene of a bus.
The camera is mounted on the bus at the position shown in fig. 5, and a world coordinate system is established, and the height H of the camera is 254 (cm). And 6 groups of points of world coordinates corresponding to the images are selected by using the calibration frame, and parameters P of the camera are calculated according to the figure 6.
Here is selectedb10, as shown in fig. 7, (a) is a background image of the channel scene, and (b) is obtainedDepth map, (c) foreground picture obtained by background removing operation, and (d) top view.
As shown in FIG. 8, select BS×BS=5×5,h10(a) a partitioning operation result graph, wherein a white rectangular frame in (b) is a local maximum region set, a white rectangular frame in (c) is an expanded local maximum region, and a white rectangular frame in (d) is a set after filtering rectangular frame processing.

Claims (7)

1.一种基于RGB-D相机的人头锁定方法,其特征在于,包括以下步骤:1. A head locking method based on RGB-D camera, is characterized in that, comprises the following steps: 步骤一:在通道场景中架设RGB-D相机,对相机进行标定,计算相机的参数矩阵,通道包括A方向和B方向,二者方向相反;Step 1: Set up an RGB-D camera in the channel scene, calibrate the camera, and calculate the parameter matrix of the camera. The channel includes the A direction and the B direction, and the two directions are opposite; 步骤二:利用相机对包含人体目标的通道进行连续拍摄,获取N幅深度图;求取每幅深度图的俯视图;利用求取的所有俯视图求取背景图IbStep 2: Use the camera to continuously shoot the channel containing the human target to obtain N depth maps; obtain the top view of each depth map; use all the obtained top views to obtain the background image I b ; 步骤三:利用相机对包含人体目标的通道进行拍摄,获取某一时刻m的深度图;针对该幅深度图获取其对应的俯视图;针对俯视图进行去背景操作得到前景图片,针对前景图片进行分块操作得到分块后的图片,针对分块后的图片进行寻找局部最大区域操作得到局部最大区域集合,针对局部最大区域集合进行扩展局部最大区域操作得到扩展后的局部最大区域集合,针对扩展后的局部最大区域集合进行过滤矩形框处理,得到包含有多个元素的一个矩形框集合SFm,实现人头锁定的目的。Step 3: Use the camera to shoot the channel containing the human target to obtain a depth map of m at a certain moment; obtain its corresponding top view for the depth map; perform background removal operations on the top view to obtain a foreground image, and divide the foreground image into blocks The operation is to obtain the divided picture, and the operation of finding the local maximum area is performed on the divided picture to obtain the local maximum area set, and the local maximum area set is extended to obtain the expanded local maximum area set, and for the expanded local maximum area set The local maximum area set is processed by filtering the rectangular frame, and a rectangular frame set S Fm containing multiple elements is obtained to achieve the purpose of head locking. 2.如权利要求1所述的基于RGB-D相机的人数统计方法,其特征在于,步骤二和步骤三中的求取每幅深度图的俯视图,采用的公式如下:2. the people counting method based on RGB-D camera as claimed in claim 1, is characterized in that, in step 2 and step 3, obtain the top view of every depth map, the formula that adopts is as follows: xx GG ythe y GG == pp 1111 -- pp 3131 xx pp 1212 -- pp 3232 xx pp 21twenty one -- pp 3131 ythe y pp 22twenty two -- pp 3232 ythe y -- 11 pp 3434 xx -- pp 1414 pp 3434 ythe y -- pp 24twenty four sthe s ii nno (( &theta;&theta; )) == Hh cc // xx GG 22 ++ ythe y GG 22 ++ Hh cc 22 len=m*rlen=m*r zz PP == Hh cc -- ll ee nno ** sthe s ii nno (( &theta;&theta; )) xx pp == xx GG (( 11 -- zz pp Hh CC )) ythe y pp == ythe y GG (( 11 -- zz pp Hh CC )) 其中,θ为深度图上经过P(xp,yp,zp)点的对应射线与地平面的夹角;G(xG,yG,0)为过P点的斜线与地平面的交点;HC为相机高度;m(0<m<D)为P点在深度图中的深度值,其中D为用户设定的最大像素值;r为单位深度值所对应的世界空间中的距离;Among them, θ is the angle between the corresponding ray passing through point P(x p ,y p ,z p ) on the depth map and the ground plane; G(x G ,y G ,0) is the angle between the oblique line passing through point P and the ground plane H C is the height of the camera; m (0<m<D) is the depth value of point P in the depth map, where D is the maximum pixel value set by the user; r is the world space corresponding to the unit depth value distance; 利用以下公式得到俯视图I:Use the following formula to get the top view I: uu == rr xx xx pp ++ dd xx vv == rr ythe y ythe y pp ++ dd ythe y II (( uu ,, vv )) == zz pp 其中,(u,v)表示深度图上的点P对应的俯视图I中的像素点,I(u,v)表示像素点(u,v)处的像素值;Among them, (u, v) represents the pixel point in the top view I corresponding to the point P on the depth map, and I(u, v) represents the pixel value at the pixel point (u, v); 针对深度图中的每一个点,得到该点对应的俯视图中的像素点和该像素点处的像素值,所有的像素值形成俯视图I。For each point in the depth map, the pixel point in the top view corresponding to the point and the pixel value at the pixel point are obtained, and all pixel values form the top view I. 3.如权利要求2所述的基于RGB-D相机的人数统计方法,其特征在于,所述步骤三中的针对俯视图进行去背景操作得到前景图片,采用的公式如下:3. the people counting method based on RGB-D camera as claimed in claim 2, is characterized in that, in described step 3, carry out background removal operation to obtain foreground picture for overhead view, the formula that adopts is as follows: 其中,δF为用户设定的用于提取前景的阈值,IF(u,v)表示前景图片IF中像素点(u,v)处的像素值,Ib(u,v)为背景图Ib在像素点(u,v)位置处的像素值,Im(u,v)表示俯视图Im像素点(u,v)处的像素值。Among them, δ F is the threshold value set by the user for extracting the foreground, I F (u, v) represents the pixel value at the pixel point (u, v) in the foreground picture I F , and I b (u, v) is the background The pixel value at the pixel point (u, v) in the image I b , and Im (u, v) represents the pixel value at the pixel point (u, v) in the top view Im . 4.如权利要求3所述的基于RGB-D相机的人数统计方法,其特征在于,所述步骤三中的针对前景图片进行分块操作得到分块后的图片,采用的公式如下:4. the people counting method based on RGB-D camera as claimed in claim 3, is characterized in that, in described step 3, carry out block operation at foreground picture and obtain the picture after block, the formula that adopts is as follows: II BB (( xx ,, ythe y )) == &Sigma;&Sigma; uu == ww bb xx ww bb xx ++ ww bb &Sigma;&Sigma; vv == ww bb ythe y ww bb ythe y ++ ww bb II Ff (( uu ,, vv )) ww bb 22 其中,IF(u,v)为前景图片IF坐标为(u,v)的像素值,IB(x,y)为图片IB在像素点(x,y)位置处的像素值,划定的块的大小为wb×wbWherein, I F (u, v) is the pixel value of the foreground picture I F coordinate (u, v), and I B (x, y) is the pixel value of the picture I B at the pixel point (x, y) position, The size of the defined block is w b ×w b . 5.如权利要求4所述的基于RGB-D相机的人数统计方法,其特征在于,所述步骤三中的针对分块后的图片进行寻找局部最大区域操作得到局部最大区域集合,具体包括以下步骤:5. the people counting method based on RGB-D camera as claimed in claim 4, it is characterized in that, the picture in described step 3 is searched for local maximum area operation and obtains local maximum area set for the picture after the block, specifically comprises the following step: 针对图片IB上的像素点(x,y),查找该像素点周围的的八个像素点,如果该像素点对应的像素值比八个像素点对应的像素值都要大,将该像素点放入局部最大区域集合SL中,利用SL (i)表示SL的成员,且SL (i)=(ui,vi,di),(ui,vi)表示该像素点,di为像素点(ui,vi)在图片IB中的像素值。For the pixel point (x, y) on the picture I B , find the eight pixel points around the pixel point, if the pixel value corresponding to the pixel point is larger than the pixel value corresponding to the eight pixel points, the pixel point Points are placed in the local maximum area set SL , and S L (i) is used to represent the members of S L , and S L (i) = (u i , v i , d i ), (u i , v i ) represents the pixel, d i is the pixel value of the pixel (u i , v i ) in the picture I B. 6.如权利要求5所述的基于RGB-D相机的人数统计方法,其特征在于,所述步骤三中的针对局部最大区域集合进行扩展局部最大区域操作得到扩展后的局部最大区域集合,具体包括以下步骤:6. the people counting method based on RGB-D camera as claimed in claim 5, it is characterized in that, in described step 3, carry out expansion local maximum area operation for local maximum area set to obtain expanded local maximum area set, specifically Include the following steps: 针对局部最大区域集合SL的每个元素SL (i),寻找SL (i)在前景图片IF中对应的像素位置,采用的公式为:For each element SL (i) of the local maximum area set SL , find the corresponding pixel position of SL (i) in the foreground picture I F , the formula adopted is: xx ii == uu ii ww bb ++ ww bb 22 ythe y ii == vv ii ww bb ++ ww bb 22 zz ii == dd ii 其中,(xi,yi)是SL (i)对应于前景图片IF中的位置;令SS (i)=(xi,yi,zi),(xi,yi)表示SL (i)对应于前景图片IF的像素点,得到集合SS,SS (i)为集合SS的元素;Among them, ( xi , y i ) is the position of S L (i) corresponding to the foreground picture I F ; let S S (i) = ( xi , y i , z i ), ( xi , y i ) Represent S L (i) corresponding to the pixel point of foreground picture I F , obtain set S S , S S (i) is the element of set S S ; 针对SS中的每个成员SS (i)=(xi,yi,zi),以SS (i)为种子,利用种子填充法,向外扩展,扩展的条件为:若|IF(xi,yi)-zi|≤δE,则使用一个矩形框SE (i)=(ui,vi,Hi,Wi,zi)框选中所有满足条件的像素点,其中(ui,vi)为矩形框左上角点,(Hi,Wi)为矩形框的高和宽,zi为原始像素值,δE为规定的阈值,形成一个扩展后区域的集合SE,SE (i)为集合SE的元素。For each member S S ( i) = ( xi , y i , zi ) in S S, use S S (i) as the seed, and use the seed filling method to expand outward. The expansion condition is: if| I F (x i ,y i )-z i |≤δ E , then use a rectangular box S E ( i ) = (u i ,v i ,H i ,W i ,zi ) to select all the Pixels, where (u i , v i ) is the upper left corner of the rectangular frame, (H i , W i ) is the height and width of the rectangular frame, z i is the original pixel value, and δ E is the specified threshold, forming an extended The set S E of the back region, S E (i) is an element of the set S E . 7.如权利要求6所述的基于RGB-D相机的人数统计方法,其特征在于,所述步骤三中的针对扩展后的局部最大区域集合进行过滤矩形框处理,得到包含有多个元素的矩形框集合,包括以下步骤:7. the people counting method based on RGB-D camera as claimed in claim 6, it is characterized in that, in described step 3, carry out filter rectangular frame processing for the local maximum area collection after expansion, obtain the multi-element A collection of rectangular boxes, including the following steps: 采用两个过滤条件对集合SE中的元素进行过滤:Use two filter conditions to filter the elements in the set S E : (1)若元素SE (i)符合以下条件:则将该元素删除;(1) If the element S E (i) meets the following conditions: then delete the element; (2)若两个矩形框SE (i)=(ui,vi,Hi,Wi,zi)和SE (j)=(uj,vj,Hj,Wj,zj),满足则判定SE (i)和SE (j)重合,如果重合,则保留zi和zj较大的矩形框;(2) If two rectangular frames S E (i) = (u i , v i , H i , W i , z i ) and S E (j) = (u j , v j , H j , W j , z j ), satisfy Then it is determined that S E (i) and S E (j) are coincident, and if they coincide, the rectangular frame with larger z i and z j is reserved; 将保留下的矩形框形成矩形框集合SFm,矩形框集合SFm中的元素为SFm (i),其中,m表示时刻。The remaining rectangular frames are formed into a rectangular frame set S Fm , and the elements in the rectangular frame set S Fm are S Fm (i) , where m represents a moment.
CN201610937575.5A 2016-10-25 2016-10-25 A head locking method based on RGB-D camera Expired - Fee Related CN106600576B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610937575.5A CN106600576B (en) 2016-10-25 2016-10-25 A head locking method based on RGB-D camera

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610937575.5A CN106600576B (en) 2016-10-25 2016-10-25 A head locking method based on RGB-D camera

Publications (2)

Publication Number Publication Date
CN106600576A true CN106600576A (en) 2017-04-26
CN106600576B CN106600576B (en) 2020-02-18

Family

ID=58589489

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610937575.5A Expired - Fee Related CN106600576B (en) 2016-10-25 2016-10-25 A head locking method based on RGB-D camera

Country Status (1)

Country Link
CN (1) CN106600576B (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102542289A (en) * 2011-12-16 2012-07-04 重庆邮电大学 Pedestrian volume statistical method based on plurality of Gaussian counting models
CN104217208A (en) * 2013-06-03 2014-12-17 株式会社理光 Target detection method and device
US20160133025A1 (en) * 2014-11-12 2016-05-12 Ricoh Company, Ltd. Method for detecting crowd density, and method and apparatus for detecting interest degree of crowd in target position
CN106228560A (en) * 2016-07-28 2016-12-14 长安大学 A kind of demographic method under complex scene

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102542289A (en) * 2011-12-16 2012-07-04 重庆邮电大学 Pedestrian volume statistical method based on plurality of Gaussian counting models
CN104217208A (en) * 2013-06-03 2014-12-17 株式会社理光 Target detection method and device
US20160133025A1 (en) * 2014-11-12 2016-05-12 Ricoh Company, Ltd. Method for detecting crowd density, and method and apparatus for detecting interest degree of crowd in target position
CN106228560A (en) * 2016-07-28 2016-12-14 长安大学 A kind of demographic method under complex scene

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
QING LIN等: "Vision-Based Obstacle Avoidance in Sidewalk Environment Using Top-View Transform and Optical-Flow", 《JOURNAL OF MEASUREMENT SCIENCE AND INSTRUMENTATION》 *
唐春晖等: "一种快速的俯视行人检测方法", 《系统仿真学报》 *

Also Published As

Publication number Publication date
CN106600576B (en) 2020-02-18

Similar Documents

Publication Publication Date Title
CN104484668B (en) A kind of contour of building line drawing method of the how overlapping remote sensing image of unmanned plane
CN110221603B (en) A long-distance obstacle detection method based on lidar multi-frame point cloud fusion
JP3513219B2 (en) Corresponding point matching method and apparatus
CN112102409B (en) Target detection method, device, equipment and storage medium
CN108876723B (en) Method for constructing color background of gray target image
US10477178B2 (en) High-speed and tunable scene reconstruction systems and methods using stereo imagery
CN103810685B (en) A kind of super-resolution processing method of depth map
CN104091324A (en) Quick checkerboard image feature matching algorithm based on connected domain segmentation
WO2021017588A1 (en) Fourier spectrum extraction-based image fusion method
CN106340036A (en) Binocular stereoscopic vision-based stereo matching method
CN106600643B (en) A kind of demographic method based on trajectory analysis
CN108171715B (en) Image segmentation method and device
CN105608667A (en) Method and device for panoramic stitching
CN106570883B (en) A People Counting Method Based on RGB-D Camera
CN106599776B (en) A People Counting Method Based on Trajectory Analysis
CN104517317A (en) Three-dimensional reconstruction method of vehicle-borne infrared images
CN116778288A (en) A multi-modal fusion target detection system and method
CN104134194A (en) Image defogging method and image defogging system
CN109544635A (en) It is a kind of based on the automatic camera calibration method for enumerating exploration
CN110910421A (en) Weak and small moving target detection method based on block representation and variable neighborhood clustering
CN109063632A (en) A kind of parking position Feature Selection method based on binocular vision
CN109214288B (en) Inter-frame scene matching method and device based on multi-rotor unmanned aerial vehicle aerial video
Chaloeivoot et al. Building detection from terrestrial images
CN103810692B (en) Video monitoring equipment carries out method and this video monitoring equipment of video tracking
Lv et al. Semantically guided multi-view stereo for dense 3d road mapping

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
TA01 Transfer of patent application right
TA01 Transfer of patent application right

Effective date of registration: 20191224

Address after: Room 501, No. 85, Xiaguang Dongli, Haicang District, Xiamen City, Fujian Province

Applicant after: Xiamen reliable intellectual property service Co.,Ltd.

Address before: 200000 Shanghai City, Pudong New Area Wanxiang town Hongxiang road 83 Lane No. 1-42 building 20 F zone, room 1126

Applicant before: SHANGHAI SAGE ELECTRONIC TECHNOLOGY CO.,LTD.

GR01 Patent grant
GR01 Patent grant
TA01 Transfer of patent application right
TA01 Transfer of patent application right

Effective date of registration: 20200122

Address after: 435, 436, building 1, South Ruifeng garden, guimiao Road, Nanshan street, Nanshan District, Shenzhen City, Guangdong Province

Applicant after: Shenzhen xiudan Technology Co.,Ltd.

Address before: Room 501, No. 85, Xiaguang Dongli, Haicang District, Xiamen City, Fujian Province

Applicant before: Xiamen reliable intellectual property service Co.,Ltd.

CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20200218