[go: up one dir, main page]

WO2019033572A1 - Procédé de détection de situation de visage bloqué, dispositif et support d'informations - Google Patents

Procédé de détection de situation de visage bloqué, dispositif et support d'informations Download PDF

Info

Publication number
WO2019033572A1
WO2019033572A1 PCT/CN2017/108751 CN2017108751W WO2019033572A1 WO 2019033572 A1 WO2019033572 A1 WO 2019033572A1 CN 2017108751 W CN2017108751 W CN 2017108751W WO 2019033572 A1 WO2019033572 A1 WO 2019033572A1
Authority
WO
WIPO (PCT)
Prior art keywords
face
lip
image
eye
model
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/CN2017/108751
Other languages
English (en)
Chinese (zh)
Inventor
陈林
张国辉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Ping An Technology Shenzhen Co Ltd
Original Assignee
Ping An Technology Shenzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ping An Technology Shenzhen Co Ltd filed Critical Ping An Technology Shenzhen Co Ltd
Publication of WO2019033572A1 publication Critical patent/WO2019033572A1/fr
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features

Definitions

  • the present application relates to the field of computer vision processing technologies, and in particular, to a face occlusion detection method, apparatus, and computer readable storage medium.
  • Face recognition is a biometric recognition technology based on human facial feature information for identity authentication.
  • the detected face is matched and recognized by collecting an image or a video stream containing a face and detecting and tracking the face in the image.
  • the application field of face recognition is very extensive, and plays a very important role in many fields such as financial payment, access control, and identification, which brings great convenience to people's lives.
  • the general product in the industry judges that the face occlusion is determined by the deep learning training method, but the judgment method has high requirements on the sample size, and if the occlusion is predicted by the deep learning method, the calculation amount is large and the speed is relatively slow.
  • the present application provides a face occlusion detection method, apparatus, and computer readable storage medium, the main purpose of which is to quickly detect a face occlusion situation in a real-time facial image.
  • the present application provides an electronic device, including: a memory, a processor, and an imaging device, wherein the memory includes a face occlusion detection program, and the face occlusion detection program is executed by the processor Implement the following steps:
  • Image acquisition step acquiring a real-time image captured by the camera device, and extracting a real-time face image from the real-time image by using a face recognition algorithm;
  • a feature point recognition step inputting the real-time facial image into a pre-trained facial average model, and using the facial average model to identify t facial feature points from the real-time facial image;
  • a feature region determining step determining an eye region and a lip region according to position information of the t facial feature points, and inputting the eye region and the lip region into an eye part class model of a pre-trained face, a face
  • the lip part type model determines the authenticity of the eye area and the lip area, and determines whether the face in the real-time image is occluded according to the judgment result.
  • the following steps are further implemented:
  • a determining step determining an eye part class model of the face, a lip part class model of the face Whether the judgment results of the eye area and the lip area are true.
  • the following steps are further implemented:
  • the present application further provides a method for detecting a face occlusion, the method comprising:
  • Image acquisition step acquiring a real-time image captured by the camera device, and extracting a real-time face image from the real-time image by using a face recognition algorithm;
  • a feature point recognition step inputting the real-time facial image into a pre-trained facial average model, and using the facial average model to identify t facial feature points from the real-time facial image;
  • a feature region determining step determining an eye region and a lip region according to position information of the t facial feature points, and inputting the eye region and the lip region into an eye part class model of a pre-trained face, a face
  • the lip part type model determines the authenticity of the eye area and the lip area, and determines whether the face in the real-time image is occluded according to the judgment result.
  • the method further comprises:
  • the determining step is: determining whether the judgment result of the eye part class model and the lip part class model of the human face on the eye region and the lip region is true.
  • the method further comprises:
  • the present application further provides a computer readable storage medium including a face occlusion detection program, where the face occlusion detection program is executed by a processor, as described above Any of the steps in the face occlusion detection method described.
  • the face occlusion detection method, the electronic device and the computer readable storage medium proposed by the present application identify the facial feature points in the real-time facial image by inputting the real-time facial image into the face averaging model, and use the eye part class of the human face
  • the lip model of the model and the face determines the authenticity of the eye region and the lip region determined by the facial feature point, and determines whether the face in the real-time facial image occurs according to the authenticity of the eye region and the lip region. Occlusion.
  • FIG. 1 is a schematic diagram of an application environment of a preferred embodiment of a method for detecting a face occlusion of an applicant
  • FIG. 2 is a block diagram of a face occlusion detection program of FIG. 1;
  • FIG. 3 is a flow chart of a preferred embodiment of the applicant's face occlusion detection method.
  • the present application provides a face occlusion detection method applied to an electronic device 1 .
  • FIG. 1 it is a schematic diagram of an application environment of a preferred embodiment of the applicant's face occlusion detection method.
  • the electronic device 1 may be a terminal device having a computing function, such as a server, a smart phone, a tablet computer, a portable computer, or a desktop computer.
  • a computing function such as a server, a smart phone, a tablet computer, a portable computer, or a desktop computer.
  • the electronic device 1 includes a processor 12, a memory 11, an imaging device 13, a network interface 14, and a communication bus 15.
  • the camera device 13 is installed in a specific place, such as an office place and a monitoring area, and real-time images are taken in real time for the target entering the specific place, and the captured real-time image is transmitted to the processor 12 through the network.
  • Network interface 14 may optionally include a standard wired interface, a wireless interface (such as a WI-FI interface).
  • Communication bus 15 is used to implement connection communication between these components.
  • the memory 11 includes at least one type of readable storage medium.
  • the at least one type of readable storage medium may be a non-volatile storage medium such as a flash memory, a hard disk, a multimedia card, a card type memory, or the like.
  • the readable storage medium may be an internal storage unit of the electronic device 1, such as a hard disk of the electronic device 1.
  • the readable storage medium may also be an external memory of the electronic device 1, such as a plug-in hard disk equipped on the electronic device 1, a smart memory card (SMC), Secure Digital (SD) card, Flash Card, etc.
  • SMC smart memory card
  • SD Secure Digital
  • the readable storage medium of the memory 11 is generally used to store a face occlusion detection program 10 installed on the electronic device 1, a face image sample library, a human eye sample library, and a human lip sample.
  • the memory 11 can also be used to temporarily store data that has been output or is about to be output.
  • the processor 12 may be a Central Processing Unit (CPU), microprocessor or other data processing chip for running program code or processing data stored in the memory 11, such as performing face occlusion Test procedure 10, etc.
  • CPU Central Processing Unit
  • microprocessor or other data processing chip for running program code or processing data stored in the memory 11, such as performing face occlusion Test procedure 10, etc.
  • Figure 1 shows only the electronic device 1 with components 11-15, but it should be understood that not all illustrated components may be implemented, and more or fewer components may be implemented instead.
  • the electronic device 1 may further include a user interface, and the user interface may include an input unit such as a keyboard, a voice input device such as a microphone, and the like.
  • the device for voice recognition function, voice output device such as audio, earphone, etc. optionally the user interface may also include a standard wired interface, a wireless interface.
  • the electronic device 1 may further include a display, which may also be appropriately referred to as a display screen or a display unit.
  • a display may also be appropriately referred to as a display screen or a display unit.
  • it may be an LED display, a liquid crystal display, a touch liquid crystal display, and an Organic Light-Emitting Diode (OLED) touch sensor.
  • the display is used to display information processed in the electronic device 1 and a user interface for displaying visualizations.
  • the electronic device 1 further comprises a touch sensor.
  • the area provided by the touch sensor for the user to perform a touch operation is referred to as a touch area.
  • the touch sensor described herein may be a resistive touch sensor, a capacitive touch sensor, or the like.
  • the touch sensor includes not only a contact type touch sensor but also a proximity type touch sensor or the like.
  • the touch sensor may be a single sensor or a plurality of sensors arranged, for example, in an array.
  • the area of the display of the electronic device 1 may be the same as or different from the area of the touch sensor.
  • a display is stacked with the touch sensor to form a touch display. The device detects a user-triggered touch operation based on a touch screen display.
  • the electronic device 1 may further include a radio frequency (RF) circuit, a sensor, an audio circuit, and the like, and details are not described herein.
  • RF radio frequency
  • an operating system and a face occlusion detecting program 10 may be included in the memory 11 as a computer storage medium; when the processor 12 executes the face occlusion detecting program 10 stored in the memory 11, Implement the following steps:
  • Image acquisition step acquiring a real-time image captured by the camera device 13, and extracting a real-time face image from the real-time image using a face recognition algorithm.
  • the camera 13 When the camera 13 captures a real-time image, the camera 13 transmits the real-time image to the processor 12.
  • the processor 12 receives the real-time image, it first acquires the size of the image to create a grayscale image of the same size. Converting the acquired color image into a grayscale image and creating a memory space; equalizing the grayscale image histogram, reducing the amount of grayscale image information, speeding up the detection speed, and then loading the training library to detect the person in the image Face, and return an object containing face information, obtain the data of the location of the face, and record the number; finally obtain the area of the avatar and save it, thus completing a real-time facial image extraction process.
  • the face recognition algorithm for extracting the real-time facial image from the real-time image may be a geometric feature-based method, a local feature analysis method, a feature face method, an elastic model-based method, a neural network method, or the like.
  • Feature point identification step input the real-time facial image into a pre-trained facial average model, and use the facial average model to identify t facial feature points from the real-time facial image.
  • a first sample library having n face images, and marking t facial feature points in each face image, the t facial feature points including: t 1 eye feature points representing eye positions, t 2 eye feature points and t 3 lip feature points representing the position of the lips.
  • t 1 eye feature points representing eye positions
  • t 2 eye feature points representing eye positions
  • t 3 lip feature points representing the position of the lips.
  • the feature points constitute a shape feature vector S, and the n shape feature vectors S of the face are obtained.
  • the face feature recognition model is trained by using the t facial feature points to obtain a face average model.
  • the face feature recognition model is an Ensemble of Regression Tress (ERT) algorithm.
  • ERT Regression Tress
  • t represents the cascading sequence number
  • ⁇ t ( ⁇ , ⁇ ) represents the regression of the current stage.
  • Each regression is composed of a number of regression trees, and the purpose of training is to obtain these regression trees.
  • each regression ⁇ t ( ⁇ , ⁇ ) predicts an increment based on the input images I and S(t) Add this increment to the current shape estimate to improve the current model.
  • Each level of regression is based on feature points for prediction.
  • the training data set is: (I1, S1), ..., (In, Sn) where I is the input sample image and S is the shape feature vector composed of the feature points in the sample image.
  • the number of face images in the first sample library is t, assuming that each sample image has 34 feature points, the feature vector x 1 to x 12 represent the abscissa of the eyelid feature point, x 13 to x 14 represent the abscissa of the eye feature point, and x 15 to x 34 represent the abscissa of the lip feature point.
  • Take some feature points of all sample images for example, randomly take 25 feature points out of 34 feature points of each sample picture) to train the first regression tree, and predict the first regression tree and the partial features.
  • the residual of the true value of the point (the weighted average of the 25 feature points taken from each sample picture) is used to train the second tree...
  • the face average model is obtained, and the model file and the sample library are saved in the memory. Since the sample of the training model marks 12 eyelid feature points, 2 eyeball feature points, and 20 lip feature points, the face average model of the trained face can be used to identify 12 eyelid feature points from the face image, 2 eye feature points and 20 lip feature points.
  • the trained facial average model is called from the memory 11
  • the real-time facial image is aligned with the facial average model
  • the feature extraction algorithm is used to search the real-time facial image with the facial average model.
  • the 20 lip feature points are evenly distributed on the lip.
  • a feature region determining step determining an eye region and a lip region according to position information of the t facial feature points, the eye region and the eye region and the lip region according to the position information of the t facial feature points, The eye region and the lip region input an eye part class model of a pre-trained face, a lip part class model of the face, determine the authenticity of the eye region and the lip region, and determine the result according to the judgment result Whether the face in the live image is occluded.
  • a first number of human eye positive sample images and a second number of human eye negative sample images are collected, and local features of each human eye positive sample image and human eye negative sample image are extracted.
  • the human eye positive sample image refers to an eye sample containing a human eye, and the binocular portion can be extracted from the face image sample library as an eye sample.
  • the ocular negative eye sample image refers to an image of a broken eye region, and a plurality of human eye positive sample images and negative sample images form a second sample library.
  • a third number of lip positive sample images and a fourth number of lip negative sample images are collected, extracting local features of each lip positive sample image, lip negative sample image.
  • the lip positive sample image refers to an image containing the human's lips, and the lip portion can be extracted from the face image sample library as a lip positive sample image.
  • the lip negative sample image refers to an image of a person's lip region being defective, or the lip of the image is not the lip of a human (eg, an animal), and the plurality of lip positive sample images and the negative sample image form a third sample bank.
  • the local feature is a Histogram of Oriented Gradient (HOG) feature, which is extracted from a human eye sample image and a lip sample image by a feature extraction algorithm. Since the color information in the sample image is not very effective, it is usually converted into a grayscale image, and the entire image is normalized, the gradient of the horizontal and vertical directions of the image is calculated, and the gradient direction of each pixel position is calculated accordingly. Values to capture outlines, silhouettes, and some texture information, and further weaken the effects of lighting. Then the whole image is divided into individual Cell cells (8*8 pixels), and a gradient direction histogram is constructed for each Cell cell to calculate the local image gradient information and quantize to obtain the feature description vector of the local image region.
  • HOG Oriented Gradient
  • the support vector machine is trained by using the positive and negative sample images and the extracted HOG features in the second sample library and the third sample library to obtain the eye part class model of the face and The lip part class model of the face.
  • an eye region may be determined according to the 12 eyelid feature points and the two eyeball feature points, according to The 20 lip feature points define a lip region, and then the determined eye region and the lip region are input into the eye part class model of the trained face and the lip part class model of the face, and are judged according to the result obtained by the model.
  • the determined authenticity of the eye region and the lip region, that is, the result of the model output may be all false or all true, and may include both true and false.
  • the determining step is: determining whether the judgment result of the eye part class model and the lip part class model of the human face on the eye region and the lip region is true. When the eye part class model of the face and the lip part class model of the face output the result, whether the judgment result contains only true.
  • the eye part type model of the face and the lip part type model of the face are true to the eye area and the lip area, it is determined that the face in the real-time face image is not occluded. That is to say, when the eye region and the lip region determined according to the facial feature points are both the human eye region or the human lip region, it is considered that the human face in the real-time facial image is not blocked.
  • the face in the real-time face image is prompted to be occluded.
  • the face in the real-time facial image is considered to be occluded, prompting the real-time The face in the face image is occluded.
  • the eye area in the image is considered to be occluded
  • the lip part class model output result of the face is false
  • the lip in the image is considered The area is occluded and prompted accordingly.
  • the face recognition is performed after detecting whether the face is occluded, when the face in the real-time face image is occluded, the face in the current face image is occluded and re-acquired.
  • the real-time image captured by the imaging device 13 is subjected to subsequent steps.
  • the electronic device 1 of the embodiment extracts a real-time facial image from a real-time image, and uses a facial average model to identify a facial feature point in the real-time facial image, and uses an eye-part model of the human face and a lip of the human face.
  • the classification model analyzes the eye region and the lip region determined by the facial feature points, and quickly determines whether the face in the current image is occluded according to the authenticity of the eye region and the lip region.
  • the face occlusion detection program 10 can also be partitioned into one or more modules that are stored in the memory 11 and executed by the processor 12 to complete the application.
  • a module as referred to in this application refers to a series of computer program instructions that are capable of performing a particular function. Referring to FIG. 2, it is a block diagram of the face occlusion detection program 10 of FIG.
  • the face occlusion detection program 10 can be divided into: an acquisition module 110, an identification module 120, a determination module 130, and a prompting module 140.
  • the functions or operational steps implemented by the modules 110-140 are similar to the above, and are not described in detail herein, by way of example, for example:
  • the acquiring module 110 is configured to acquire a real-time image captured by the camera device 13 and extract a real-time face image from the real-time image by using a face recognition algorithm;
  • the identification module 120 is configured to input the real-time facial image into a pre-trained facial average model, and use the facial average model to identify t facial feature points from the real-time facial image;
  • the determining module 130 is configured to determine an eye region and a lip region according to position information of the t facial feature points, and input the eye region and the lip region into an eye part class model and a face of the pre-trained face a lip part type model for judging the authenticity of the eye area and the lip area, and determining whether the face in the real-time image is occluded according to the judgment result;
  • the prompting module 140 is configured to prompt the real-time facial image when the eye part class model of the human face and the lip part class model of the human face include the unreality of the judgment result of the eye region and the lip region The face is occluded.
  • this application also provides a face occlusion detection method.
  • this application A flowchart of a first embodiment of a face occlusion detection method. The method can be performed by a device that can be implemented by software and/or hardware.
  • the face occlusion detection method includes:
  • Step S10 Acquire a real-time image captured by the camera device, and extract a real-time face image from the real-time image by using a face recognition algorithm.
  • the camera captures a real-time image
  • the camera transmits the real-time image to the processor.
  • the processor When the processor receives the real-time image, first acquires the size of the image to create a grayscale image of the same size; Color image, converted into gray image, and create a memory space; equalize the gray image histogram, reduce the amount of gray image information, speed up the detection, then load the training library, detect the face in the picture, and return An object containing face information, obtains the data of the location of the face, and records the number; finally obtains the area of the avatar and saves it, thus completing a real-time facial image extraction process.
  • the face recognition algorithm for extracting the real-time facial image from the real-time image may be a geometric feature-based method, a local feature analysis method, a feature face method, an elastic model-based method, a neural network method, or the like.
  • Step S20 input the real-time facial image into a pre-trained facial average model, and use the facial average model to identify t facial feature points from the real-time facial image.
  • a first sample library having n face images, and marking t facial feature points in each face image, the t facial feature points including: t 1 eye feature points representing eye positions, t 2 eye feature points and t 3 lip feature points representing the position of the lips.
  • t 1 eye feature points representing eye positions
  • t 2 eye feature points representing eye positions
  • t 3 lip feature points representing the position of the lips.
  • the feature points constitute a shape feature vector S, and the n shape feature vectors S of the face are obtained.
  • the face feature recognition model is trained by using the t facial feature points to obtain a face average model.
  • the face feature recognition model is an ERT algorithm.
  • the ERT algorithm is expressed as follows:
  • t represents the cascading sequence number
  • ⁇ t ( ⁇ , ⁇ ) represents the regression of the current stage.
  • Each regression is composed of a number of regression trees, and the purpose of training is to obtain these regression trees.
  • each regression ⁇ t ( ⁇ , ⁇ ) predicts an increment based on the input images I and S(t) Add this increment to the current shape estimate to improve the current model.
  • Each level of regression is based on feature points for prediction.
  • the training data set is: (I1, S1), ..., (In, Sn) where I is the input sample image and S is the shape feature vector composed of the feature points in the sample image.
  • the number of face images in the first sample library is t, assuming that each sample image has 34 feature points, the feature vector x 1 to x 12 represent the abscissa of the eyelid feature point, x 13 to x 14 represent the abscissa of the eye feature point, and x 15 to x 34 represent the abscissa of the lip feature point.
  • Take some feature points of all sample images for example, randomly take 25 feature points out of 34 feature points of each sample picture) to train the first regression tree, and predict the first regression tree and the partial features.
  • the residual of the true value of the point (the weighted average of the 25 feature points taken from each sample picture) is used to train the second tree...
  • the face average model is obtained, and the model file and the sample library are saved in the memory. Since the sample of the training model marks 12 eyelid feature points, 2 eyeball feature points, and 20 lip feature points, the face average model of the trained face can be used to identify 12 eyelid feature points from the face image, 2 eye feature points and 20 lip feature points.
  • the trained facial average model is called from the memory, and the real-time facial image is aligned with the facial average model, and the feature extraction algorithm is used to search the real-time facial image with the facial average model.
  • Step S30 determining an eye region and a lip region according to position information of the t facial feature points, and inputting the eye region and the lip region into an eye part type model of a pre-trained face, and a lip of the face
  • the classification model determines the authenticity of the eye region and the lip region, and determines whether the face in the real-time image is occluded according to the judgment result.
  • a first number of human eye positive sample images and a second number of human eye negative sample images are collected, and local features of each human eye positive sample image and human eye negative sample image are extracted.
  • the human eye positive sample image refers to an eye sample containing a human eye, and the binocular portion can be extracted from the face image sample library as an eye sample, and the human eye negative eye sample image refers to an image of a broken eye region, and multiple human eye positive samples.
  • the image and the negative sample image form a second sample library.
  • a third number of lip positive sample images and a fourth number of lip negative sample images are collected, extracting local features of each lip positive sample image, lip negative sample image.
  • the lip positive sample image refers to an image containing the human's lips, and the lip portion can be extracted from the face image sample library as a lip positive sample image.
  • the lip negative sample image refers to an image of a person's lip region being defective, or the lip of the image is not the lip of a human (eg, an animal), and the plurality of lip positive sample images and the negative sample image form a third sample bank.
  • the local feature is a direction gradient histogram (HOG) feature, which is extracted from a human eye sample image and a lip sample image by a feature extraction algorithm. Since the color information in the sample image is not very effective, it is usually converted into a grayscale image, and the entire image is normalized, the gradient of the horizontal and vertical directions of the image is calculated, and the gradient direction of each pixel position is calculated accordingly. Values to capture outlines, silhouettes, and some texture information, and further weaken the effects of lighting. Then the whole image is divided into individual Cell cells (8*8 pixels), and a gradient direction histogram is constructed for each Cell cell to calculate the local image gradient information and quantize to obtain the feature description vector of the local image region. Then the cell cells are combined into a large block.
  • HOG direction gradient histogram
  • the support vector machine classifier is trained to obtain the eye part class model of the face and the lip part class model of the face.
  • an eye region may be determined according to the 12 eyelid feature points and the two eyeball feature points, according to The 20 lip feature points define a lip region, and then the determined eye region and the lip region are input into the eye part class model of the trained face and the lip part class model of the face, and are judged according to the result obtained by the model.
  • the determined authenticity of the eye region and the lip region, that is, the result of the model output may be all false or all true, and may include both true and false.
  • step S40 it is determined whether the judgment result of the eye part class model and the lip part class model of the human face on the eye area and the lip area is true.
  • the eye part class model of the face and the lip part class model of the face output the result, whether the judgment result contains only true.
  • Step S50 when the eye part class model of the face and the lip part class model of the face face the determination result of the eye area and the lip area are true, determining that the face in the real-time face image is not Occlusion occurs. That is to say, when the eye region and the lip region determined according to the facial feature points are both the human eye region or the human lip region, it is considered that the human face in the real-time facial image is not blocked.
  • Step S60 when the eye part class model of the face and the lip part class model of the face face the judgment result of the eye area and the lip area to be untrue, prompting the face in the real-time face image to occur Occlusion.
  • the face in the real-time facial image is considered to be occluded, prompting the real-time The face in the face image is occluded.
  • the eye area in the image is considered to be occluded
  • the lip part class model output result of the face is false
  • the lip in the image is considered The area is occluded and prompted accordingly.
  • the step S50 further includes:
  • the facial facial feature model of the facial feature point is used to identify the key facial feature points in the real-time facial image, and the facial part class model of the human face and the lip partial class model of the human face are used.
  • the eye area and the lip area determined by the point are analyzed, and according to the authenticity of the eye area and the lip area, whether the face in the current image is occluded is detected, and the occlusion of the face in the real-time face image is quickly detected.
  • the embodiment of the present application further provides a computer readable storage medium, the computer readable storage medium
  • the storage medium includes a face occlusion detection program, and when the face occlusion detection program is executed by the processor, the following operations are implemented:
  • Image acquisition step acquiring a real-time image captured by the camera device, and extracting a real-time face image from the real-time image by using a face recognition algorithm;
  • a feature point recognition step inputting the real-time facial image into a pre-trained facial average model, and using the facial average model to identify t facial feature points from the real-time facial image;
  • a feature region determining step determining an eye region and a lip region according to position information of the t facial feature points, and inputting the eye region and the lip region into an eye part class model of a pre-trained face, a face
  • the lip part type model determines the authenticity of the eye area and the lip area, and determines whether the face in the real-time image is occluded according to the judgment result.
  • the determining step is: determining whether the judgment result of the eye part class model and the lip part class model of the human face on the eye region and the lip region is true.
  • the training steps of the eye part class model and the lip part class model of the face include:
  • the human eye negative sample image Using the human eye positive sample image, the human eye negative sample image and its local features to train the support vector machine classifier to obtain the eye part class model of the face;
  • the support vector machine classifier is trained by using the lip positive sample image, the lip negative sample image and its local features to obtain the lip part class model of the face.
  • the training step of the facial average model includes:
  • t facial feature points including: t 1 eye feature points representing eye positions, t 2 eye feature points and t 3 lip feature points representing the position of the lips;
  • the face feature recognition model is trained by using the t facial feature points to obtain a face average model.
  • a disk including a number of instructions for causing a terminal device (which may be a mobile phone, a computer, a server, or a network device, etc.) to perform the methods described in the various embodiments of the present application.
  • a terminal device which may be a mobile phone, a computer, a server, or a network device, etc.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

L'invention concerne un procédé permettant de détecter si un visage est bloqué, un dispositif électronique et un support d'informations lisible par ordinateur. Ledit procédé consiste à : acquérir une image en temps réel capturée par un dispositif de caméra, et extraire une image de visage en temps réel à partir de l'image en temps réel (S10) ; entrer l'image de visage en temps réel dans un modèle de visage moyen, et identifier les points de caractéristique de visage dans l'image de visage en temps réel (S20) ; déterminer, en fonction des informations de position des points de caractéristique de visage, une région d'œil et une région de lèvres, entrer la région d'œil et la région de lèvres dans un modèle de classification d'œil préalablement entraîné du visage et un modèle de classification de lèvres du visage (S30), déterminer l'authenticité de la région d'œil et de la région de lèvres (S40), et déterminer, selon un résultat de détermination, si le visage dans l'image en temps réel est bloqué (S50, S60). Le procédé susmentionné peut déterminer rapidement si le visage dans une image de visage est bloqué.
PCT/CN2017/108751 2017-08-17 2017-10-31 Procédé de détection de situation de visage bloqué, dispositif et support d'informations Ceased WO2019033572A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201710707944.6 2017-08-17
CN201710707944.6A CN107633204B (zh) 2017-08-17 2017-08-17 人脸遮挡检测方法、装置及存储介质

Publications (1)

Publication Number Publication Date
WO2019033572A1 true WO2019033572A1 (fr) 2019-02-21

Family

ID=61099639

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/108751 Ceased WO2019033572A1 (fr) 2017-08-17 2017-10-31 Procédé de détection de situation de visage bloqué, dispositif et support d'informations

Country Status (2)

Country Link
CN (1) CN107633204B (fr)
WO (1) WO2019033572A1 (fr)

Cited By (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110119674A (zh) * 2019-03-27 2019-08-13 深圳和而泰家居在线网络科技有限公司 一种作弊检测的方法、装置、计算设备及计算机存储介质
CN110414394A (zh) * 2019-07-16 2019-11-05 公安部第一研究所 一种面部遮挡人脸图像重建方法以及用于人脸遮挡检测的模型
CN110543823A (zh) * 2019-07-30 2019-12-06 平安科技(深圳)有限公司 基于残差网络的行人再识别方法、装置和计算机设备
CN111353404A (zh) * 2020-02-24 2020-06-30 支付宝实验室(新加坡)有限公司 一种人脸识别方法、装置及设备
CN111414879A (zh) * 2020-03-26 2020-07-14 北京字节跳动网络技术有限公司 人脸遮挡程度识别方法、装置、电子设备及可读存储介质
CN111444887A (zh) * 2020-04-30 2020-07-24 北京每日优鲜电子商务有限公司 口罩佩戴的检测方法、装置、存储介质及电子设备
CN111461047A (zh) * 2020-04-10 2020-07-28 北京爱笔科技有限公司 身份识别的方法、装置、设备及计算机存储介质
CN111486961A (zh) * 2020-04-15 2020-08-04 贵州安防工程技术研究中心有限公司 基于宽光谱人体额头成像与距离感知的高效额温估计方法
CN111489373A (zh) * 2020-04-07 2020-08-04 北京工业大学 一种基于深度学习的遮挡物体分割方法
CN111626240A (zh) * 2020-05-29 2020-09-04 歌尔科技有限公司 一种人脸图像识别方法、装置、设备及可读存储介质
CN111639596A (zh) * 2020-05-29 2020-09-08 上海锘科智能科技有限公司 基于注意力机制和残差网络的抗眼镜遮挡人脸识别方法
CN111814571A (zh) * 2020-06-12 2020-10-23 深圳禾思众成科技有限公司 一种基于背景过滤的口罩人脸识别方法及系统
CN111860047A (zh) * 2019-04-26 2020-10-30 美澳视界(厦门)智能科技有限公司 一种基于深度学习的人脸快速识别方法
CN111881740A (zh) * 2020-06-19 2020-11-03 杭州魔点科技有限公司 人脸识别方法、装置、电子设备及介质
CN112016464A (zh) * 2020-08-28 2020-12-01 中移(杭州)信息技术有限公司 检测人脸遮挡的方法、装置、电子设备及存储介质
CN112052730A (zh) * 2020-07-30 2020-12-08 广州市标准化研究院 一种3d动态人像识别监控设备及方法
CN112116525A (zh) * 2020-09-24 2020-12-22 百度在线网络技术(北京)有限公司 换脸识别方法、装置、设备和计算机可读存储介质
CN112597886A (zh) * 2020-12-22 2021-04-02 成都商汤科技有限公司 乘车逃票检测方法及装置、电子设备和存储介质
CN112633183A (zh) * 2020-12-25 2021-04-09 平安银行股份有限公司 影像遮挡区域自动检测方法、装置及存储介质
CN112766214A (zh) * 2021-01-29 2021-05-07 北京字跳网络技术有限公司 一种人脸图像处理方法、装置、设备及存储介质
CN113111817A (zh) * 2021-04-21 2021-07-13 中山大学 语义分割的人脸完整度度量方法、系统、设备及存储介质
CN113449562A (zh) * 2020-03-26 2021-09-28 北京沃东天骏信息技术有限公司 人脸位姿校正方法和装置
CN113449696A (zh) * 2021-08-27 2021-09-28 北京市商汤科技开发有限公司 一种姿态估计方法、装置、计算机设备以及存储介质
CN114255491A (zh) * 2020-09-11 2022-03-29 北京眼神智能科技有限公司 眼睛遮挡判断方法、装置、计算机可读存储介质及设备
CN114399813A (zh) * 2021-12-21 2022-04-26 马上消费金融股份有限公司 人脸遮挡检测方法、模型训练方法、装置及电子设备

Families Citing this family (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108664908A (zh) * 2018-04-27 2018-10-16 深圳爱酷智能科技有限公司 人脸识别方法、设备及计算机可读存储介质
CN110472459B (zh) * 2018-05-11 2022-12-27 华为技术有限公司 提取特征点的方法和装置
CN108551552B (zh) * 2018-05-14 2020-09-01 Oppo广东移动通信有限公司 图像处理方法、装置、存储介质及移动终端
CN108763897A (zh) * 2018-05-22 2018-11-06 平安科技(深圳)有限公司 身份合法性的校验方法、终端设备及介质
CN108985159A (zh) * 2018-06-08 2018-12-11 平安科技(深圳)有限公司 人眼模型训练方法、人眼识别方法、装置、设备及介质
CN110084191B (zh) * 2019-04-26 2024-02-23 广东工业大学 一种眼部遮挡检测方法及系统
CN110348331B (zh) * 2019-06-24 2022-01-14 深圳数联天下智能科技有限公司 人脸识别方法及电子设备
CN110428399B (zh) 2019-07-05 2022-06-14 百度在线网络技术(北京)有限公司 用于检测图像的方法、装置、设备和存储介质
CN112183173B (zh) * 2019-07-05 2024-04-09 北京字节跳动网络技术有限公司 一种图像处理方法、装置及存储介质
CN112929638B (zh) * 2019-12-05 2023-12-15 北京芯海视界三维科技有限公司 眼部定位方法、装置及多视点裸眼3d显示方法、设备
CN111428581B (zh) * 2020-03-05 2023-11-21 平安科技(深圳)有限公司 人脸遮挡检测方法及系统
CN111598018A (zh) * 2020-05-19 2020-08-28 北京嘀嘀无限科技发展有限公司 面部遮挡物的佩戴检测方法、装置、设备及存储介质
CN111598021B (zh) * 2020-05-19 2021-05-28 北京嘀嘀无限科技发展有限公司 面部遮挡物的佩戴检测方法、装置、电子设备及存储介质
CN111626193A (zh) * 2020-05-26 2020-09-04 北京嘀嘀无限科技发展有限公司 一种面部识别方法、面部识别装置及可读存储介质
CN113963393A (zh) * 2020-07-03 2022-01-21 北京君正集成电路股份有限公司 一种戴墨镜情况下的人脸识别方法
CN113963394A (zh) * 2020-07-03 2022-01-21 北京君正集成电路股份有限公司 一种下半部遮挡情况下的人脸识别方法
CN114078270B (zh) * 2020-08-19 2024-09-06 上海新氦类脑智能科技有限公司 基于遮挡环境下人脸身份验证方法、装置、设备和介质
CN112418190B (zh) * 2021-01-21 2021-04-02 成都点泽智能科技有限公司 移动端医学防护遮蔽人脸识别方法、装置、系统及服务器
CN112949418A (zh) * 2021-02-05 2021-06-11 深圳市优必选科技股份有限公司 说话对象的确定方法、装置、电子设备及存储介质
CN112966654B (zh) * 2021-03-29 2023-12-19 深圳市优必选科技股份有限公司 唇动检测方法、装置、终端设备及计算机可读存储介质
CN113762136A (zh) * 2021-09-02 2021-12-07 北京格灵深瞳信息技术股份有限公司 人脸图像遮挡判断方法、装置、电子设备和存储介质
CN114462495B (zh) * 2021-12-30 2023-04-07 浙江大华技术股份有限公司 一种脸部遮挡检测模型的训练方法及相关装置
CN117275075B (zh) * 2023-11-01 2024-02-13 浙江同花顺智能科技有限公司 一种人脸遮挡检测方法、系统、装置和存储介质
CN117282038B (zh) * 2023-11-22 2024-02-13 杭州般意科技有限公司 眼部光疗装置的光源调整方法、装置、终端及存储介质
CN118058717A (zh) * 2024-02-29 2024-05-24 首都医科大学附属北京积水潭医院 俯卧位手术患者面部动态监测预警系统、方法及设备

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102270308A (zh) * 2011-07-21 2011-12-07 武汉大学 一种基于五官相关aam模型的面部特征定位方法
CN102306304A (zh) * 2011-03-25 2012-01-04 杜利利 人脸遮挡物识别方法及其装置
CN104463172A (zh) * 2014-12-09 2015-03-25 中国科学院重庆绿色智能技术研究院 基于人脸特征点形状驱动深度模型的人脸特征提取方法
CN105868689A (zh) * 2016-02-16 2016-08-17 杭州景联文科技有限公司 一种基于级联卷积神经网络的人脸遮挡检测方法

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102542246A (zh) * 2011-03-29 2012-07-04 广州市浩云安防科技股份有限公司 Atm机异常人脸检测方法
CN105654049B (zh) * 2015-12-29 2019-08-16 中国科学院深圳先进技术研究院 人脸表情识别的方法及装置
CN106056079B (zh) * 2016-05-31 2019-07-05 中国科学院自动化研究所 一种图像采集设备及人脸五官的遮挡检测方法
CN106295566B (zh) * 2016-08-10 2019-07-09 北京小米移动软件有限公司 人脸表情识别方法及装置
CN106485215B (zh) * 2016-09-29 2020-03-06 西交利物浦大学 基于深度卷积神经网络的人脸遮挡检测方法
CN106910176B (zh) * 2017-03-02 2019-09-13 中科视拓(北京)科技有限公司 一种基于深度学习的人脸图像去遮挡方法

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102306304A (zh) * 2011-03-25 2012-01-04 杜利利 人脸遮挡物识别方法及其装置
CN102270308A (zh) * 2011-07-21 2011-12-07 武汉大学 一种基于五官相关aam模型的面部特征定位方法
CN104463172A (zh) * 2014-12-09 2015-03-25 中国科学院重庆绿色智能技术研究院 基于人脸特征点形状驱动深度模型的人脸特征提取方法
CN105868689A (zh) * 2016-02-16 2016-08-17 杭州景联文科技有限公司 一种基于级联卷积神经网络的人脸遮挡检测方法

Cited By (39)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110119674A (zh) * 2019-03-27 2019-08-13 深圳和而泰家居在线网络科技有限公司 一种作弊检测的方法、装置、计算设备及计算机存储介质
CN111860047B (zh) * 2019-04-26 2024-06-11 美澳视界(厦门)智能科技有限公司 一种基于深度学习的人脸快速识别方法
CN111860047A (zh) * 2019-04-26 2020-10-30 美澳视界(厦门)智能科技有限公司 一种基于深度学习的人脸快速识别方法
CN110414394A (zh) * 2019-07-16 2019-11-05 公安部第一研究所 一种面部遮挡人脸图像重建方法以及用于人脸遮挡检测的模型
CN110414394B (zh) * 2019-07-16 2022-12-13 公安部第一研究所 一种面部遮挡人脸图像重建方法以及用于人脸遮挡检测的模型
CN110543823A (zh) * 2019-07-30 2019-12-06 平安科技(深圳)有限公司 基于残差网络的行人再识别方法、装置和计算机设备
CN110543823B (zh) * 2019-07-30 2024-03-19 平安科技(深圳)有限公司 基于残差网络的行人再识别方法、装置和计算机设备
CN111353404A (zh) * 2020-02-24 2020-06-30 支付宝实验室(新加坡)有限公司 一种人脸识别方法、装置及设备
CN111353404B (zh) * 2020-02-24 2023-12-01 支付宝实验室(新加坡)有限公司 一种人脸识别方法、装置及设备
CN111414879A (zh) * 2020-03-26 2020-07-14 北京字节跳动网络技术有限公司 人脸遮挡程度识别方法、装置、电子设备及可读存储介质
CN111414879B (zh) * 2020-03-26 2023-06-09 抖音视界有限公司 人脸遮挡程度识别方法、装置、电子设备及可读存储介质
CN113449562A (zh) * 2020-03-26 2021-09-28 北京沃东天骏信息技术有限公司 人脸位姿校正方法和装置
CN111489373A (zh) * 2020-04-07 2020-08-04 北京工业大学 一种基于深度学习的遮挡物体分割方法
CN111489373B (zh) * 2020-04-07 2023-05-05 北京工业大学 一种基于深度学习的遮挡物体分割方法
CN111461047A (zh) * 2020-04-10 2020-07-28 北京爱笔科技有限公司 身份识别的方法、装置、设备及计算机存储介质
CN111486961A (zh) * 2020-04-15 2020-08-04 贵州安防工程技术研究中心有限公司 基于宽光谱人体额头成像与距离感知的高效额温估计方法
CN111444887A (zh) * 2020-04-30 2020-07-24 北京每日优鲜电子商务有限公司 口罩佩戴的检测方法、装置、存储介质及电子设备
CN111626240B (zh) * 2020-05-29 2023-04-07 歌尔科技有限公司 一种人脸图像识别方法、装置、设备及可读存储介质
CN111639596B (zh) * 2020-05-29 2023-04-28 上海锘科智能科技有限公司 基于注意力机制和残差网络的抗眼镜遮挡人脸识别方法
CN111626240A (zh) * 2020-05-29 2020-09-04 歌尔科技有限公司 一种人脸图像识别方法、装置、设备及可读存储介质
CN111639596A (zh) * 2020-05-29 2020-09-08 上海锘科智能科技有限公司 基于注意力机制和残差网络的抗眼镜遮挡人脸识别方法
CN111814571A (zh) * 2020-06-12 2020-10-23 深圳禾思众成科技有限公司 一种基于背景过滤的口罩人脸识别方法及系统
CN111881740B (zh) * 2020-06-19 2024-03-22 杭州魔点科技有限公司 人脸识别方法、装置、电子设备及介质
CN111881740A (zh) * 2020-06-19 2020-11-03 杭州魔点科技有限公司 人脸识别方法、装置、电子设备及介质
CN112052730B (zh) * 2020-07-30 2024-03-29 广州市标准化研究院 一种3d动态人像识别监控设备及方法
CN112052730A (zh) * 2020-07-30 2020-12-08 广州市标准化研究院 一种3d动态人像识别监控设备及方法
CN112016464A (zh) * 2020-08-28 2020-12-01 中移(杭州)信息技术有限公司 检测人脸遮挡的方法、装置、电子设备及存储介质
CN112016464B (zh) * 2020-08-28 2024-04-12 中移(杭州)信息技术有限公司 检测人脸遮挡的方法、装置、电子设备及存储介质
CN114255491A (zh) * 2020-09-11 2022-03-29 北京眼神智能科技有限公司 眼睛遮挡判断方法、装置、计算机可读存储介质及设备
CN112116525A (zh) * 2020-09-24 2020-12-22 百度在线网络技术(北京)有限公司 换脸识别方法、装置、设备和计算机可读存储介质
CN112597886A (zh) * 2020-12-22 2021-04-02 成都商汤科技有限公司 乘车逃票检测方法及装置、电子设备和存储介质
CN112633183B (zh) * 2020-12-25 2023-11-14 平安银行股份有限公司 影像遮挡区域自动检测方法、装置及存储介质
CN112633183A (zh) * 2020-12-25 2021-04-09 平安银行股份有限公司 影像遮挡区域自动检测方法、装置及存储介质
CN112766214A (zh) * 2021-01-29 2021-05-07 北京字跳网络技术有限公司 一种人脸图像处理方法、装置、设备及存储介质
CN113111817A (zh) * 2021-04-21 2021-07-13 中山大学 语义分割的人脸完整度度量方法、系统、设备及存储介质
CN113449696A (zh) * 2021-08-27 2021-09-28 北京市商汤科技开发有限公司 一种姿态估计方法、装置、计算机设备以及存储介质
CN113449696B (zh) * 2021-08-27 2021-12-07 北京市商汤科技开发有限公司 一种姿态估计方法、装置、计算机设备以及存储介质
CN114399813B (zh) * 2021-12-21 2023-09-26 马上消费金融股份有限公司 人脸遮挡检测方法、模型训练方法、装置及电子设备
CN114399813A (zh) * 2021-12-21 2022-04-26 马上消费金融股份有限公司 人脸遮挡检测方法、模型训练方法、装置及电子设备

Also Published As

Publication number Publication date
CN107633204A (zh) 2018-01-26
CN107633204B (zh) 2019-01-29

Similar Documents

Publication Publication Date Title
WO2019033572A1 (fr) Procédé de détection de situation de visage bloqué, dispositif et support d'informations
US10534957B2 (en) Eyeball movement analysis method and device, and storage medium
US10445562B2 (en) AU feature recognition method and device, and storage medium
WO2019033571A1 (fr) Procédé de détection de point de caractéristique faciale, appareil et support de stockage
CN107423690B (zh) 一种人脸识别方法及装置
US8792722B2 (en) Hand gesture detection
US8750573B2 (en) Hand gesture detection
WO2019033570A1 (fr) Procédé d'analyse de mouvement labial, appareil et support d'informations
WO2019109526A1 (fr) Procédé et dispositif de reconnaissance de l'âge de l'image d'un visage, et support de stockage
WO2019033568A1 (fr) Procédé de saisie de mouvement labial, appareil et support d'informations
WO2019061658A1 (fr) Procédé et dispositif de localisation de lunettes, et support d'informations
WO2019071664A1 (fr) Procédé et appareil de reconnaissance de visage humain combinés à des informations de profondeur, et support de stockage
WO2019041519A1 (fr) Procédé et dispositif de suivi de cible, et support de stockage lisible par ordinateur
US9633284B2 (en) Image processing apparatus and image processing method of identifying object in image
WO2019033567A1 (fr) Procédé de capture de mouvement de globe oculaire, dispositif et support d'informations
US20130202159A1 (en) Apparatus for real-time face recognition
WO2019056503A1 (fr) Procédé d'évaluation de surveillance de magasin, dispositif, et support d'informations
CN114898116A (zh) 一种基于嵌入式平台的车库管理方法、系统及存储介质
CN111582118A (zh) 一种人脸识别方法及装置
WO2023165616A1 (fr) Procédé et système de détection d'une porte arrière dissimulée d'un modèle d'image, support de stockage et terminal
Rai et al. Software development framework for real-time face detection and recognition in mobile devices
US12272175B2 (en) Evaluating method and system for face verification, and computer storage medium
HK1246921B (zh) 人脸遮挡检测方法、装置及存储介质
HK1246921A1 (en) Face shielding detection method, device and storage medium
HK1246921A (en) Face shielding detection method, device and storage medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17921768

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS (EPO FORM 1205A DATED 25.09.2020)

122 Ep: pct application non-entry in european phase

Ref document number: 17921768

Country of ref document: EP

Kind code of ref document: A1