[go: up one dir, main page]

WO2019017720A1 - Système de caméra permettant la protection de la confidentialité et procédé correspondant - Google Patents

Système de caméra permettant la protection de la confidentialité et procédé correspondant Download PDF

Info

Publication number
WO2019017720A1
WO2019017720A1 PCT/KR2018/008196 KR2018008196W WO2019017720A1 WO 2019017720 A1 WO2019017720 A1 WO 2019017720A1 KR 2018008196 W KR2018008196 W KR 2018008196W WO 2019017720 A1 WO2019017720 A1 WO 2019017720A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
resolution
low
unit
camera
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/KR2018/008196
Other languages
English (en)
Korean (ko)
Inventor
양현종
유상원
김기윤
김명언
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Egovid Inc
Original Assignee
Egovid Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from KR1020170092203A external-priority patent/KR101876433B1/ko
Priority claimed from KR1020170092204A external-priority patent/KR101911900B1/ko
Application filed by Egovid Inc filed Critical Egovid Inc
Publication of WO2019017720A1 publication Critical patent/WO2019017720A1/fr
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/09Supervised learning
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast

Definitions

  • the present invention relates to a technology for recognizing human behavior in a low-resolution image, a camera system for automatically adjusting a shooting resolution of the camera based on the technology, and a low- To a camera system capable of anonymizing video.
  • the present invention relates to a camera system for shooting a low-resolution image in a normal state to protect the privacy of a subject, a camera system for capturing a high-resolution image when a dangerous situation is determined through a behavior recognition technology for a low-resolution image, And an automatic behavior recognition method.
  • the present invention provides a camera system capable of automatically reducing the camera battery consumption by automatically adjusting the shooting resolution of the camera according to the importance of the situation through the low-resolution image behavior recognition technology, And to provide the above objects.
  • the present invention takes into account the characteristics of low-resolution image conversion, and learns a convolution neural network by mapping a plurality of low-resolution images transformed from a high-resolution image into an embedding space, And to provide an automatic behavior recognition method capable of increasing the human behavior recognition rate.
  • Another object of the present invention is to provide a camera system and a facial recognition-based image anonymization method that can protect privacy by separating a face region and an area other than a face region with respect to a high-resolution image by performing an ultra-low resolution processing only on the face region do.
  • a camera system generates learning data based on recognition of human behavior in a low-resolution image based on a high-resolution learning image, ; Resolution image, and when a crisis is detected through the first behavior recognition that recognizes human behavior in a low-resolution image using the human behavior recognition model transmitted from the data server, the image is changed to a high- And a camera for photographing.
  • the camera determines whether the image captured by changing to the high-resolution mode is a crisis state through a second behavior recognition that recognizes human behavior in the image using the human behavior recognition model, And stores the image photographed at a high resolution when the situation is not satisfied and the image photographed at a low resolution when the situation is not a crisis.
  • the camera may further include: a photographing unit photographing a low-resolution image; A communication unit for receiving a human behavior recognition model used for recognizing human behavior in the low resolution image from the data server and transmitting the received human behavior recognition model to the image analysis unit; An image analyzing unit for determining whether the user is in a crisis state through a first behavior recognition process for recognizing a human behavior in a low resolution image using the human behavior recognition model received from the communication unit; A control unit for controlling the photographing unit to change to a higher resolution mode than the existing resolution and to photograph the photographing unit if a crisis state is determined by the image analysis unit; And a storage unit for storing a low-resolution image or a high-resolution image photographed by the photographing unit.
  • the image analyzing unit can determine whether the image is a crisis state through a second behavior recognition that recognizes human behavior in the image using the human behavior recognition model for the image changed to the high-resolution mode.
  • the data server may include: a resolution conversion unit that converts a high-resolution human behavior image into a plurality of low-resolution images; A plurality of low resolution images generated by the resolution conversion unit are received and divided into a spatial stream and a time stream, and convolution and pooling are performed for each stream, and a fully connected layer is added A convolution neural network learning unit for learning a convolutional neural network; And a human behavior recognition model generation unit for generating a human behavior recognition model based on the data learned by the convolution neural network learning unit.
  • an automatic resolution adjusting method comprising the steps of: (a) capturing an image of a camera at a low resolution; (b) receiving a human behavior recognition model used by the communication unit of the camera for recognizing human behavior in a low-resolution image from an external data server and transmitting the human behavior recognition model to an image analysis unit of the camera; (c) determining whether the image analyzing unit is in a crisis state through a first behavior recognition that recognizes human behavior in a low-resolution image using the human behavior recognition model received from the camera communication unit; (d) when the image analyzing unit determines that the image capturing unit is in a crisis state, the control unit of the camera controls the image capturing unit to change the image capturing unit to a higher resolution mode than the existing resolution image capturing unit; And (e) storing a low-resolution image or a high-resolution image captured by the photographing unit by a storage unit of the camera.
  • the automatic resolution adjusting method may further include a step of, between the steps (d) and (e), performing a human action in the image using the human behavior recognition model And a second step of recognizing whether or not the user is in a crisis state by recognizing the second behavior.
  • an automatic behavior recognition method comprising: (a) converting a high-resolution human behavior image into a plurality of low-resolution images by a resolution conversion unit of a data server; (b) a convolutional neural network learning unit of the data server receives the plurality of low-resolution images and divides them into a spatial stream and a time stream, performs convolution and pooling for each stream, layer is further applied to learn a Convolutional Neural Network; (c) generating a human behavior recognition model based on data learned by the convolutional neural network learning unit; (d) photographing a low-resolution image of a photographing part of the camera; And (e) recognizing human behavior in the low-resolution image photographed by the photographing unit using the human behavior recognition model received from the data server by the image analysis unit of the camera.
  • a camera system including: a data server that performs learning for face recognition in an image using a learning image and generates a face recognition model; And a controller for receiving the facial recognition model from the data server, capturing a high resolution image to perform low resolution conversion, recognizing a face in the converted image using the facial recognition model transmitted from the data server, And a camera for outputting the anonymized image to the face region by re-converting the detected face region and the region other than the face region so as to have different resolutions.
  • a camera including: a communication unit for receiving a face recognition model for face recognition in an image from an external data server; A photographing unit for photographing a high-resolution image; A low-resolution conversion module that receives the high-resolution image photographed by the photographing unit, receives the position information of the face region and the target resolution value from the processor, and converts the high-resolution image; A processor for recognizing a face in the image converted by the low resolution conversion module using the face recognition model transmitted from the communication unit and updating the position of the face region and the target resolution value according to the face recognition result; And an output unit for outputting the image converted by the low resolution conversion module.
  • the processor when the face is detected in the image converted by the low resolution conversion module through face recognition, the processor records the detected face region and stores the position information, and the remaining region excluding the stored face region And the low resolution conversion module controls to convert the remaining area excluding the stored face area to the target resolution to be converted again.
  • a camera including: a communication unit for receiving a face recognition model for face recognition in an image from an external data server; A photographing unit for photographing a high-resolution image; A low-resolution conversion module that receives a high-resolution image captured by the imaging unit, receives a target resolution value from the processor, and converts the high-resolution image; A processor for recognizing a face in the image converted by the low resolution conversion module using the face recognition model transmitted from the communication unit and updating the target resolution value according to the face recognition result; And an output unit for outputting the image converted by the low resolution conversion module.
  • the processor may control the low-resolution conversion module to output the low-resolution image converted at the detection time when a face is detected in the image converted by the low-resolution conversion module.
  • an image anonymizing method comprising the steps of: (a) capturing a high resolution image of a photographing unit of a camera; (b) a communication unit of the camera receives a face recognition model for face recognition in an image from an external data server; (c) receiving a high-resolution image photographed by the photographing unit by the low resolution conversion module of the camera, receiving the position of the face region and the target resolution value from the processor, and converting the high resolution image; (d) recognizing a face in the image converted by the low resolution conversion module using the face recognition model received from the communication unit of the camera, and updating the position of the face region and the target resolution value according to the face recognition result ; And (e) the output of the camera outputting the image converted by the low resolution conversion module.
  • an image anonymizing method comprising the steps of: (a) capturing a high resolution image of a photographing unit of a camera; (b) a communication unit of the camera receives a face recognition model for face recognition in an image from an external data server; (c) receiving a high resolution image photographed by the photographing unit by the low resolution conversion module of the camera, receiving the target resolution value from the processor, and converting the high resolution image; (d) recognizing a face in the image converted by the low resolution conversion module using the face recognition model received from the communication unit of the camera processor, and updating the target resolution value according to the face recognition result; And (e) the output of the camera outputting the image converted by the low resolution conversion module.
  • a low-resolution image is normally photographed and a behavior recognition technology for a low-resolution image is determined to be a crisis, a high-resolution image is photographed, thereby protecting privacy of a subject.
  • the present invention it is possible to reduce the camera battery consumption and efficiently utilize the storage space by automatically adjusting the shooting resolution of the camera according to the importance of the situation through the low-resolution image behavior recognition technology.
  • a plurality of low-resolution images transformed from a high-resolution image are mapped into an embedding space in consideration of the characteristics of low-resolution image conversion, thereby learning a convolution neural network,
  • the behavior recognition rate can be increased.
  • the present invention it is possible to protect the privacy of the high-resolution image by distinguishing the face area from the area other than the face area and performing the ultra-low resolution processing only on the face area.
  • FIG. 1 is a block diagram of a camera system according to a first embodiment of the present invention.
  • FIG. 2 is a view for explaining the operational concept of the camera system according to the first embodiment of the present invention.
  • FIG. 3 is a diagram illustrating a configuration of a camera according to a first embodiment of the present invention.
  • FIG. 4 is a flowchart illustrating a flow of operation of the camera according to the first embodiment of the present invention.
  • FIG. 5 is a detailed block diagram of a data server in a camera system according to a first embodiment of the present invention.
  • FIG. 6 is a view for explaining that even when a high-resolution image is converted to a low-resolution image, a low-resolution image derived from the same high-resolution image may have a different value.
  • FIG. 7 is a view showing a structured manner in which the convolutional neural network learning unit according to the first embodiment of the present invention divides an input image into a spatial stream and a temporal stream.
  • FIG. 8 is a diagram showing the structure of a two-stream convolutional neural network applied to convolutional neural network learning according to the first embodiment of the present invention.
  • FIG. 9 is a diagram for explaining how the convolutional neural network learning unit according to the first embodiment of the present invention performs learning for intra-image behavior recognition using an embedding space.
  • FIG. 10 is a diagram illustrating a structure of a multi-siamese convolutional neural network applied to convolutional neural network learning according to the first embodiment of the present invention.
  • FIG. 11 is a flowchart illustrating a process of learning a convolutional neural network among methods of automatically recognizing behaviors in a video camera system according to the first embodiment of the present invention.
  • FIG. 12 is a diagram showing a comparison of accuracy of behavior recognition between a case in which a method of automatically detecting an intra-video motion in a camera system according to the first embodiment of the present invention is applied and a conventional technique.
  • FIG. 13 is a block diagram illustrating the configuration of a camera system according to a second embodiment of the present invention.
  • FIG. 14 is a diagram showing the structure of a convolutional neural network (CNN) used by a data server of a camera system according to a second embodiment of the present invention to generate a face recognition model.
  • CNN convolutional neural network
  • 15 is a diagram illustrating an example of a video image processed by the video anonymization method according to the second embodiment of the present invention.
  • 16 is a block diagram showing a configuration of a camera according to a second embodiment of the present invention.
  • FIG. 17 is a view simply showing an anonymizing process of an image photographed by a camera according to the second embodiment of the present invention in real time.
  • FIG. 18 is a flowchart illustrating a process of performing an image anonymization method according to the second embodiment of the present invention.
  • FIG. 19 is a flowchart illustrating a process of performing an image anonymizing method according to the third embodiment of the present invention.
  • a component when referred to as being connected or connected to another component, it may be directly connected or connected to the other component, but it should be understood that there may be other components in between.
  • FIGS. 13 to 19 relate to a second embodiment, And a camera system for recognizing and anonymizing a corresponding part.
  • FIG. 1 is a view illustrating a configuration of a camera system according to an embodiment of the present invention
  • FIG. 2 is a diagram illustrating an operation concept of a camera system according to an embodiment of the present invention.
  • the camera system 1 of the present invention includes a behavior recognition-based resolution automatic adjustment camera 10 and a data server 20.
  • the behavior recognition-based resolution automatic adjustment camera 10 photographs a subject
  • the data server 20 receives a high-resolution image, performs learning for human behavior recognition, and recognizes human behavior in a low-resolution image based on the learned data To generate a human behavior recognition model.
  • the motion recognition based resolution control camera 10 of the present invention may include various camera devices. For example, it can be a smart home camera, a robot camera, a security CCTV, a wearable camera, a car black box camera, and the like.
  • the data server 20 can perform learning for recognizing human behavior in a low-resolution image from a high-resolution image stored on-line or a high-resolution image on-line.
  • the data server 20 can generate learning data by learning a convolutional neural network, and can generate a human behavior recognition model based on the generated learning data.
  • FIGS. 5 to 10. FIG.
  • a camera 10 (hereinafter, referred to as a 'camera') photographs a low-resolution moving image in a usual manner to prevent privacy invasion of a subject.
  • the camera 10 of the present invention recognizes human behavior in a low-resolution image (for example, 16x12 pixels) by applying a behavior recognition technique to the photographed image.
  • the camera 10 can pick up an image of a high resolution (for example, 4K pixels) by selecting an optimal camera resolution according to a result of behavior recognition, and can store the captured image in a memory or transmit it to the data server 20. Therefore, the camera 10 of the present invention can solve the problem that the privacy of the subject can be violated by continuously photographing a high-resolution image even in a conventional camera or the like.
  • the present invention reduces battery consumption of the camera 10 and enables efficient use of the storage space. The specific operation of the camera 10 of the present invention will be described with reference to FIGS. 3 to 4. FIG.
  • FIG. 3 is a diagram illustrating a configuration of a camera according to a first embodiment of the present invention.
  • the camera 10 may include a photographing unit 100, a communication unit 200, an image analysis unit 300, a storage unit 400, and a control unit 500.
  • the shooting unit 100 can shoot a subject at a high resolution.
  • the communication unit 200 receives the human behavior recognition model used for recognizing human behavior in the low resolution image from the data server 20 and transmits the received human behavior recognition model to the image analysis unit 300.
  • the data server 20 can perform learning to recognize human behavior in a low-resolution image from a high-resolution image stored on-line or a high-resolution image on-line.
  • the high-resolution image is used to learn a Convolutional Neural Network to perform human behavior recognition in a low-resolution image photographed by the photographing unit 100.
  • a high-resolution image as a learning image may be a publicly available source.
  • a publicly available source can be, for example, an open source such as YouTube.
  • high-resolution images can be obtained from sources that publicly provide video online.
  • the communication unit 200 may transmit the image stored in the storage unit 400 to the data server 20.
  • the communication unit 200 may include at least one of a wireless communication module and a wired communication module.
  • the wireless communication module may include a wireless network communication module, a wireless local area network (WLAN), a wireless communication module such as a Wi-Fi, a wireless fidelity or WiMAX, and a wireless personal area network (WPAN) Or the like.
  • WLAN wireless local area network
  • WPAN wireless personal area network
  • the image analysis unit 300 can receive the human behavior recognition model from the communication unit 200 and recognize the human behavior in the low resolution image photographed by the photographing unit 100. [ The image analyzing unit 300 can recognize the human behavior in the low-resolution image photographed by the photographing unit 100 using the human behavior recognition model, and judge whether or not the recognized human behavior corresponds to the crisis situation recognition).
  • the storage unit 400 may store a low-resolution image or a high-resolution image photographed by the photographing unit 100.
  • the storage unit 400 stores images (weapons, naked images, drugs, blood, etc.) to be recognized or analyzed by the image or image analysis unit 300 taken by the image sensing unit 100 as a dangerous situation, And may be in the form of a flash memory for providing images in a real-time streaming manner.
  • the storage unit 400 includes a main storage unit and an auxiliary storage unit, and stores an application program necessary for the functional operation of the camera 10.
  • the storage unit 400 may include a program area and a data area.
  • the program area is an area for storing an operating system or the like for booting the camera 10
  • the data area is an area for storing an image photographed by the photographing part 100.
  • the control unit 500 may control the photographing unit 100, the communication unit 200, the image analysis unit 300, and the storage unit 400.
  • the control unit 500 can control the photographing unit 100 to change the mode to a higher resolution mode than the existing resolution and take a picture if the crisis is determined by the image analysis unit 300. [ This is to maximize the privacy of the photographed subject and to minimize the battery consumption of the camera 10 and the utilization of the storage space. That is, the normal photographing unit 100 photographs with a low resolution image (ex. 16x12 pixels), and when receiving the control signal, it can photograph with a high resolution image (ex. 720x480, 1280x720, 1920x1080, 4K, etc.).
  • the control unit 500 may control the photographing unit 100 to operate at a predetermined cycle to minimize battery consumption.
  • FIG. 4 is a flowchart for explaining the flow of operation of the camera according to the first embodiment of the present invention.
  • the photographing unit 100 photographs a normal low-resolution image and transmits the photographed image to the image analysis unit 300 (S210).
  • the image analyzing unit 300 determines whether the user is in a crisis state by recognizing a first action in the low resolution image using the human behavior recognition model received from the communication unit 200 at step S220.
  • the control unit 500 transmits a control signal so that the low resolution image is stored in the storage unit 400 when the image analysis unit 300 determines that the human action in the low resolution image does not correspond to the crisis situation.
  • the control unit 500 transmits a control signal to the imaging unit 100 to capture the subject as a high-resolution image (S240).
  • the image analyzing unit 300 determines whether a high-resolution image photographed by the photographing unit 100 is in a crisis state through a second behavior recognition that recognizes human behavior in the image using the human behavior recognition model at step S250.
  • the control unit 500 transmits a control signal to store the high-resolution image in the storage unit 400, A control signal is transmitted so that an image photographed at a low resolution is stored in the storage unit 400 (S260).
  • the camera 10 of the present invention can recognize the human behavior by performing two human behavior recognizations on the photographed image, and finally store the photographed image. Therefore, it is possible to reduce the error of the behavior recognition, to protect the privacy of the object and to appropriately use the storage space.
  • FIG. 5 is a detailed block diagram of a data server in a camera system according to a first embodiment of the present invention.
  • the data server 20 may include a resolution conversion unit 21, a convolutional neural network learning unit 22, and a human behavior recognition model generation unit 23.
  • the resolution conversion unit 21 can convert the high-resolution human behavior image received from the communication unit 200 into a plurality of low-resolution images. Multiple low-resolution images can be obtained using a technique called Inverse Super Resolution (ISR).
  • ISR Inverse Super Resolution
  • the main motivation of Inverse Super Resolution (ISR) is that a high resolution image contains a quantity of information corresponding to a low resolution image set, and a behavior recognition method can be performed by applying another low resolution image transformation to a single high resolution image. (M. S. Ryoo, B. Rothrock, C. Fleming, and H. J. Yang. Privacy-preserving human activity recognition from extreme low resolution. In AAAI, 2017.)
  • Resolution conversion section 21 generates n low-resolution images (i.e., V ik ) for each high-resolution training image X i by applying transform sets F k and D k as shown in the following equation (1) .
  • Fk is the camera motion transformation and Dk is the downsampling operator.
  • F k can be any affine transformation
  • the present invention regards the combination of transformation, scaling, and rotation as motion transformation.
  • standard mean downsampling is used for D k .
  • a sufficient number of low resolution transformed images V ik generated from the learning samples X i are provided to the convolutional neural network CNN and the classifier 330 so that efficient learning can be performed.
  • the convolutional neural network learning unit 22 receives a plurality of low resolution images generated by the resolution conversion unit 21 and divides them into a spatial stream and a time stream (optical flow), and performs convolution and pooling ) And apply a fully connected layer to learn the Convolutional Neural Network.
  • the convolutional neural network learning unit 22 can know that spatial streams and time streams derived from low-resolution images are used as parameters.
  • the spatial resolution of the low-resolution image is 16 ⁇ 12 pixels.
  • the spatial stream is configured to input the RGB pixel values of each frame (i.e., the input dimension is 16x12x3)
  • the temporal stream is the 10-frame connection of the X and Y optical flow images (i.e., 16x12x20) Lt; / RTI >
  • the X and Y optical flow images are constructed by calculating the " x (and y) optical flow magnitude " per pixel.
  • the human behavior recognition model generation unit 23 can generate a human behavior recognition model for recognizing human behavior in an image photographed by the photographing unit 100 based on data learned by the convolution neural network learning unit 22 have.
  • the human behavior recognition model may correspond to a classifier.
  • the classifier can shorten the image information and motion information included in the image based on the learned data when a new image is input, and can recognize the behavior in the image using the reduced information.
  • the image analysis technique of the present invention can be applied not only to a convolutional neural network (CNN) but also to a support vector machine (SVM).
  • FIG. 8 is a diagram illustrating a structure of a two-stream convolutional neural network applied to convolutional neural network learning according to an embodiment of the present invention.
  • the convolutional neural network learning unit 22 can perform learning for behavior recognition using a two-stream convolutional neural network.
  • a two-stream convolution neural network is applied to each frame of the image. They are summarized using time pyramids and generate a single video representation. Let a 2-stream network applied to each frame V t of video V at time t be h (V t ). Then, the expression f (V;?) Is calculated according to the following equation (2).
  • "," denotes the vector concatenation operator
  • T is the number of frames in the image V
  • fc represents the set of fully connected layers to be applied at the top of the connection.
  • the size of h (V t ) is 512-D: 256 x 2.
  • is a set of parameters in CNN that require learning from learning data.
  • max is a temporary maximum pooling operator that computes the maximum value of each element. In the example, four levels of time pyramids were used (i.e., a total of 15 maximum poolings).
  • a fully connected layer and a softmax layer can be applied to f (V) to learn the classifier.
  • Let g be such a layer, y g (f (V;?)).
  • y is an activity class label.
  • the learning g (f (V; [theta]) having classification loss using the low resolution image generated by the resolution converting unit 21 can provide a basic image classification model.
  • Figure 8 shows the overall structure of a two-stream CNN model using this time pyramid.
  • FIG. 9 is a diagram for explaining how the convolutional neural network learning unit 22 according to the first embodiment of the present invention performs learning for intra-image behavior recognition using an embedding space.
  • the two-stream network design of FIG. 8 can classify behavioral images by learning model parameters optimized for classification, but does not consider the characteristics of ultra-low resolution images in which different low resolution data are generated due to different transformations applied to the same scene .
  • the classifier In order for the classifier to better reflect the characteristics of the ultralow-resolution image, the embedding space, which maps to the same embedding location, regardless of the conversion to different low-resolution images with the same semantic content, .
  • Embedding learning is commonly optimized for both embedded learning and classification using the learned embedding in an end-to-end manner, and more generalized (i.e., overfilled) Enables learning of classifiers.
  • embedding learning is performed by minimizing the embedding distance between negative pairs while maximizing the embedding distance between positive pairs.
  • FIG. 10 is a diagram illustrating a structure of a multi-siamese convolutional neural network applied to convolutional neural network learning according to an embodiment of the present invention.
  • the resolution converting unit 21 converts a first batch composed of n low-resolution images having different resolutions and a plurality of different high-resolution images converted from the same high-resolution image, A second batch composed of n low-resolution images having the same size as the training data, and transmits the training data to the convolutional neural network learning unit 22.
  • the convolutional neural network learning unit 22 receives the first and second arrangements, divides each of the images into a spatial stream and a temporal stream, performs convolution and pooling for each stream, After the layer (Fully connected layer) is further applied, the embedding distance is mapped by mapping to the embedding space, so that learning for motion recognition in an image can be performed.
  • the Siamese neural network is a concept often used to learn similarity measures between two inputs, with two networks sharing the same parameters.
  • the goal of the Siamese network is to learn the embedding space in which similar items (low-resolution images in the case of the present invention) are placed nearby. More specifically, a sample corresponding to a positive pair that should be located close in the embedding space and a sample corresponding to a negative pair that should be located remotely are used for learning.
  • B is the batch of low resolution learning examples used
  • i and j are the index of the learning pair in the batch.
  • y ' ( i, j ) is a binary number, 1 for positive pairs and 0 for negative pairs.
  • the positive pair is composed of two low-resolution images derived from the same high-resolution image
  • the negative pair is composed of two low-resolution images derived from different high-resolution images.
  • Figure 9 shows embedding learning with contrast loss.
  • the network must be learned by using the combining loss function as shown in the following equation (4).
  • the multi-siamese convolutional neural network of the present invention has 2n network copies sharing the same parameter ⁇ for f (V; ⁇ ), unlike the standard cyamises network with only two network sharing parameters.
  • the multi-siamese convolutional neural network maps each copy to each of the n different low resolution transforms (i.e., F k ). Therefore, the embedding distance can be kept small by using the contrast loss.
  • the multi-siamuse convolutional neural network has n more network copies using images that do not match the scene of the first n branch to form a negative learning pair.
  • X ik f (V ik;? ), where V ik is obtained by applying the transform F k to X i .
  • Two types of placement are randomly generated based on placement B of the original high resolution learning image.
  • B 1 is the arrangement of the low-resolution images generated in the single high-resolution image Xi
  • B 2 is the arrangement of the randomly selected low-resolution images.
  • B 1 produces a positive pair
  • B 2 produces a negative pair.
  • the size of B 1 and B 2 should be n.
  • the low-resolution example Vj of B 2 is provided directly to the remaining n branches of the siamese network. Therefore, the new loss function is expressed by the following equation (5).
  • the multi-siamese convolution neural network model of the present invention simultaneously considers a plurality of low-resolution transforms for embedding learning.
  • the new loss function essentially takes all pairs of n low resolution transforms as positive pairs and considers the same number of negative pairs using a separate arrangement.
  • the final loss function is calculated by combining the multi-cimase contrast loss with the standard classification loss performed in equation (4).
  • FIG. 10 shows the overall process of learning multi-cyamics embedding and classifier. This can be seen as Siamese CNN, which combines multiple contrast losses of different low resolution pairs.
  • the multi-siamese convolution neural network model of the present invention uses three fully connected layers for embedding learning and classification. After applying the time pyramid, we obtain an intermediate representation of 7680-D (ie, 15 ⁇ 256 ⁇ 2) per image. Next we get a fully connected layer of size 8192. Embedding learning occurs after the second full-connect layer, where x has a dimension of 8192-D. Classification is performed by adding another fully connected layer and a soft max layer at the top.
  • FIG. 11 is a flowchart illustrating a step of learning a convolutional neural network among methods of automatically recognizing behaviors in a video camera system according to an embodiment of the present invention.
  • the data server 20 receives a high-resolution learning image (S1101).
  • the resolution conversion unit 21 converts the received high-resolution human behavior image into a plurality of low-resolution images (S1102).
  • the resolution converting unit 21 converts a first batch composed of n low resolution images having different resolutions and a plurality of different high resolution images converted from the same high resolution image, (S1103), and transmits the second batch to the convolutional neural network learning unit 22.
  • the convolutional neural network learning unit 22 generates a second batch of low-resolution images as learning data (S1103).
  • the convolutional neural network learning unit 22 receives a plurality of low-resolution images and divides them into a spatial stream and a temporal stream, performs a convolution and a pooling for each stream, and performs au, a fully connected layer, And the convolutional neural network is learned (S1104). At this time, the convolutional neural network learning unit 22 receives the first and second arrangements, divides each of the images into a spatial stream and a time stream, and performs convolution and pooling for each stream After the Fully connected layer is further applied, the embedding distance is mapped to the embedding space to perform learning for intra-image motion recognition (S1105).
  • FIG. 12 is a view comparing the accuracy of the behavior recognition between the case where the method of automatically detecting the behavior in the camera of the camera system according to the embodiment of the present invention and the accuracy of the behavior recognition between the prior art.
  • FIG. 12 shows the results of comparing the accuracy of classification using a 16x12 HMDB dataset or a DogCentric activity dataset in the case of (1) applying basic 1-stream CNN, (2) applying 2-stream CNN, Table.
  • two CNNs we divided into (i) learning without using multiple low-resolution transforms, (ii) learning with multiple low-resolution transformations applied but without embedding learning, and (iii) learning with Siamese embedding learning. Respectively.
  • the number of low-resolution transformations used in the experiment is 15. These transformations were randomly selected from a uniform pool. A total of 25 motion transforms F k were provided in the X and Y directions of ⁇ -5, -2.5, 0, +2.5, + 5 ⁇ %. In addition, a total of 75 transforms were provided with three different rotations of the angle ⁇ -5, 0, 5 ⁇ degrees. 27 transforms out of 75 were randomly selected.
  • 12B is a table showing the result of comparing the present invention with the conventional technique. As shown in the table, it can be seen that the classification accuracy according to the present invention is higher than 8%. That is, the method of applying the embedding learning to the 2-stream multi-Siamese CNN, which is one embodiment of the present invention, yields the best result in the 16x12 resolution motion recognition.
  • 12C and 12D are tables showing the results of experiments using DogCentric activity dataset.
  • FIG. 13 is a block diagram illustrating the configuration of a camera system according to a second embodiment of the present invention.
  • the privacy camera system 5 may include a camera 50 and a data server 60.
  • the camera 50 receives the face recognition model from the data server 60, captures a high-resolution image and performs low-resolution conversion.
  • the camera 50 uses the face recognition model transmitted from the data server 60, And an image that has been anonymized with respect to the face region can be output by re-converting the detected face region and the region other than the face region according to the face recognition result to have different resolutions.
  • the camera 50 of the present invention may include various camera devices. For example, it can be a smart home camera, a robot camera, a security CCTV, a wearable camera, a car black box camera, and the like. The specific configuration and function of the camera 50 will be described below with reference to FIG. 15 to FIG.
  • the data server 60 can perform learning for face recognition in an image using a learning image and generate a face recognition model.
  • the learning image may include both a high-resolution image and a low-resolution image.
  • the data server 60 uses a convolutional neural network (CNN) in generating a face recognition model.
  • CNN convolutional neural network
  • Convolutional neural network (CNN) is an algorithm with a structure that learns to analyze and classify input image (image frame).
  • the Fully convolutional network is newly applied to the ultra-low resolution face recognition.
  • Such an FCN model can analyze and classify whether or not a specific pixel of an input image (image frame) is a face pixel using learning data (a plurality of images including a 10x10 face).
  • FCN Fully convolutional network
  • FCN Fully Convolutional Network
  • This FCN is learned to be specialized for 10x10 size face recognition (detection), and learning can be performed using a face image database.
  • the learning follows a backpropagation method in the convolutional neural network.
  • the FCN judges whether a new input image corresponds to a face for each pixel to derive a probability that the corresponding pixel is a face.
  • the size of the face pixel may be 10 x 10 or larger.
  • a sliding window method is applied to find the position and size of the face.
  • a sliding window is a method of checking how many face pixels are present in a box by applying a bounding box of a predetermined size to all positions of the image. Using this, we find a box containing a lot of pixels with a high probability of a face, and non-maximum suppression of such boxes to find a final bounding box.
  • Non-maximum suppression refers to thinning edges obtained from image processing. Non-maximum suppression is performed to find a sharper line because the edges found using Gaussian mask and sob mask are blurred edges that are easily crumpled. In other words, non-maximum suppression is a process of comparing the pixel values of 8 directions based on the center pixel and removing the center pixel when the center pixel is the largest.
  • An integral image is simply an image with the sum of the next pixel plus the previous pixel.
  • the integral image is used, the sum of pixel values of a specific area can be obtained very easily.
  • FIG. 15 is a diagram illustrating an example of an image that has been anonymized by the face recognition based real-time automatic image anonymization method of the present invention.
  • the leftmost image is a high-resolution image
  • the middle image is a low-resolution image
  • the rightmost image corresponds to an anonymized image.
  • the camera and image anonymizing method according to the second embodiment of the present invention can convert only a face area into an ultra-low resolution image and photograph a region outside the face area while maintaining a high resolution.
  • 16 is a block diagram showing the configuration of a camera 50 according to the second embodiment of the present invention.
  • the camera 50 of the present invention includes a communication unit 150, a photographing unit 250, a low resolution conversion module 350, a processor 450, an output unit 550, and a storage unit 650 can do.
  • the communication unit 150 may receive the face recognition model for performing face recognition in the image from the data server 60 and transmit the received face recognition model to the processor 450.
  • the face recognition model used for recognizing a face in an image captured by the camera 50 is generated by the data server 60.
  • the communication unit 150 may transmit the image captured and stored by the camera 50 to the data server 60.
  • the communication unit 150 may include at least one of a wireless communication module and a wired communication module.
  • the wireless communication module may include a wireless network communication module, a wireless local area network (WLAN), a wireless communication module such as a Wi-Fi, a wireless fidelity or WiMAX, and a wireless personal area network (WPAN) Or the like.
  • WLAN wireless local area network
  • WPAN wireless personal area network
  • the photographing unit 250 can photograph a high-resolution image with respect to a subject.
  • the photographing unit 250 transmits the photographed high resolution image to the low resolution conversion module 350.
  • Resolution conversion module 350 receives the high resolution image from the image sensing unit 250 and receives the position of the face region and the target resolution value from the processor 450 to convert the high resolution image.
  • the position of the initial face region to be input may be an empty value that is not set, and the initial target resolution value may be a predetermined low resolution value.
  • the ultralow resolution value may be 16x12 pixels. That is, the low resolution conversion module 350 initially performs a low resolution conversion on the entire image without distinguishing the regions other than the face region and the face region.
  • the low-resolution conversion module 350 receives the position of the face region stored from the processor 450, and converts the face region and the non-face region in the image. At this time, the face area can be converted into a predetermined low resolution resolution, and the area outside the face area can be converted into the target resolution value.
  • the low resolution conversion module 350 may be implemented as a circuit and only receives the face area position information and the target resolution value from the processor 450 and converts the image frame continuously supplied from the photographing unit 250, It does not have storage space or memory.
  • the low resolution conversion module 350 may convert the high resolution image by receiving only the target resolution value from the processor 450. That is, the low resolution conversion module 350 can perform the low resolution conversion on the entire image without distinguishing the face region separately.
  • the processor 450 can recognize a face in the image converted by the low resolution conversion module 350 using the face recognition model transmitted from the communication unit 150.
  • the processor 450 may be a processing unit such as a CPU (Central Processing Unit) or a GPU (Graphics Processing Unit).
  • the processor 450 updates the position of the face area and the target resolution value according to the face recognition result. That is, when a face is detected in the image converted by the low resolution conversion module 350, the processor 450 records the detected face region, stores the position information, and searches for a target Increase the resolution value.
  • the processor 450 causes the low resolution conversion module 350 to convert again to the increased resolution for the remaining area outside the detected face area. Thereafter, the processor 450 performs face recognition again on the converted image with the increased resolution, and repeats the above process.
  • the camera 50 of the present invention repeatedly performs face recognition / detection and resolution conversion while maintaining the resolution of the face area in the image to NxN and increasing the resolution of the area outside the face area to the final target resolution.
  • An image output through this process is called an " anonymizing video ".
  • the processor 450 stops increasing the resolution of the instantaneous image in which the face is detected, instead of increasing only the resolution of the remaining region except for the detected face region when a face is detected in the image, and immediately outputs the low- .
  • the resolution of the area other than the face area is increased to speed up the calculation speed compared with the case of repeating resolution conversion and face recognition, but there is a disadvantage in that the anonymization processing is not precise.
  • the output unit 550 outputs the image converted by the low resolution conversion module 350. That is, the output unit 550 outputs the anonymized image.
  • the storage unit 650 may store an image output by the output unit 550.
  • the storage unit 650 includes a main storage device and an auxiliary storage device, and stores an application program necessary for the functional operation of the camera 50.
  • the storage unit 650 may include a program area and a data area.
  • FIG. 17 is a view simply showing an anonymizing process of an image taken by a camera 50 according to the second embodiment of the present invention in real time.
  • the photographing unit 250 transmits the photographed high resolution image to the low resolution conversion module 350.
  • the low resolution conversion module 350 converts the received high resolution image into a low resolution image and transmits the low resolution image to the processor 450 do.
  • the processor 450 performs face recognition on the low-resolution image, transfers the face region location information and the target resolution value according to the face recognition result to the low resolution conversion module 350, and performs the conversion again.
  • the target resolution value reaches the final target resolution, the image converted by the low resolution conversion module 350 is output.
  • FIG. 18 is a flowchart illustrating a process of performing an image anonymization method according to the second embodiment of the present invention.
  • the photographing unit 250 transmits the photographed high-resolution image to the low resolution conversion module 350 (S1810).
  • the low resolution conversion module 350 receives the position information of the face region and the target resolution value from the processor 450 and converts the face region in the high resolution image to the super low resolution in operation S1820.
  • Resolution conversion (S1830). However, since there is no face region position information at the initial conversion, the resolution conversion is performed at a very low resolution with respect to the entire image.
  • the processor 450 performs facial recognition on the image converted by the low resolution conversion module 350 using the facial recognition model transmitted from the communication unit 150 (S1840).
  • the processor 450 adds the position information of the detected face region to the low resolution conversion module 350 (S1850). If there is no new face in the face recognition result image, the processor 450 increases only the target resolution value and transmits it to the low resolution conversion module 350 (S1860). When the resolution of the region other than the face region in the image matches the resolution of the high resolution image, which is the original image, the processor 450 outputs the image at that point of time and ends the processing. However, if the resolution of the region other than the face region in the converted image does not match the resolution of the high resolution image, which is the original image, the face recognition is performed again (S1870).
  • FIG. 19 is a flowchart illustrating a process of performing a face recognition based real-time automatic image anonymizing method according to another embodiment of the present invention.
  • the photographing unit 250 transmits the photographed high resolution image to the low resolution conversion module 350 (S1910).
  • the low resolution conversion module 350 receives the target resolution value from the processor 450 and performs low resolution conversion on the entire image of the high resolution image (S1920). However, in the initial conversion, the resolution conversion is performed at a very low resolution for the entire image.
  • the processor 450 performs facial recognition on the image converted by the low resolution conversion module 350 using the facial recognition model transmitted from the communication unit 150 (S1930). If there is no new face in the face recognition result image, the processor 450 increases the target resolution value and transmits it to the low resolution conversion module 350 (S1940).
  • the processor 450 controls the output unit 550 to output the converted low resolution image at the face detection time and ends the processing in step S1950.
  • the processor 450 outputs the image at that point of time and ends the processing .
  • the camera system according to the first and second embodiments and the method of driving the camera system have been described above.
  • the automatic recognition of behavior-based resolution, the automatic behavior recognition method, and the face recognition based image anonymization method according to the present invention can be implemented in the form of a program command that can be executed through various computer means and recorded in a computer readable medium .
  • the computer-readable medium may include program instructions, data files, data structures, and the like, alone or in combination.
  • the program instructions recorded on the medium may be those specially designed and constructed for the present invention or may be available to those skilled in the art of computer software.
  • Examples of computer-readable media include magnetic media such as hard disks, floppy disks and magnetic tape; optical media such as CD-ROMs and DVDs; magnetic media such as floppy disks; Magneto-optical media, and hardware devices specifically configured to store and execute program instructions such as ROM, RAM, flash memory, and the like.
  • Examples of program instructions include machine language code such as those produced by a compiler, as well as high-level language code that can be executed by a computer using an interpreter or the like.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Biomedical Technology (AREA)
  • Data Mining & Analysis (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Mathematical Physics (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Processing (AREA)
  • Image Analysis (AREA)

Abstract

La présente invention concerne une technologie permettant de reconnaître l'activité humaine dans une image basse résolution, et un système de caméra permettant de régler automatiquement une résolution de photographie d'une caméra en fonction de la technologie, et la présente invention concerne également une caméra permettant l'anonymisation d'une séquence vidéo à l'aide d'une technologie de reconnaissance/détection de visage faible résolution, et un procédé d'anonymisation de cette dernière.
PCT/KR2018/008196 2017-07-20 2018-07-20 Système de caméra permettant la protection de la confidentialité et procédé correspondant Ceased WO2019017720A1 (fr)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
KR10-2017-0092203 2017-07-20
KR10-2017-0092204 2017-07-20
KR1020170092203A KR101876433B1 (ko) 2017-07-20 2017-07-20 행동인식 기반 해상도 자동 조절 카메라 시스템, 행동인식 기반 해상도 자동 조절 방법 및 카메라 시스템의 영상 내 행동 자동 인식 방법
KR1020170092204A KR101911900B1 (ko) 2017-07-20 2017-07-20 사생활 보호 카메라와 그 시스템 및 얼굴 인식 기반 실시간 자동 영상 익명화 방법

Publications (1)

Publication Number Publication Date
WO2019017720A1 true WO2019017720A1 (fr) 2019-01-24

Family

ID=65015145

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2018/008196 Ceased WO2019017720A1 (fr) 2017-07-20 2018-07-20 Système de caméra permettant la protection de la confidentialité et procédé correspondant

Country Status (1)

Country Link
WO (1) WO2019017720A1 (fr)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111178319A (zh) * 2020-01-06 2020-05-19 山西大学 基于压缩奖惩机制的视频行为识别方法
CN111970509A (zh) * 2020-08-10 2020-11-20 杭州海康威视数字技术股份有限公司 一种视频图像的处理方法、装置与系统
DE102020115697A1 (de) 2020-06-15 2021-12-16 Iav Gmbh Ingenieurgesellschaft Auto Und Verkehr Verfahren und Vorrichtung zur natürlichen Gesichtsanonymisierung in Echtzeit
US20220060617A1 (en) * 2018-12-17 2022-02-24 Hanwha Techwin Co., Ltd. Camera control device and method to process captured image using at least one camera
US12277542B2 (en) 2018-12-17 2025-04-15 Hanwha Vision Co., Ltd. Automatic payment system

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20110033680A (ko) * 2009-09-25 2011-03-31 삼성전자주식회사 로봇 시스템의 영상 처리 장치 및 그 방법
KR101458136B1 (ko) * 2014-02-28 2014-11-05 김호 영상 처리 방법, 이를 실행하는 영상 처리 서버, 감시 서버, 영상 처리 시스템 및 이를 저장한 기록 매체
KR20140131188A (ko) * 2013-05-03 2014-11-12 딕스비전 주식회사 저비용으로 사생활 보호 및 관심 강조 기능을 구비하는 통합 관제 시스템 및 방법
KR20160088224A (ko) * 2015-01-15 2016-07-25 삼성전자주식회사 객체 인식 방법 및 장치
KR101722664B1 (ko) * 2015-10-27 2017-04-18 울산과학기술원 능동형 상황 인식을 위한 다인칭 시스템, 웨어러블 카메라, cctv, 관제서버 및 방법

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20110033680A (ko) * 2009-09-25 2011-03-31 삼성전자주식회사 로봇 시스템의 영상 처리 장치 및 그 방법
KR20140131188A (ko) * 2013-05-03 2014-11-12 딕스비전 주식회사 저비용으로 사생활 보호 및 관심 강조 기능을 구비하는 통합 관제 시스템 및 방법
KR101458136B1 (ko) * 2014-02-28 2014-11-05 김호 영상 처리 방법, 이를 실행하는 영상 처리 서버, 감시 서버, 영상 처리 시스템 및 이를 저장한 기록 매체
KR20160088224A (ko) * 2015-01-15 2016-07-25 삼성전자주식회사 객체 인식 방법 및 장치
KR101722664B1 (ko) * 2015-10-27 2017-04-18 울산과학기술원 능동형 상황 인식을 위한 다인칭 시스템, 웨어러블 카메라, cctv, 관제서버 및 방법

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20220060617A1 (en) * 2018-12-17 2022-02-24 Hanwha Techwin Co., Ltd. Camera control device and method to process captured image using at least one camera
US12223503B2 (en) * 2018-12-17 2025-02-11 Hanwha Vision Co., Ltd. Camera control device and method to process captured image using at least one camera
US12277542B2 (en) 2018-12-17 2025-04-15 Hanwha Vision Co., Ltd. Automatic payment system
CN111178319A (zh) * 2020-01-06 2020-05-19 山西大学 基于压缩奖惩机制的视频行为识别方法
DE102020115697A1 (de) 2020-06-15 2021-12-16 Iav Gmbh Ingenieurgesellschaft Auto Und Verkehr Verfahren und Vorrichtung zur natürlichen Gesichtsanonymisierung in Echtzeit
CN111970509A (zh) * 2020-08-10 2020-11-20 杭州海康威视数字技术股份有限公司 一种视频图像的处理方法、装置与系统

Similar Documents

Publication Publication Date Title
WO2019017720A1 (fr) Système de caméra permettant la protection de la confidentialité et procédé correspondant
WO2019050360A1 (fr) Dispositif électronique et procédé de segmentation automatique d'être humain dans une image
WO2019172546A1 (fr) Appareil électronique et son procédé de commande
WO2021045599A1 (fr) Procédé d'application d'effet bokeh sur une image vidéo et support d'enregistrement
CN109218663B (zh) 图像处理设备、图像处理方法和存储介质
WO2020130309A1 (fr) Dispositif de masquage d'image et procédé de masquage d'image
WO2022114731A1 (fr) Système de détection de comportement anormal basé sur un apprentissage profond et procédé de détection pour détecter et reconnaître un comportement anormal
WO2019225964A1 (fr) Système et procédé de détection rapide d'objets
WO2022225102A1 (fr) Ajustement d'une valeur d'obturateur d'une caméra de surveillance par le biais d'une reconnaissance d'objets basée sur l'ia
WO2022075772A1 (fr) Procédé et dispositif de retouche d'image
WO2021006482A1 (fr) Appareil et procédé de génération d'image
WO2022225374A2 (fr) Procédé et appareil de reconstruction d'image de visage à l'aide d'un réseau de clarification d'identité vidéo
EP3707909A1 (fr) Dispositif électronique et procédé de correction d'images à l'aide d'un dispositif électronique externe
WO2023018084A1 (fr) Procédé et système de capture et de traitement automatiques d'une image d'un utilisateur
WO2022225105A1 (fr) Élimination de bruit pour une image de caméra de surveillance au moyen d'une reconnaissance d'objet basée sur l'ia
WO2020080734A1 (fr) Procédé de reconnaissance faciale et dispositif de reconnaissance faciale
WO2024167173A1 (fr) Stabilisation de multimédia à trame complète dans des dispositifs électroniques
WO2013047954A1 (fr) Dispositif et procédé de capture d'image pour stabiliser des images par utilisation d'un mouvement global obtenu à partir d'un point caractéristique dans l'arrière-plan
WO2023172031A1 (fr) Génération d'images de surveillance panoramique
WO2021025241A1 (fr) Procédé et dispositif pliable de construction d'un élément de données 3d
WO2023182794A1 (fr) Dispositif de contrôle de vision fondé sur une mémoire permettant la conservation de performances de contrôle, et procédé associé
WO2023140519A1 (fr) Système de construction d'une base de données d'apprentissage automatique à l'aide d'une technologie de protection d'informations personnelles de visage
WO2023080667A1 (fr) Traitement d'image wdr de caméra de surveillance par reconnaissance d'objets basée sur l'ia
WO2020067615A1 (fr) Procédé de commande d'un dispositif d'anonymisation vidéo permettant d'améliorer les performances d'anonymisation, et dispositif associé
WO2019198900A1 (fr) Appareil électronique et procédé de commande associé

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18834584

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18834584

Country of ref document: EP

Kind code of ref document: A1