[go: up one dir, main page]

WO2023281995A1 - Personal information masking method, and personal information masking device - Google Patents

Personal information masking method, and personal information masking device Download PDF

Info

Publication number
WO2023281995A1
WO2023281995A1 PCT/JP2022/023832 JP2022023832W WO2023281995A1 WO 2023281995 A1 WO2023281995 A1 WO 2023281995A1 JP 2022023832 W JP2022023832 W JP 2022023832W WO 2023281995 A1 WO2023281995 A1 WO 2023281995A1
Authority
WO
WIPO (PCT)
Prior art keywords
personal information
area
color image
image
exist
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/JP2022/023832
Other languages
French (fr)
Japanese (ja)
Inventor
久美生 大橋
義樹 山田
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
BITS Co Ltd
Original Assignee
BITS Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by BITS Co Ltd filed Critical BITS Co Ltd
Publication of WO2023281995A1 publication Critical patent/WO2023281995A1/en
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast

Definitions

  • the present invention relates to a personal information masking method and a personal information masking apparatus for masking a portion of personal information that identifies an individual captured in an image.
  • personal information information consisting of letters and numbers such as name, address, date of birth, gender, car registration number, etc.
  • personal information information consisting of letters and numbers such as name, address, date of birth, gender, car registration number, etc.
  • personal information is often imprinted. Such personal information may infringe on the privacy of individuals, and pose a problem of being used for crimes.
  • Patent Document 1 discloses that after an image is displayed on a display, it is determined whether or not there is personal information that the user does not want to share in the image from the viewpoint of protecting personal information. A technical idea is disclosed in which a user's determination result is accepted, and based on the determination result, a partial area of an image is masked so that personal information cannot be recognized. As a method for detecting personal information consisting of letters or numbers from an image, OCR (Optical Character Recognition) using pattern matching technology is used, as disclosed in Patent Document 2.
  • OCR Optical Character Recognition
  • Patent Document 1 it is up to the user to decide whether or not the personal information, etc. of Patent Document 1 exists.
  • OCR described in Patent Document 2 it is difficult to recognize characters or numbers such as resolution, vertical and horizontal arrangement, handwritten characters, and broken characters, especially small characters or numbers included in the image background. , was almost unrecognizable.
  • the present invention provides a personal information masking apparatus and method for erasing personal information that appears to be characters or numbers from a color image, and for outputting an image that does not give a sense of incongruity as a color image even if the personal information is erased. I will provide a.
  • a personal information masking method comprises a step of obtaining a color image, a size conversion step of converting the obtained color image into a first color image of a predetermined size, and a step of converting the first color image into a binary image. and a step. Furthermore, the masking method includes an area detection step of detecting an area in which personal information may exist in a binary image, a determination step of determining whether or not personal information exists in an area in which personal information is likely to exist, and Based on the personal information determined in the determining step and the first color image, an erasing step of drawing peripheral pixels of the personal information into the personal information of the first color image to erase the personal information, and a color image after the erasing step. and outputting.
  • the size conversion step converts the obtained color image into a first grayscale image of a predetermined size
  • the region detection step detects a region in which personal information may exist in the first grayscale image. good too.
  • a personal information masking method includes steps of obtaining a color image, a size conversion step of converting the obtained color image into a first color image and a first grayscale image of a predetermined size, and a first grayscale image. and an area detection step of detecting areas in the image where personal information may exist. Further, the masking method comprises a determination step of determining the presence or absence of personal information in an area in which personal information is likely to exist, and a pixel surrounding the personal information based on the personal information determined in the determination step and the first color image. An erasing step of drawing in personal information of one color image and erasing the personal information, and a step of outputting the color image after the erasing step are provided.
  • the masking method includes the steps of detecting a protected area of a person in the first color image, storing the protected area when the protected area is detected, and removing the protected area from the first color image after the erasing step. and an overwriting step of overwriting.
  • the masking method comprises the steps of denoising the first color image, processing a color space transformation on the first color image, and after the steps of denoising and processing the color space transformation, the first masking method. and a converting step of converting the color image into a second grayscale image.
  • the masking method further comprises processing a local histogram equalization on the second grayscale image and converting to a third grayscale image, wherein the region detection step comprises the second grayscale image or the third grayscale image. A region in which personal information may exist in the image may be detected.
  • the judgment process is (a) A filter for determining the presence or absence of personal information based on whether the logical product of the area where personal information is likely to exist and the area obtained by rotating the area by a predetermined angle is greater than a first threshold; (b) a filter for determining the presence or absence of personal information based on whether the standard deviation of the distance from the thin line that forms the framework of the area where personal information is likely to exist to the boundary line of the area is greater than a second threshold; (c) a filter that determines the presence or absence of personal information based on whether the vertical or horizontal length of an area where personal information is likely to exist is greater than a third threshold for a predetermined size; (d) A filter that determines the presence or absence of personal information based on whether the ratio of the area of the thin line that forms the framework of the area where personal information is likely to exist and the area of the area where personal information is likely to exist is greater than a fourth threshold; It is preferred to apply at least one filter treatment of
  • a personal information detection step may be provided for detecting personal information for the first color image.
  • a finishing step of finishing with an edge preserving filter may be provided.
  • the personal information masking apparatus includes an acquisition unit that acquires a first color image of a predetermined size, a binary conversion unit that converts the first color image into a binary image, and a personal information masking unit for the binary image.
  • an area detection step for detecting an area in which information may exist; a determination unit for determining the presence or absence of personal information in an area in which personal information is likely to exist; and the personal information determined by the determination unit and the first color.
  • An erasing section for erasing the personal information by drawing surrounding pixels of the personal information into the personal information of the first color image based on the image, and an outputting section for outputting the erased color image.
  • the personal information masking device further includes a protected area detection unit for detecting a protected area of a person in the first color image, the protected area detection unit comprising: i) the ability to detect facial regions using deep learning; ii) the ability to detect specific objects using deep learning; iii) the ability to detect skin areas by skin color in the first color image; iv) a function of setting a protection frame to be protected on the first color image and detecting an area within the protection frame; It is preferable to have at least one of
  • personal information means characters (including multilingual characters and numbers, hereinafter referred to as "characters") captured in images (still images/videos).
  • the personal information is erased by drawing the peripheral pixels of the personal information in the personal information of the color image, so that the personal information is protected.
  • FIG. 1 is a block diagram of one embodiment of a personal information masking device; FIG. It is a flow chart using a personal information masking device.
  • A is a diagram explaining color space conversion by a color space conversion unit.
  • B is an image diagram of outline detection and outer frame detection by the area detection unit.
  • C is a diagram explaining a mechanism for erasing personal information.
  • 4 is a flow chart showing the function of detection of a protected area; 4 is a conceptual diagram when a user sets a protection area on the personal computer 23 or the mobile information terminal 25.
  • FIG. 10 is a flow chart showing a method of determining personal information (whether or not it is characters).
  • (A) is a diagram showing a state in which a character or graphic is rotated and superimposed on the original character or graphic.
  • (B) is a diagram showing a state in which a character or graphic is shifted by a fixed distance and rotated, and superimposed on the original character or graphic.
  • (A) is a diagram explaining the distance from the center line of a character or figure to the boundary line.
  • (B) is an image diagram of outline detection and outer frame detection by the center line of a character or figure and the center line area detection unit. It is an example of a color image acquired by the personal information masking device 100 . It is an example (Example 1) of the output color image. It is an example (Example 2) of the output color image. It is an example (Example 3) of the output color image. It is an example (Example 4) of the output color image.
  • a personal information masking apparatus detects characters, which are personal information, from color images such as everyday scenery captured by a camera/video camera.
  • OCR Optical Character Recognition
  • the personal information masking apparatus does not use OCR, it is not necessary to register all characters to be recognized in advance, and multilingual personal information can be deleted.
  • the personal information masking device can delete fine characters or corrupted characters that cannot be detected by an OCR (optical character reader), such as fine personal information contained in the background.
  • FIG. 1 shows an outline of a system in which a user H uploads images (still images and moving images) taken at school, for example, to a personal information masking device 100, and downloads or receives images from which personal information has been deleted from the personal information masking device 100. It is a diagram showing. Even if the user H uploads the image from which the personal situation has been deleted to an SNS or the like, the personal information within the school included in the image has been deleted, so the personal information will not be leaked.
  • User H captures a still image of a predetermined place in the school with the digital camera 21, for example.
  • a still image captured by the camera 21 is recorded in the personal computer 23 .
  • User H can use the personal computer 23 to upload a still image to the personal information masking device 100 via a communication network NET such as the Internet.
  • the user H can take a moving image using the portable information terminal 25 with a camera function such as a tablet PC or a smartphone, and upload the moving image to the personal information masking device 100 via the communication network NET.
  • the user H may register a user name, password, credit card number, etc. in advance.
  • the personal information masking device 100 depicted in FIG. 1 shows a block configuration of one embodiment.
  • a cloud server or the like is suitable for the hardware configuration of the personal information masking device 100. Specifically, it has one or more processors, one or more memories, a communication interface, and the like.
  • a color image acquisition unit, an image size conversion unit, and the like, which will be described below, are composed of a processor, a memory, a communication interface, programs executed by the processor, and the like.
  • the personal information masking device 100 includes a color image acquisition unit 101 that acquires a color image (still image or moving image) from an external device such as the personal computer 23 or the mobile information terminal 25, and an image size converter that converts the size of the acquired image into a predetermined size.
  • the image size conversion unit 102 also has a function of converting a color image before removing color noise into a gray scale.
  • the personal information masking apparatus 100 includes a grayscale conversion unit 106 that converts the noise-removed color image into a grayscale image, a local histogram flattening unit 107 that flattens the brightness of the grayscale image, and a flattening of the brightness.
  • a binary conversion unit 108 that converts the converted grayscale image into a binary image
  • a noise removal unit (binary) 109 that removes black and white noise included in the binary image
  • personal information is extracted from the binary image or the grayscale image.
  • an area detection unit 110 for detecting an area in which is likely to exist. The area detection unit 110 detects boundaries and outlines due to line drawings, shadows of objects, noise, etc.
  • the area detection unit 110 can detect fine characters that are difficult for the human eye to notice, and can be applied to characters of any language, not limited to a specific language. On the other hand, the area detection unit 110 also detects shadows of objects other than characters or noise as areas in which personal information is likely to exist.
  • the personal information masking apparatus 100 includes a personal information determination unit 111 that determines whether personal information exists in an area where personal information is likely to exist, and a personal information determination unit 111 that erases the personal information determined from the color image converted to a predetermined size.
  • Protected area overwriting for overwriting a protected area saved when a person's face, skin or clock is detected by the information erasing section 112 and the protected area detecting section 103 into a color image from which personal information has been erased.
  • a finishing unit 114 for finishing the color image; and a color image output unit 115 for outputting the color image to the outside such as the personal computer 23 or the portable information terminal 25 .
  • the personal information masking apparatus 100 may also include a personal information detection unit 116 such as an in-scene character detection program (EAST: An Efficient and Accurate Scene Text Detector) using a neural network.
  • EAST An Efficient and Accurate Scene Text Detector
  • the personal information masking device 100 may not only have the plurality of components described above physically at the same location, but may also be a collection of a plurality of servers. Also, a part of the plurality of configurations described above may be performed by the personal computer 23 or the portable information terminal 25 to which the application software is downloaded.
  • FIG. 2 is a flow chart of the personal information masking method from acquisition of a color image to output of a color image from which personal information has been erased.
  • the color image acquisition unit 101 acquires a color image from the personal computer 23 or the mobile information terminal 25 (step S21).
  • this color image includes HD (1280*720), FHD (1920*1088) and 6M wide (3264*1836) for still images, and HD (1280*720) and FHD (1920) for moving images. *1088) and 4K (3840*2160) images.
  • the image size conversion unit 102 converts color images of various sizes into images of a predetermined size (S22). For example, in this embodiment, the image size conversion unit 102 converts the image into an FHD size color image. The color image converted to the predetermined size is stored in a memory (not shown) or the like. The image size conversion unit 102 also has a function of converting a color image into grayscale.
  • the protected area detection unit 103 detects a protected area included in the color image of a predetermined size (S23). Also, when no protected area is detected, it is preferable to add a flag indicating that there is no protected area to the color image.
  • the protection area detection function will be described later.
  • the color space conversion unit 105 converts the color space so that the personal information buried in colors can be easily seen (S24).
  • Color image expression methods include the RGB method, which expresses color images using the three primary colors of light, and the HSV method, which expresses color images using hue, saturation, and value.
  • a color space conversion unit 105 converts a color image (RGB) of a predetermined size into a color image (HSV) and erases a dark background color to highlight personal information.
  • FIG. 3(A) shows a color photograph (A-1) with a name written in black on a red cap, a grayscale image of the color photograph (A-2), and a saturation (S) of zero.
  • FIG. 1 shows a color photograph (A-1) with a name written in black on a red cap, a grayscale image of the color photograph (A-2), and a saturation (S) of zero.
  • FIG. 10 is a diagram showing an image (A-3);
  • the color photograph (A-1) in which personal information is drawn in black against a dark background color (especially red, blue, gray, etc.), is converted to grayscale or binary, the black personal information becomes a dark background color. This makes it difficult to detect (see A-2).
  • the image (A-3) with the saturation (S) set to zero the personal information stands out and stands out.
  • noise is removed from the color image by the noise removal unit (color) 104 (S25). Since the area detection unit 110 detects a line drawing, a shadow of an object, noise, or the like as an area in which personal information is likely to exist and the amount of calculation increases, it is preferable to remove noise contained in the color image as much as possible.
  • the noise included in the color image is, for example, if the subject is a leather surface or the like, the area detection unit 110 detects the texture of the leather surface or the like as a shadow of an object, etc. Roughness on the surface of the object is also processed as noise contained in the color image.
  • the noise removal unit (color) 104 of the present embodiment preferably applies edge-preserving smoothing filter processing (bilateral filter, mean shift filter, adaptive bilateral filter).
  • the grayscale conversion unit 106 converts the color space-converted image into a grayscale image (S26).
  • the local histogram flattening unit 107 flattens the grayscale image so that personal information buried in the dark/bright areas of the grayscale image can be easily detected (S27). For example, if a grayscale image has a bright object on an overall dark background, if the overall contrast is flattened, the bright object will become pure white, and the personal information present in the bright object will disappear. Sometimes I end up Therefore, applying the local histogram equalization process to the grayscale image makes it easier to detect personal information buried in dark and bright areas.
  • Local histogram equalization specifically includes adaptive histogram equalization, contrast-limited adaptive histogram equalization (CLAHE), multi-peak histogram equalization (MPHE), and multi-objective beta-optimized bihistogram equalization. (MBOBHE) processing and the like.
  • CLAHE contrast-limited adaptive histogram equalization
  • MPHE multi-peak histogram equalization
  • MBOBHE multi-objective beta-optimized bihistogram equalization.
  • the binary conversion unit 108 converts the grayscale image whose brightness has been flattened into a binary image (S28). If the background of the binary image is dark and the personal information is drawn in black, the background and the personal information will have the same value if the binary conversion is performed with a specific threshold value. area cannot be detected. Therefore, the binarization conversion unit 108 binarizes the grayscale image using the average value of the entire grayscale image (Otsu's binarization process (Discriminant Analysis method)), or for each part of the grayscale image (comparison Adaptive Gaussian Thresholding or Adaptive Mean Thresholding is preferably used to make dark black characters stand out.
  • a noise removal unit (binary) 109 removes noise remaining in the binary image. Specifically, the noise removal unit (binary) 109 repeats erosion for removing pixels on the boundary of the object in the binary image and dilation for adding pixels to the boundary. to remove binary noise.
  • the area detection unit 110 detects contours in which personal information is likely to exist from the binary image (S30). Contours in which personal information is likely to exist are detected by detecting contours such as lines in a binary image. In general, the area detector 110 scans the binary image for pixels of the object and determines whether it is an outer boundary or a hole boundary. By scanning and repeating the determination, the outermost boundary or contour is detected. for example.
  • FIG. 3B-1 is an image diagram in which a dashed-dotted line contour 50 is detected in a line drawing included in a binary image. More details are disclosed in the paper Topological structural analysis of digitized binary images by border following; by Satoshi Suzuki (Computer Vision, Graphics, and Image Processing Volume 30, Issue 1, April 1985, Pages 32-46).
  • the area detection unit 110 detects an outer frame in which personal information is likely to exist from the grayscale image (S31). Any of the grayscale images generated in steps S22, S26 or S27 may be used as the grayscale image. Schematically, the area detection unit 110 is used for area division of a grayscale image, in which areas having similar luminance values in the grayscale image are grouped into one area. Based on the idea of grouping together areas with similar luminance values, it is interpreted as an outer frame with a stable distribution, and its representative points are derived as feature quantities. for example.
  • FIG. 3B-2 is an image diagram in which the outer frame 51 of the two-dot chain line is detected in the line drawing included in the grayscale image.
  • 3B-2 is particularly different from FIG. 3B-1 in that the inside of "0" is also detected as an outer frame. More details are disclosed in the paper Efficient Maximally Stable Extremal Region (MSER) Tracking; by M. Donoser et al (2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)).
  • MSER Maximally Stable Extremal Region
  • the personal information detection unit 116 detects personal information (characters) by applying an in-scene character detection program (EAST: An Efficient and Accurate Scene Text Detector) or the like to the color image of a predetermined size generated in step S22.
  • EAST An Efficient and Accurate Scene Text Detector
  • the personal information detection unit 116 can detect relatively large characters and the like included in the color screen, but it is difficult to detect small characters and the like included in the background.
  • the personal information judging unit 111 judges whether or not there is personal information, that is, letters and numbers, in the outline or outer frame area that is likely to contain personal information detected in steps S30 and S31 (S33). . Also for characters detected by the personal information detection unit 116 . Similarly, the personal information determining unit 111 may determine whether or not there is personal information, that is, characters (S33). A method for determining personal information will be described later with reference to FIG.
  • a route for determining whether personal information exists from a binary image (S29, S30, S33) and a route for determining whether personal information exists from a grayscale image (S22 (or S26, S27) , S31, S33) and a route (S22, S32, S33) for determining whether personal information exists from the color image and personal information (characters) is stored in a memory (not shown) in each route. Then, the personal information (characters) stored in each route is integrated to obtain the personal information contained in the color image.
  • the present invention is not limited to this, and at least one route out of these three routes may be applied, or two routes out of the three routes may be selected and integrated.
  • FIG. 3(C) is a color photograph (C-1) in which personal information (a linear flaw in FIG. 3(C)) is drawn on a color image of a predetermined size.
  • Data (C-2) in which the personal information determined in step S33 is outlined on a black background, and a color image (C-3) in which the personal information is inpainted with surrounding pixels are shown. It is a diagram.
  • the personal information erasing unit 112 prepares a color image (C-1) of a predetermined size and an image (C-2) of the same size in which the personal information is outlined against a black background and erases the personal information.
  • the personal information is erased (C-3) by gradually drawing in the personal information with surrounding pixels from the boundary toward the inside (Inpainting).
  • step S34 it is preferable to devise ways to reduce the amount of calculation.
  • the personal information erasing unit 112 converts an image of a predetermined size (eg, FHD) into a reduced size image (eg, HD). Based on the color photograph (C-1), the white data (C-2), and the reduced size, pixels surrounding the personal information are drawn in (inpainting). Then, the personal information erasing unit 112 creates a reduced-size image (C-3) from which the personal information has been erased.
  • a predetermined size eg, FHD
  • HD reduced size image
  • the reduced size image (C-3) from which the personal information has been deleted is returned to a predetermined size, and the color image (C-1) of the predetermined size is blanked with the image (C-3) restored to the predetermined size.
  • the image at position (C-2) should be overwritten.
  • Inpainting inpainting
  • FMM Fast Marching Method
  • Another method is to draw pixels by searching along edges from pixel values in a known area around the personal information to the area of the personal information. More details are disclosed in the paper Navier-stokes, fluid dynamics, and image and video inpainting; by Marcelo Bertalmio et al (Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.).
  • the protected area overwrite unit 113 overwrites the saved protected area in the color image from which the personal information has been deleted (S35).
  • the area detection unit 110 detects that areas in which personal information is likely to exist include, for example, the temples of eyeglasses and the shadow of the nose. is disturbed. Therefore, the protected area is protected by overwriting the saved protected area with the color image from which the personal information has been deleted. If no protected area is detected in step S23, step S35 may be skipped based on the flag.
  • the finishing section 114 performs finishing processing on the color photograph in which the protected area has been overwritten (S36). Specifically, a pattern preserving filter (non-local mean filter) that suppresses smoothing of luminance values across boundaries or Gaussian blurring is used for color images.
  • the boundary with the overwritten protected area can be made inconspicuous, and minute personal information (characters) that could not be erased can be made illegible. There is no problem even if there is no finishing treatment.
  • the color image output unit 115 outputs the finished color image to the personal computer 23 or the mobile information terminal 25 (S37).
  • the protection area detection unit 103 uses an AI capable of deep learning, etc., based on the color image that has been changed to a predetermined size, from the features of the human face such as the positions of the eyes, nose, and mouth. , to detect the face area.
  • the face area is a protected area that does not contain personal information (characters).
  • the face area is saved in step S235 and overwritten in the color image from which the personal information has been erased as described in S35.
  • the protected area detection unit 103 uses AI or the like to label the objects included in the color image. Therefore, when a color image includes a clock, calendar, or table, the protected area detection unit 103 outputs a label such as "clock", "wall calendar", or “table". For example, clocks and calendars have time and date numbers (characters) drawn on them, but objects that contain characters but do not contain personal information, clocks, etc., are registered in advance. back. A clock or the like is detected as a protected area, and the protected area of the clock or the like is saved in step S235.
  • step S233 the protected area detection unit 103 detects whether the color image resized to the predetermined size has skin color. For example, the spaces (shadows) between the fingers are prevented from being recognized as a single line character (for example, "I" of the alphabet).
  • the skin area is a protected area that does not contain personal information, and the skin area is saved in step S235.
  • the protected area detection unit 103 may detect the color of the skin using, for example, color image display methods RGB, HSV, and YCbCr for a color image (RGB) of a predetermined size. More details are disclosed in the paper Human Skin Detection Using RGB, HSV and YCbCr Color Models; by S. Kolkur et al (ICCASP/ICMMD-2016. Published by Atlantic Press).
  • step S234 the protected area detection unit 103 displays a protected frame so that the user H can set a protected area on the screen of the personal computer 23 or the mobile information terminal 25.
  • FIG. 5 is an example of an uploaded color image 51 displayed on the screen of the mobile information terminal 25, and the color image 51 is part of a school classroom.
  • the classroom there are two desks and one shelf, the desk has a personal computer on which character information is displayed, and the shelf has a plurality of books and files.
  • the clock has a dial and the placard has the words "Congratulations on winning".
  • the protected area detection unit 103 causes the portable information terminal 25 of FIG.
  • the frame 57 is displayed on the screen.
  • the user H changes the size and position of the frame 57, for example, sets a frame 57a around the placard and sets a frame 57b at the bottom of the shelf.
  • User H clicks the completion button 54 when the setting of the frame 57 is completed.
  • the area set by the frame 57a or 57b in this manner is detected as a protected area, and the protected area is saved in step S235.
  • the protected area is overwritten in the color image from which the personal information has been erased. Note that even if the clock on the wall of the classroom in FIG.
  • the clock is made a protected area when step 232 is executed.
  • FIG. 5 has been described on the premise of a still image, the protection area can also be set for a moving image. For example, a frame area designated by the user H is held as a template image, and template matching processing is performed to check areas in the moving image that match the template image.
  • steps S231 to S234 may be changed for all protected areas, or only one of the steps may be executed, such as detecting only the face area.
  • the personal information area detection unit 110 detects the area of personal information, and the detection area detected by the personal information detection unit 111 includes a protection area such as a face area or a skin area. You may detect whether it exists.
  • the personal information determination unit 111 determines whether or not the area ratio between the area of the estimated character rotated and superimposed on the original estimated character and the area of the original estimated character is greater than a certain threshold k1 ( S331). If the area ratio is larger than the threshold k1, it is determined to be a figure (other than a character) and the process proceeds to step S336.
  • Fig. 7(A) shows an example of rotating the estimated characters.
  • alphabetic characters "C", “O” and “Z” and figures of perfect circles, squares and equilateral triangles are used as examples of estimated characters.
  • FIG. 7A from the left, 0° (not rotated), 90° rotation, superposition of 0° estimated character and 90° rotated estimated character, 180° rotation, 0° rotation Superposition of the estimated character and the estimated character rotated by 180°, superposition of the estimated character rotated by 270°, and the estimated character rotated by 0° and the estimated character rotated by 270° are shown. Note that the rotation is performed around the position of the center of gravity of the character area or figure area.
  • the personal information determination unit 111 supports not only the alphabet but also multiple languages. It has a threshold value k13 for superposition of ° and 270°, respectively, and it is determined whether or not the estimated character is a character by combining the superposition of the three rotation angles. If the personal information determination unit 111 performs character determination on an estimated character of a certain specific language, for example, only the threshold value k11 for overlapping 0° and 90° may be used. Moreover, it is preferable to make the threshold k1 variable, and it is preferable to make the threshold k1 value variable for each language.
  • the overlapping area of "C" is about 80%, and the overlapping area of the equilateral triangle is about 50%. It's getting smaller. Since it is difficult to determine whether the estimated character is a character, the estimated character may be rotated as shown in FIG. 7B.
  • an alphabet "C”, a perfect circle, and an equilateral triangle are examples of estimated characters.
  • the rotation is performed around the position of the center of gravity of the character area or figure area.
  • the overlap area of the "C" is about 10%
  • the overlap area of the new circle is about 70%
  • the overlap area of the equilateral triangle is about 50%.
  • Whether or not the estimated character is a character can be determined by checking whether or not the overlap region of the estimated character at 0 degrees and the estimated character shifted by the distance S and rotated by 180 degrees is larger than the threshold value k12.
  • a technique of shifting and rotating 180°, rotating 90°, and rotating 270° may be combined.
  • step S332 the personal information determination unit 111 calculates the standard deviation of the distance from the center line of the estimated character to the boundary line of the estimated character, compares the standard deviation with a threshold value k2, and determines that the character is a character. If the standard deviation is larger than the threshold k2, it is determined to be a figure or the like, and the process proceeds to step S336. Although it has already been determined in step S331 that the character is a character, in step S332 another method may be used to determine whether the estimated character is a character.
  • FIG. 8(A) is an example showing the alphabet "J” and a trapezoidal figure similar to J.
  • the distance L1 from the center line (framework) 61 of "J" to the boundary line 63 is almost constant and its standard deviation is small.
  • J is Gothic, but the standard deviation is small even in other typefaces such as New Century.
  • the distance L2 from the center line (framework) 61 of the trapezoidal figure to the boundary line 63 varies greatly, and its standard deviation increases.
  • the threshold k2 may be set to 0.6.
  • step S333 the personal information determination unit 111 calculates the area ratio of the area of the estimated character itself and the area of the entire area surrounding the estimated character, and determines that it is a character if it is smaller than the threshold value k3. If it is larger than the threshold k3, it is determined to be a figure or the like and the process proceeds to step S336. If it is smaller than the threshold k3, it is determined to be a character and the process proceeds to step S334. Although it has already been determined that the character is a character in steps S331 and S332, another method may be used to determine whether the estimated character is a character in step S333.
  • FIG. 8(B) is a specific example of step S333, showing an alphabet "J" and a spoon-shaped figure similar to J.
  • the area ratio of the area 66 of the J character itself to the area 65 of the entire area surrounding J is about 60%
  • the area ratio with 65 is about 80%.
  • the threshold k3 may be set to 70%.
  • steps S331 to S333 may be changed, or it may be determined that the estimated character is a character in only one of the steps.
  • steps S334 and S335 it is determined whether or not the character is personal information because it was determined in step S331, S332 or S333 that the character is a character.
  • step S334 it is determined whether or not the estimated character is larger than a certain threshold k4 for an image of a predetermined size. If it is larger than the threshold k4, it is determined that the character is large, and the process proceeds to step S338. If it is smaller than the threshold k4, it is determined that the character is small, and the process proceeds to step S336. More specifically, when the image size is FHD (1920*1088), at least one of the estimated characters has a vertical pixel count or a horizontal pixel count of, for example, 5 percent (96*54) or more of the image size. It is determined whether or not. If larger, the process proceeds to step S339.
  • the percentage of the image size is preferably variable.
  • a function that can change the threshold value k4 may be displayed on a web screen or the like for uploading a color image so that the user can determine the size of characters that the user wants to leave.
  • step S335 it is determined whether or not the character is drawn with a thick line width greater than a certain threshold value k5. If it is larger than the threshold k5, it is determined that the information is not personal information, and the process proceeds to step S338.
  • the threshold k5 is also variable.
  • a function that can change the threshold value k5 may be displayed on a web screen or the like for uploading a color image so that the user can determine the characters with a thick line width that the user wants to leave.
  • FIG. 8(C) is a specific example of step S335, showing a thick line width "I” and a thin line width "I” of the alphabet.
  • the area of the "I” character itself 68 and the area of the three centerlines (framework) 67 of the "I” are calculated, and it is determined whether the ratio of the areas is greater than a threshold k5, for example 10.
  • step S336 of FIG. 6 the estimated characters determined as not being characters in steps S331, S332, or S333 are considered to be graphics or the like, and are therefore left as part of the color image.
  • step S337 it is determined that the estimated characters are characters in step S331 or S332, and that they are neither large characters nor thick line width characters in step S334 or S335. Therefore, these characters are deleted from the color image as personal information. subject to
  • step S338 the estimated characters are determined to be characters in step S332 or S333, and are determined to be large characters or characters with a thick line width in step S334 or S335.
  • the "Graduation Ceremony" signboard hanging at the entrance of the school and the letters drawn in thick lines on the T-shirt are preserved in the color image.
  • large characters or characters with a thick line width are not treated as personal information, even if they are names, addresses, dates of birth, sex, numbers on license plates, and the like.
  • FIG. 9 is a color image acquired by the personal information masking device 100 from the personal computer 23.
  • FIG. 10, 11, 12 and 13 are color images output to the personal computer 23 by the personal information masking device 100.
  • FIG. 10 is a color image acquired by the personal information masking device 100 from the personal computer 23.
  • FIG. 10, 11, 12 and 13 are color images output to the personal computer 23 by the personal information masking device 100.
  • FIG. 10 is a color image acquired by the personal information masking device 100 from the personal computer 23.
  • FIG. 10, 11, 12 and 13 are color images output to the personal computer 23 by the personal information masking device 100.
  • the color image shown in FIG. 9 includes license plates of more than 10 cars, 8 languages (English, Thai, Arabic, Cyrillic, Japanese, Georgian, Korean, Burmese) drawn on the car. ) language characters, English characters such as news titles, and reporter faces.
  • FIG. 10 is an output example of Example 1, and Example 1 is an example of processing a color image without using steps S31 and S32 of FIG. That is, FIG. 10 is a color image obtained by executing steps S21-S30 and S33-S37 in FIG. In step S23, only S231 (face area protection) in FIG. 3 is executed.
  • FIG. 11 is an output example of Example 2, in which Steps S24-30 and S32 of FIG. 2 are not executed. That is, FIG. 11 is a color image obtained by performing step S31 on the grayscale image converted in step S22. In step S23, only S231 (face area protection) in FIG. 3 is executed.
  • FIG. 12 is an output example of Example 3.
  • Example 3 all the flowcharts of FIG. 2 are executed except for step S32. Therefore, FIG. 12 is equivalent to the photograph in which FIGS. 10 and 11 are superimposed.
  • step S23 only S231 (face area protection) in FIG. 3 is executed.
  • FIG. 13 is an output example of Example 4.
  • steps S24-30 and S32 are not executed as in Example 2, and the grayscale image converted in step S22 is executed in step S31. This is a color image.
  • step S23 only S234 (setting of protected area) in FIG. 3 is executed, and the protected area set by user H is the reporter frame in which the reporter is captured.
  • the personal information masking device of this embodiment processes everything from acquiring a color image to outputting a color image from which personal information has been erased.
  • the personal computer 23 or the mobile information terminal 25 may be responsible for part of the processing.
  • the personal computer 23 or the mobile information terminal 25 downloads an application from the cloud server in advance, and the personal computer 23 or the mobile information terminal 25 converts the color image into a predetermined size and protects the protected area such as the face area. may be detected and stored before being uploaded to the personal information masking device 100 .
  • a color image may be output from the personal information masking apparatus 100 and the protected area may be overwritten by the personal computer 23 or the portable information terminal 25 .
  • the personal computer 23 or the portable information terminal 25 may be responsible for all software processing of the personal information masking device 100 .
  • personal computer 25 portable information terminal 100 personal information masking device 101 color image acquisition unit 102 image size conversion unit 103 protected area detection unit; 104 Noise remover (color) 105 color space converter 106 grayscale converter 107 local histogram equalizer 108 binary converter 109 noise remover (binary) 110 Area detection unit 111 Personal information determination unit 112 Personal information deletion unit 113 Protected area overwrite unit 114 Finishing unit 115 Color image output unit 116 Personal information detection unit

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Closed-Circuit Television Systems (AREA)
  • Studio Devices (AREA)
  • Image Analysis (AREA)

Abstract

[problem] To provide a personal information masking method which erases personal information of characters or numbers from a color image. [Solution] This personal information masking method comprises: a step (S21) for acquiring a color image; a size change step (S22) for changing the acquired color image to a first color image of a prescribed size; a step (S28) for changing the first color image to a binary value image; an area detection step (S30) for detecting, from the binary image, an area in which personal information is possibly present; a decision step (S33) for deciding whether the personal information is present with respect to the area in which the personal information is likely to be present; and an erasing step (S34) for drawing surrounding pixels of the personal information on the personal information of the first color image and erasing the personal information.

Description

個人情報マスキング方法、及び個人情報マスキング装置Personal information masking method and personal information masking device

  本発明は、画像に写し込まれた個人を特定する個人情報の箇所にマスキングを付す個人情報マスキング方法、及び個人情報マスキング装置に係る。 The present invention relates to a personal information masking method and a personal information masking apparatus for masking a portion of personal information that identifies an individual captured in an image.

  近年、ユーザが、企業もしくは学校等での活動報告をクラウド上のSNS(ソーシャルネットワークサービス)に報告したり、人物を含む風景をSNSに投稿したりすることが多い。その活動報告や投稿には、テキストデータの文章だけでなく、静止画もしくは動画の画像情報もSNSにアップロードされることも多い。   In recent years, users often report activity reports at companies, schools, etc. to SNSs (social network services) on the cloud, and post scenery including people to SNSs. In the activity reports and posts, not only sentences of text data but also image information of still images or moving images are often uploaded to the SNS.

  しかしながら、その画像(静止画・動画)の中に、個人情報(名前、住所、生年月日、性別、自動車登録番号等の文字及び数字からなる情報、以下、本明細書では文字及び数字からなる情報を個人情報という。)が写し込まれることも多い。このような個人情報は、個人のプライバシーを侵害する恐れがあり、また犯罪に利用されてしまう問題を生じている。 However, in the image (still image / video), personal information (information consisting of letters and numbers such as name, address, date of birth, gender, car registration number, etc.) information is called personal information.) is often imprinted. Such personal information may infringe on the privacy of individuals, and pose a problem of being used for crimes.

  このような問題を生じないようにするために、特許文献1には、ディスプレイに画像が表示された後、個人情報の保護の観点から画像に共有したくない個人情報が存在するか否かをユーザの判定結果を受け付け、その判定結果に基づいて、個人情報を認知できないように、画像の一部領域をマスキングするように構成する技術思想が開示されている。また文字もしくは数字からなる個人情報を画像から検出する手法は、特許文献2に開示されるように、パターンマッチング技術を用いたOCR(Optical Character Recognition)が使用されている。 In order to avoid such a problem, Patent Document 1 discloses that after an image is displayed on a display, it is determined whether or not there is personal information that the user does not want to share in the image from the viewpoint of protecting personal information. A technical idea is disclosed in which a user's determination result is accepted, and based on the determination result, a partial area of an image is masked so that personal information cannot be recognized. As a method for detecting personal information consisting of letters or numbers from an image, OCR (Optical Character Recognition) using pattern matching technology is used, as disclosed in Patent Document 2.

特開2020-156033号公報JP 2020-156033 A 特開2015-028735号公報JP 2015-028735 A

  しかしながら、特許文献1の個人情報等が存在するか否かをユーザの判断に委ねられており、個人情報等が多く含まれる場合には、ユーザに著しい負担が生じる。またユーザが精査しないと気が付かない個人情報等もある可能性があり、気が付かない場合には個人情報等についてはマスキングされず、個人情報等が保護されないという問題点があった。また、特許文献2に記載されているOCRでは解像度、縦横の配列、手書き文字、くずし文字などは文字もしくは数字を認識することが困難であり、特に画像背景に含まれている小さな文字もしくは数字は、ほとんど認識できなかった。 However, it is up to the user to decide whether or not the personal information, etc. of Patent Document 1 exists. In addition, there is a possibility that there may be personal information, etc., that the user does not notice unless the user carefully examines it. If the user does not notice, the personal information, etc. is not masked, and there is a problem that the personal information, etc. is not protected. In addition, in the OCR described in Patent Document 2, it is difficult to recognize characters or numbers such as resolution, vertical and horizontal arrangement, handwritten characters, and broken characters, especially small characters or numbers included in the image background. , was almost unrecognizable.

 また欧州では、個人情報やプライバシーの保護に関して、GDPR(General Data Protection Regulation)が施行され、欧州以外でも個人情報のさらなる保護規制が進められているため、個人情報を保護する要望が高くなっている。 In addition, in Europe, the GDPR (General Data Protection Regulation) has been enforced to protect personal information and privacy, and further regulations for the protection of personal information are being promoted outside of Europe, so there is a growing demand for the protection of personal information. .

  本発明は、上記の問題点を踏まえ、カラー画像から文字もしくは数字と思われる個人情報を消去するとともに、個人情報を消去してもカラー画像として違和感がない画像を出力する個人情報マスキング装置及び方法を提供する。 In view of the above problems, the present invention provides a personal information masking apparatus and method for erasing personal information that appears to be characters or numbers from a color image, and for outputting an image that does not give a sense of incongruity as a color image even if the personal information is erased. I will provide a.

  本実施形態に係る個人情報マスキング方法は、カラー画像を取得する工程と、取得したカラー画像を所定サイズの第1カラー画像に変換するサイズ変換工程と、第1カラー画像を二値画像に変換する工程とを備える。さらにマスキング方法は、二値画像に対して個人情報が存在する可能性がある領域を検出する領域検出工程と、個人情報が存在しそうな領域に対して個人情報の有無を判定する判定工程と、判定工程で判定された個人情報と第1カラー画像とに基づいて、個人情報の周辺画素を第1カラー画像の個人情報に描き入れ個人情報を消去する消去工程と、消去工程後のカラー画像を出力する工程と、を備える。
 このサイズ変換工程は、取得したカラー画像を所定サイズの第1グレースケール画像に変換し、領域検出工程は、第1グレースケール画像に対して個人情報が存在する可能性のある領域を検出してもよい。
A personal information masking method according to the present embodiment comprises a step of obtaining a color image, a size conversion step of converting the obtained color image into a first color image of a predetermined size, and a step of converting the first color image into a binary image. and a step. Furthermore, the masking method includes an area detection step of detecting an area in which personal information may exist in a binary image, a determination step of determining whether or not personal information exists in an area in which personal information is likely to exist, and Based on the personal information determined in the determining step and the first color image, an erasing step of drawing peripheral pixels of the personal information into the personal information of the first color image to erase the personal information, and a color image after the erasing step. and outputting.
The size conversion step converts the obtained color image into a first grayscale image of a predetermined size, and the region detection step detects a region in which personal information may exist in the first grayscale image. good too.

  別の実施形態に係る個人情報マスキング方法は、カラー画像を取得する工程と、取得したカラー画像を所定サイズの第1カラー画像及び第1グレースケール画像に変換するサイズ変換工程と、第1グレースケール画像に対して個人情報が存在する可能性のある領域を検出する領域検出工程と、を備える。さらにマスキング方法は、個人情報が存在しそうな領域に対して個人情報の有無を判定する判定工程と、判定工程で判定された個人情報と第1カラー画像とに基づいて個人情報の周辺画素を第1カラー画像の個人情報に描き入れ個人情報を消去する消去工程と、消去工程後のカラー画像を出力する工程と、を備える。 A personal information masking method according to another embodiment includes steps of obtaining a color image, a size conversion step of converting the obtained color image into a first color image and a first grayscale image of a predetermined size, and a first grayscale image. and an area detection step of detecting areas in the image where personal information may exist. Further, the masking method comprises a determination step of determining the presence or absence of personal information in an area in which personal information is likely to exist, and a pixel surrounding the personal information based on the personal information determined in the determination step and the first color image. An erasing step of drawing in personal information of one color image and erasing the personal information, and a step of outputting the color image after the erasing step are provided.

 さらにマスキング方法は、第1カラー画像に対して人物の保護領域を検出する工程と、保護領域を検出した場合に保護領域を保存する工程と、消去工程後の第1カラー画像に、保護領域を上書きする上書き工程と、を備えてもよい。
 さらにマスキング方法は、第1カラー画像に対してノイズ除去する工程と、第1カラー画像に対して色空間変換を処理する工程と、ノイズ除去する工程及び色空間変換を処理する工程の後に第1カラー画像を第2グレースケール画像に変換する変換工程と、を備えても良い。
 さらにマスキング方法は、第2グレースケール画像に対して局所的ヒストグラム平坦化を処理し、第3グレースケール画像に変換する工程、を備え、領域検出工程は、第2グレースケール画像又は第3グレースケール画像に対して個人情報が存在する可能性のある領域を検出しても良い。
Further, the masking method includes the steps of detecting a protected area of a person in the first color image, storing the protected area when the protected area is detected, and removing the protected area from the first color image after the erasing step. and an overwriting step of overwriting.
Further, the masking method comprises the steps of denoising the first color image, processing a color space transformation on the first color image, and after the steps of denoising and processing the color space transformation, the first masking method. and a converting step of converting the color image into a second grayscale image.
The masking method further comprises processing a local histogram equalization on the second grayscale image and converting to a third grayscale image, wherein the region detection step comprises the second grayscale image or the third grayscale image. A region in which personal information may exist in the image may be detected.

 判定工程は、
(a) 個人情報が存在しそうな領域とかかる領域を所定角度回転させた領域との論理積が第1閾値より大きいかに基づいて、個人情報の有無を判定するフィルター、
(b) 個人情報が存在しそうな領域の骨組みとなる細線から領域の境界線までの距離の標準偏差が第2閾値より大きいかに基づいて、個人情報の有無を判定するフィルター、
(c) 個人情報が存在しそうな領域の縦長さ又は横長さが所定サイズに対する第3閾値より大きいかに基づいて、個人情報の有無を判定するフィルター、
(d) 個人情報が存在しそうな領域の骨組みとなる細線の面積と個人情報が存在しそうな領域の面積との比が第4閾値より大きいかに基づいて、個人情報の有無を判定するフィルター、
 の少なくとも1つのフィリター処理を適用することが好ましい。
The judgment process is
(a) A filter for determining the presence or absence of personal information based on whether the logical product of the area where personal information is likely to exist and the area obtained by rotating the area by a predetermined angle is greater than a first threshold;
(b) a filter for determining the presence or absence of personal information based on whether the standard deviation of the distance from the thin line that forms the framework of the area where personal information is likely to exist to the boundary line of the area is greater than a second threshold;
(c) a filter that determines the presence or absence of personal information based on whether the vertical or horizontal length of an area where personal information is likely to exist is greater than a third threshold for a predetermined size;
(d) A filter that determines the presence or absence of personal information based on whether the ratio of the area of the thin line that forms the framework of the area where personal information is likely to exist and the area of the area where personal information is likely to exist is greater than a fourth threshold;
It is preferred to apply at least one filter treatment of

 第1カラー画像に対して個人情報を検出する個人情報検出工程を、備えても良い。
 上書き工程後に、エッジ保存フィルターにより仕上げ処理する仕上げ工程、を備えても良い。
A personal information detection step may be provided for detecting personal information for the first color image.
After the overwriting step, a finishing step of finishing with an edge preserving filter may be provided.

 本実施形態に係る個人情報マスキング装置は、所定サイズの第1カラー画像を取得する取得部と、第1カラー画像を二値画像に変換する二値変換部と、二値画像に対して、個人情報が存在する可能性がある領域を検出する領域検出工程と、個人情報が存在しそうな領域に対して個人情報の有無を判定する判定部と、判定部で判定された個人情報と第1カラー画像とに基づいて、個人情報の周辺画素を第1カラー画像の個人情報に描き入れ個人情報を消去する消去部と、消去後のカラー画像を出力する出力部と、を備える。 The personal information masking apparatus according to the present embodiment includes an acquisition unit that acquires a first color image of a predetermined size, a binary conversion unit that converts the first color image into a binary image, and a personal information masking unit for the binary image. an area detection step for detecting an area in which information may exist; a determination unit for determining the presence or absence of personal information in an area in which personal information is likely to exist; and the personal information determined by the determination unit and the first color. An erasing section for erasing the personal information by drawing surrounding pixels of the personal information into the personal information of the first color image based on the image, and an outputting section for outputting the erased color image.

 また個人情報マスキング装置は、第1カラー画像に対して人物の保護領域を検出する保護領域検出部をさらに備え、該保護領域検出部は、
  i)ディープラーニングを使用した顔領域の検出する機能、
 ii)ディープラーニングを使用した特定物体の検出する機能、
iii)第1カラー画像にある皮膚の色による皮膚領域の検出する機能、
 iv)第1カラー画像に保護すべき保護枠を設定させて、その保護枠内の領域を検出する機能、
 の少なくとも1つを有していることが好ましい。
The personal information masking device further includes a protected area detection unit for detecting a protected area of a person in the first color image, the protected area detection unit comprising:
i) the ability to detect facial regions using deep learning;
ii) the ability to detect specific objects using deep learning;
iii) the ability to detect skin areas by skin color in the first color image;
iv) a function of setting a protection frame to be protected on the first color image and detecting an area within the protection frame;
It is preferable to have at least one of

 本明細書において、個人情報とは、画像(静止画・動画)に撮影された文字(多言語の文字及び数字を含め、以下、“文字”という。)を意味する。 In this specification, personal information means characters (including multilingual characters and numbers, hereinafter referred to as "characters") captured in images (still images/videos).

  本発明の個人情報マスキング装置及び個人情報マスキング方法によれば、個人情報の周辺画素をカラー画像の個人情報に描き入れて個人情報を消去するので、個人情報が保護される。 According to the personal information masking device and the personal information masking method of the present invention, the personal information is erased by drawing the peripheral pixels of the personal information in the personal information of the color image, so that the personal information is protected.

個人情報マスキング装置の一実施形態のブロック図である。1 is a block diagram of one embodiment of a personal information masking device; FIG. 個人情報マスキング装置を使ったフローチャートである。It is a flow chart using a personal information masking device. (A)は色空間変換部による色空間変換について説明した図である。(B)は領域検出部による輪郭の検出、外枠の検出のイメージ図である。(C)は個人情報を消去する仕組みを説明した図である。(A) is a diagram explaining color space conversion by a color space conversion unit. (B) is an image diagram of outline detection and outer frame detection by the area detection unit. (C) is a diagram explaining a mechanism for erasing personal information. 保護領域の検出の機能を示したフローチャートである。4 is a flow chart showing the function of detection of a protected area; パーソナルコンピュータ23又は携帯情報端末25で保護領域をユーザが設定する際の概念図である。4 is a conceptual diagram when a user sets a protection area on the personal computer 23 or the mobile information terminal 25. FIG. 個人情報(文字であるか否か)の判定の手法を示したフローチャートである。10 is a flow chart showing a method of determining personal information (whether or not it is characters). (A)は文字又は図形を回転させて、元の文字又は図形に重ね合わせた状態を示した図である。(B)は文字又は図形を一定の距離をシフトさせて且つ回転させて、元の文字又は図形に重ね合わせた状態を示した図である。(A) is a diagram showing a state in which a character or graphic is rotated and superimposed on the original character or graphic. (B) is a diagram showing a state in which a character or graphic is shifted by a fixed distance and rotated, and superimposed on the original character or graphic. (A)は文字又は図形の中心線から境界線までの距離について説明した図である。(B)は文字又は図形の中心線と中心線領域検出部による輪郭の検出、外枠の検出のイメージ図である。(A) is a diagram explaining the distance from the center line of a character or figure to the boundary line. (B) is an image diagram of outline detection and outer frame detection by the center line of a character or figure and the center line area detection unit. 個人情報マスキング装置100に取得されるカラー画像の一例である。It is an example of a color image acquired by the personal information masking device 100 . 出力したカラー画像の一例(実施例1)である。It is an example (Example 1) of the output color image. 出力したカラー画像の一例(実施例2)である。It is an example (Example 2) of the output color image. 出力したカラー画像の一例(実施例3)である。It is an example (Example 3) of the output color image. 出力したカラー画像の一例(実施例4)である。It is an example (Example 4) of the output color image.

  以下、実施形態の個人情報マスキング装置について図を参照しながら詳しく説明する。なお、本明細書及び図面において、実質的に同一の機能構成を有する構成要素については、同一の符号を付することにより重複説明を省略する。   Hereinafter, the personal information masking device of the embodiment will be described in detail with reference to the drawings. In the present specification and drawings, constituent elements having substantially the same functional configuration are denoted by the same reference numerals, thereby omitting redundant description.

<<個人情報マスキング装置の概要>>
 本実施形態に係る個人情報マスキング装置は、カメラ・ビデオカメラで撮影された日常風景などのカラー画像から、個人情報である文字を検出する。画像から文字を認識する方法としてOCR(Optical Character Recognition)が広く使用されている。しかし本実施形態に係る個人情報マスキング装置は、OCRを使用しないため、予め認識したい文字をすべて登録しておく必要がなく、多言語の個人情報を削除することができる。また、個人情報マスキング装置は、OCR(光学式文字読み取り装置)では検出できない微細な文字もしくは崩れた文字、例えば背景に含まれる微細な個人情報を削除することができる。
<<Overview of personal information masking device>>
A personal information masking apparatus according to the present embodiment detects characters, which are personal information, from color images such as everyday scenery captured by a camera/video camera. OCR (Optical Character Recognition) is widely used as a method of recognizing characters from an image. However, since the personal information masking apparatus according to the present embodiment does not use OCR, it is not necessary to register all characters to be recognized in advance, and multilingual personal information can be deleted. In addition, the personal information masking device can delete fine characters or corrupted characters that cannot be detected by an OCR (optical character reader), such as fine personal information contained in the background.

 図1は、ユーザHが例えば学校で撮影した画像(静止画、動画)を個人情報マスキング装置100にアップロードし、個人情報マスキング装置100から個人情報が消去された画像をダウンロードもしくは受信するシステム概要を示した図である。ユーザHは、この個人状況が消去された画像をSNS等にアップロードしても、その画像に含まれる学校内の個人情報が消去されているため、個人情報の漏洩にならない。 FIG. 1 shows an outline of a system in which a user H uploads images (still images and moving images) taken at school, for example, to a personal information masking device 100, and downloads or receives images from which personal information has been deleted from the personal information masking device 100. It is a diagram showing. Even if the user H uploads the image from which the personal situation has been deleted to an SNS or the like, the personal information within the school included in the image has been deleted, so the personal information will not be leaked.

 ユーザHは例えばデジタルカメラ21で学校の所定の場所の静止画を撮像する。そしてカメラ21で撮影された静止画は、パーソナルコンピュータ23に記録される。ユーザHはそのパーソナルコンピュータ23を使って、インターネット等の通信ネットワークNETを介して、個人情報マスキング装置100に静止画をアップロードすることができる。またユーザHは、タブレット型PC又はスマートフォン等のカメラ機能付き携帯情報端末25を使って動画を撮影し、通信ネットワークNETを介して、個人情報マスキング装置100に動画をアップロードすることができる。特に説明しないが、個人情報マスキング装置100にアクセスするためには、ユーザHはユーザ名、パスワード、クレジットカード番号等を事前に登録するようにしてもよい。 User H captures a still image of a predetermined place in the school with the digital camera 21, for example. A still image captured by the camera 21 is recorded in the personal computer 23 . User H can use the personal computer 23 to upload a still image to the personal information masking device 100 via a communication network NET such as the Internet. Also, the user H can take a moving image using the portable information terminal 25 with a camera function such as a tablet PC or a smartphone, and upload the moving image to the personal information masking device 100 via the communication network NET. Although not specifically described, in order to access the personal information masking device 100, the user H may register a user name, password, credit card number, etc. in advance.

<<個人情報マスキング装置100の構成の概要>>
 図1に描かれた個人情報マスキング装置100は、一実施形態のブロック構成が示されている。個人情報マスキング装置100は、ハードウェア構成としては、クラウドサーバー等が適しており、具体的には1以上のプロセッサ、1以上のメモリー及び通信インターフェースなどを有している。以下に説明するカラー画像取得部、画像サイズ変換部等はプロセッサ、メモリー及び通信インターフェース並びにプロセッサで実行されるプログラム等で構成される。
<<Overview of Configuration of Personal Information Masking Device 100>>
The personal information masking device 100 depicted in FIG. 1 shows a block configuration of one embodiment. A cloud server or the like is suitable for the hardware configuration of the personal information masking device 100. Specifically, it has one or more processors, one or more memories, a communication interface, and the like. A color image acquisition unit, an image size conversion unit, and the like, which will be described below, are composed of a processor, a memory, a communication interface, programs executed by the processor, and the like.

  個人情報マスキング装置100は、パーソナルコンピュータ23又は携帯情報端末25などの外部からカラー画像(静止画もしくは動画)を取得するカラー画像取得部101と、取得した画像のサイズを所定サイズに変換する画像サイズ変換部102と、所定サイズに変換されたカラー画像中に人物の顔や時計等(保護領域)があるか検出する保護領域検出部103と、カラー画像に含まれるノイズを除去するノイズ除去部(カラー)104と、背景色に埋もれて見え難くなった個人情報を見やすくする色空間変換部105とを有している。なお画像サイズ変換部102は、カラーノイズを除去する前のカラー画像をグレースケールに変換する機能も有している。 The personal information masking device 100 includes a color image acquisition unit 101 that acquires a color image (still image or moving image) from an external device such as the personal computer 23 or the mobile information terminal 25, and an image size converter that converts the size of the acquired image into a predetermined size. A conversion unit 102, a protection area detection unit 103 that detects whether there is a person's face, clock, etc. (protection area) in the color image converted to a predetermined size, and a noise removal unit that removes noise contained in the color image ( color) 104, and a color space conversion unit 105 that makes it easier to see personal information that is hidden in the background color and is difficult to see. Note that the image size conversion unit 102 also has a function of converting a color image before removing color noise into a gray scale.

 さらに個人情報マスキング装置100は、ノイズ除去されたカラー画像をグレースケール画像に変換するグレースケール変換部106と、グレースケール画像の明度を平坦化する局所的ヒストグラム平坦化部107と、明度が平坦化されたグレースケール画像を二値画像に変換する二値変換部108と、二値画像に含まれる白黒ノイズを除去するノイズ除去部(二値)109と、二値画像もしくはグレースケール画像から個人情報が存在しそうな領域を検出する領域検出部110とを備える。領域検出部110は、二値画像もしくはグレースケール画像に存在する線画、物の影もしくはノイズなどによる境界や外郭を個人情報が存在しそうな領域として検出する。このため、領域検出部110は、人間の目では気づきにくい微細な文字も検出することができ、また、特定の言語に限らず、あらゆる言語の文字に適用できるようになっている。その一方で、領域検出部110は、文字ではない物の影もしくはノイズなども個人情報が存在しそうな領域として検出する。 Further, the personal information masking apparatus 100 includes a grayscale conversion unit 106 that converts the noise-removed color image into a grayscale image, a local histogram flattening unit 107 that flattens the brightness of the grayscale image, and a flattening of the brightness. A binary conversion unit 108 that converts the converted grayscale image into a binary image, a noise removal unit (binary) 109 that removes black and white noise included in the binary image, and personal information is extracted from the binary image or the grayscale image. and an area detection unit 110 for detecting an area in which is likely to exist. The area detection unit 110 detects boundaries and outlines due to line drawings, shadows of objects, noise, etc. existing in a binary image or a grayscale image as areas in which personal information is likely to exist. Therefore, the area detection unit 110 can detect fine characters that are difficult for the human eye to notice, and can be applied to characters of any language, not limited to a specific language. On the other hand, the area detection unit 110 also detects shadows of objects other than characters or noise as areas in which personal information is likely to exist.

 さらに個人情報マスキング装置100は、個人情報が存在しそうな領域に個人情報が存在するかを判定する個人情報判定部111と,所定サイズに変換されたカラー画像から判定された個人情報を消去する個人情報消去部112と、保護領域検出部103で人物の顔、皮膚又は時計などを検出した場合に保存しておいた保護領域を、個人情報が消去されたカラー画像に中に上書きする保護領域上書き部113と、カラー画像を仕上げする仕上げ部114と、カラー画像をパーソナルコンピュータ23又は携帯情報端末25などの外部に出力するカラー画像出力部115とを備える。また、個人情報マスキング装置100は、ニューラルネットワークを利用した情景内文字検出プログラム(EAST:An Efficient and Accurate Scene Text Detector)等の個人情報の検出部116を有しても良い。 Further, the personal information masking apparatus 100 includes a personal information determination unit 111 that determines whether personal information exists in an area where personal information is likely to exist, and a personal information determination unit 111 that erases the personal information determined from the color image converted to a predetermined size. Protected area overwriting for overwriting a protected area saved when a person's face, skin or clock is detected by the information erasing section 112 and the protected area detecting section 103 into a color image from which personal information has been erased. a finishing unit 114 for finishing the color image; and a color image output unit 115 for outputting the color image to the outside such as the personal computer 23 or the portable information terminal 25 . The personal information masking apparatus 100 may also include a personal information detection unit 116 such as an in-scene character detection program (EAST: An Efficient and Accurate Scene Text Detector) using a neural network.

  個人情報マスキング装置100は、上記複数の構成要素を物理的に同一の場所に有するだけでなく、複数のサーバの集合体であってもよい。また上記複数の構成の一部を、アプリケーションソフトをダウンロードしたパーソナルコンピュータ23もしくは携帯情報端末25に担わせても良い。   The personal information masking device 100 may not only have the plurality of components described above physically at the same location, but may also be a collection of a plurality of servers. Also, a part of the plurality of configurations described above may be performed by the personal computer 23 or the portable information terminal 25 to which the application software is downloaded.

<<個人情報マスキング方法のフローチャート>>
 次に、本実施形態の個人情報マスキング方法の動作を説明する。図2は個人情報マスキング方法のカラー画像の取得から、個人情報を消去したカラー画像を出力するまでのフローチャートである。
<<Flow chart of personal information masking method>>
Next, the operation of the personal information masking method of this embodiment will be described. FIG. 2 is a flow chart of the personal information masking method from acquisition of a color image to output of a color image from which personal information has been erased.

 まず、カラー画像取得部101が、パーソナルコンピュータ23もしくは携帯情報端末25からカラー画像を取得する(ステップS21)。このカラー画像には、例えば、静止画であればHD(1280*720)、FHD(1920*1088)、6Mワイド(3264*1836)等、動画であればHD(1280*720)、FHD(1920*1088)、4K(3840*2160)等の画像を取得する。 First, the color image acquisition unit 101 acquires a color image from the personal computer 23 or the mobile information terminal 25 (step S21). For example, this color image includes HD (1280*720), FHD (1920*1088) and 6M wide (3264*1836) for still images, and HD (1280*720) and FHD (1920) for moving images. *1088) and 4K (3840*2160) images.

 次に、画像サイズ変換部102が種々のサイズのカラー画像を所定のサイズの画像に変換する(S22)。例えば、本実施形態では画像サイズ変換部102はFHDサイズのカラー画像に変換する。この所定サイズに変換されたカラー画像は、不図示のメモリー等に保存される。画像サイズ変換部102は、カラー画像をグレースケールに変換する機能も有している。 Next, the image size conversion unit 102 converts color images of various sizes into images of a predetermined size (S22). For example, in this embodiment, the image size conversion unit 102 converts the image into an FHD size color image. The color image converted to the predetermined size is stored in a memory (not shown) or the like. The image size conversion unit 102 also has a function of converting a color image into grayscale.

 次に、保護領域検出部103が所定サイズのカラー画像に含まれる保護領域を検出する(S23)。また保護領域が検出されなかった場合にはそのカラー画像には保護領域が無いことを示すフラグが追加されることが好ましい。保護領域の検出の機能については、後述する。 Next, the protected area detection unit 103 detects a protected area included in the color image of a predetermined size (S23). Also, when no protected area is detected, it is preferable to add a flag indicating that there is no protected area to the color image. The protection area detection function will be described later.

 次に、色彩に埋もれてしまっている個人情報を見やすくするように、色空間変換部105が色空間変換する(S24)。カラー画像の表現方式には、光の三原色でカラー画像を表すRGB方式、色相(hue)、彩度(Saturation)及び明度(Value)でカラー画像を表すHSV方式がある。色空間変換部105は、所定サイズのカラー画像(RGB)をカラー画像(HSV)に変換するとともに、濃い背景色を消去することで個人情報を浮き立たせる。図3(A)は、赤色の帽子に黒字で名前が書いてあるカラー写真(A-1)と、そのカラー写真のグレースケール画像(A-2)と、彩度(S)をゼロにした画像(A-3)とを示した図である。濃い背景色(特に赤色、青色、灰色等)に黒色で個人情報が描かれているカラー写真(A-1)が、グレースケールもしくは二値に変換されると、黒色の個人情報は濃い背景色に埋もれて逆に検出しにくくなる(A-2を参照)。一方、彩度(S)をゼロにした画像(A-3)では、個人情報が浮き出て目立っている。 Next, the color space conversion unit 105 converts the color space so that the personal information buried in colors can be easily seen (S24). Color image expression methods include the RGB method, which expresses color images using the three primary colors of light, and the HSV method, which expresses color images using hue, saturation, and value. A color space conversion unit 105 converts a color image (RGB) of a predetermined size into a color image (HSV) and erases a dark background color to highlight personal information. FIG. 3(A) shows a color photograph (A-1) with a name written in black on a red cap, a grayscale image of the color photograph (A-2), and a saturation (S) of zero. FIG. 10 is a diagram showing an image (A-3); When the color photograph (A-1), in which personal information is drawn in black against a dark background color (especially red, blue, gray, etc.), is converted to grayscale or binary, the black personal information becomes a dark background color. This makes it difficult to detect (see A-2). On the other hand, in the image (A-3) with the saturation (S) set to zero, the personal information stands out and stands out.

 次に、カラー画像に対して、ノイズ除去部(カラー)104によりノイズを除去する(S25)。領域検出部110が、線画、物の影もしくはノイズなどを個人情報が存在しそうな領域として検出し計算量が増えるため、カラー画像に含まれるノイズをできるだけ除去することが好ましい。ここで、カラー画像に含まれるノイズとは、たとえば被写体が革表面等であると、領域検出部110が革表面等の質感(テクスチャー)も物の影等として検出するため、カラー画像に写った物体表面のざらつきもカラー画像に含まれるノイズとして処理される。 Next, noise is removed from the color image by the noise removal unit (color) 104 (S25). Since the area detection unit 110 detects a line drawing, a shadow of an object, noise, or the like as an area in which personal information is likely to exist and the amount of calculation increases, it is preferable to remove noise contained in the color image as much as possible. Here, the noise included in the color image is, for example, if the subject is a leather surface or the like, the area detection unit 110 detects the texture of the leather surface or the like as a shadow of an object, etc. Roughness on the surface of the object is also processed as noise contained in the color image.

 カラー画像に含まれるノイズを除去する方法としてガウシアンぼかし (Gaussian Blur)処理と呼ばれるガウス関数を用いてカラー画像全体をぼかす処理がある。しかし、ガウシアンぼかし処理は、カラー画像内に含まれる個人情報のエッジ部分まで平滑化されるため、個人情報を検出できなくなる可能性もある。このため本実施形態のノイズ除去部(カラー)104は、エッジ保持平滑化フィルター処理(バイラテラルフィルター、ミーンシフトフィルター、アダプティブバイラテラルフィルター)を適用することが好ましい。 As a method of removing noise contained in color images, there is a process called Gaussian Blur processing that uses a Gaussian function to blur the entire color image. However, since Gaussian blur processing smoothes even the edge portion of personal information contained in a color image, there is a possibility that personal information cannot be detected. Therefore, the noise removal unit (color) 104 of the present embodiment preferably applies edge-preserving smoothing filter processing (bilateral filter, mean shift filter, adaptive bilateral filter).

 次にグレースケール変換部106が、色空間変換された画像をグレースケール画像に変換する(S26)。
 引き続き、グレースケール画像の暗部・明部に埋もれた個人情報を検出しやすいように、局所的ヒストグラム平坦化部107がグレースケール画像を平坦化する(S27)。例えば、全体的に暗い背景に明るい対象物がグレースケール画像に存在する場合に、全体的にコントラストを平坦化すると、明るい対象物が真っ白になってしまい、明るい対象物に存在する個人情報が消えてしまうことがある。このため、局所的ヒストグラム平坦化の処理をグレースケール画像に適用することにより、暗部・明部に埋もれた個人情報を検出しやすいようにする。局所的ヒストグラム平坦化は、具体的には、適応ヒストグラム平坦化(adaptive histogram equalization)、コントラスト制限適応ヒストグラム平坦化(CLAHE)、マルチピークヒストグラム平坦化(MPHE)、および多目的ベータ最適化バイヒストグラム平坦化(MBOBHE)処理等がある。
Next, the grayscale conversion unit 106 converts the color space-converted image into a grayscale image (S26).
Subsequently, the local histogram flattening unit 107 flattens the grayscale image so that personal information buried in the dark/bright areas of the grayscale image can be easily detected (S27). For example, if a grayscale image has a bright object on an overall dark background, if the overall contrast is flattened, the bright object will become pure white, and the personal information present in the bright object will disappear. Sometimes I end up Therefore, applying the local histogram equalization process to the grayscale image makes it easier to detect personal information buried in dark and bright areas. Local histogram equalization specifically includes adaptive histogram equalization, contrast-limited adaptive histogram equalization (CLAHE), multi-peak histogram equalization (MPHE), and multi-objective beta-optimized bihistogram equalization. (MBOBHE) processing and the like.

 次に二値変換部108が、明度が平坦化されたグレースケール画像を二値画像に変換する(S28)。二値画像の背景が暗い場合に個人情報が黒色で描かれていると、特定の閾値で二値変換してしまうと背景と個人情報とが同値になってしまい、領域検出部110が個人情報の領域を検出できなくなる。そこで、二値化変換部108は、グレースケール画像の全体の平均値を利用して2値化する(大津の二値化処理(Discriminant Analysis method))、もしくはグレースケール画像の一部毎(比較的狭い領域内)の平均値を利用して2値化する(Adaptive Gaussian ThresholdingまたはAdaptive Mean Thresholding)適応的閾値処理で、暗部の黒文字を浮き立たせることが好ましい。 Next, the binary conversion unit 108 converts the grayscale image whose brightness has been flattened into a binary image (S28). If the background of the binary image is dark and the personal information is drawn in black, the background and the personal information will have the same value if the binary conversion is performed with a specific threshold value. area cannot be detected. Therefore, the binarization conversion unit 108 binarizes the grayscale image using the average value of the entire grayscale image (Otsu's binarization process (Discriminant Analysis method)), or for each part of the grayscale image (comparison Adaptive Gaussian Thresholding or Adaptive Mean Thresholding is preferably used to make dark black characters stand out.

 次にノイズ除去部(二値)109が、二値画像に残っているノイズを除去する。具体的には、ノイズ除去部(二値)109は、二値画像内の対象物の境界のピクセルを除去する収縮(Erosion)とその境界にピクセルを加える膨張(Dilation)とを繰り返すモルフォロジー変換して、二値ノイズを除去する。 Next, a noise removal unit (binary) 109 removes noise remaining in the binary image. Specifically, the noise removal unit (binary) 109 repeats erosion for removing pixels on the boundary of the object in the binary image and dilation for adding pixels to the boundary. to remove binary noise.

 次に、領域検出部110は二値画像から個人情報が存在しそうな輪郭を検出する(S30)。個人情報が存在しそうな輪郭は、二値画像に線などがある輪郭を検出することである。概略的には、領域検出部110は、二値画像をスキャンして対象物のピクセルを見つけ、それが外側の境界か孔の境界かを決定していく。その決定をスキャンして繰り返していくことで最も外側の境界すなわち輪郭を検出していく。例えば。図3(B-1)は、二値画像中に含まれる線画に一点鎖線の輪郭50が検出されたイメージ図である。より詳細は、論文Topological structural analysis of digitized binary images by border following;Satoshi Suzuki著(Computer Vision, Graphics, and Image Processing Volume 30, Issue 1, April 1985, Pages 32-46)に開示されている。 Next, the area detection unit 110 detects contours in which personal information is likely to exist from the binary image (S30). Contours in which personal information is likely to exist are detected by detecting contours such as lines in a binary image. In general, the area detector 110 scans the binary image for pixels of the object and determines whether it is an outer boundary or a hole boundary. By scanning and repeating the determination, the outermost boundary or contour is detected. for example. FIG. 3B-1 is an image diagram in which a dashed-dotted line contour 50 is detected in a line drawing included in a binary image. More details are disclosed in the paper Topological structural analysis of digitized binary images by border following; by Satoshi Suzuki (Computer Vision, Graphics, and Image Processing Volume 30, Issue 1, April 1985, Pages 32-46).

 また、ステップS30と同時に、領域検出部110はグレースケール画像から個人情報が存在しそうな外枠を検出する(S31)。グレースケール画像は、ステップS22、S26もしくはS27で生成されたグレースケール画像のいずれを使用してもよい。概略的には、領域検出部110は、グレースケール画像中の輝度値が近い所を1つの領域にまとめていく,グレースケール画像の領域分割に使用する。輝度値が近い所をまとめていくという発想から,安定した分布のある外枠として解釈して、その代表点を特徴量として導出している。例えば。図3(B-2)は、グレースケール画像中に含まれる線画に二点鎖線の外枠51が検出されたイメージ図である。図3(B-2)が図3(B-1)と特に違う点は、“0”の内側も外枠として検出する点である。より詳細は、論文Efficient Maximally Stable Extremal Region (MSER) Tracking;M. Donoser et al著(2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06))に開示されている。 At the same time as step S30, the area detection unit 110 detects an outer frame in which personal information is likely to exist from the grayscale image (S31). Any of the grayscale images generated in steps S22, S26 or S27 may be used as the grayscale image. Schematically, the area detection unit 110 is used for area division of a grayscale image, in which areas having similar luminance values in the grayscale image are grouped into one area. Based on the idea of grouping together areas with similar luminance values, it is interpreted as an outer frame with a stable distribution, and its representative points are derived as feature quantities. for example. FIG. 3B-2 is an image diagram in which the outer frame 51 of the two-dot chain line is detected in the line drawing included in the grayscale image. FIG. 3B-2 is particularly different from FIG. 3B-1 in that the inside of "0" is also detected as an outer frame. More details are disclosed in the paper Efficient Maximally Stable Extremal Region (MSER) Tracking; by M. Donoser et al (2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)).

 個人情報の検出部116でカラー画像から個人情報を検出する(S32)。個人情報の検出部116は、ステップS22で生成された所定サイズのカラー画像に情景内文字検出プログラム(EAST:An Efficient and Accurate Scene Text Detector)等を適用して個人情報(文字)を検出する。個人情報の検出部116は、カラー画面に含まれる比較的大きな文字等を検出することができるが、背景などに含まれる小さな文字等は検出しにくい。 Personal information is detected from the color image by the personal information detection unit 116 (S32). The personal information detection unit 116 detects personal information (characters) by applying an in-scene character detection program (EAST: An Efficient and Accurate Scene Text Detector) or the like to the color image of a predetermined size generated in step S22. The personal information detection unit 116 can detect relatively large characters and the like included in the color screen, but it is difficult to detect small characters and the like included in the background.

 次に、ステップS30、S31で検出された個人情報が存在しそうな輪郭又は外枠領域に対して、個人情報判定部111で個人情報、つまり文字・数字があるか否かを判定する(S33)。また個人情報の検出部116で検出された文字に対しても。同様に個人情報判定部111で個人情報、つまり文字があるか否かを判定してもよい(S33)。個人情報の判定の手法に関しては、図6で後述する。 Next, the personal information judging unit 111 judges whether or not there is personal information, that is, letters and numbers, in the outline or outer frame area that is likely to contain personal information detected in steps S30 and S31 (S33). . Also for characters detected by the personal information detection unit 116 . Similarly, the personal information determining unit 111 may determine whether or not there is personal information, that is, characters (S33). A method for determining personal information will be described later with reference to FIG.

 なお、本実施形態では、二値画像から個人情報が存在するかを判定するルート(S29、S30、S33)、グレースケール画像から個人情報が存在するか判定するルート(S22(又はS26、S27),S31、S33)、カラー画像から個人情報が存在するかを判定するルート(S22,S32,S33)があり、それぞれのルートで個人情報(文字)を不図示のメモリーに記憶している。そしてそれぞれのル-トで記憶された個人情報(文字)を統合して、カラー画像に含まれている個人情報としている。しかし、これに限らず、これら3ルートの少なくとも1ルートだけを適用してもよく、3ルート中の2ルートを選択して統合してもよい。 In this embodiment, a route for determining whether personal information exists from a binary image (S29, S30, S33) and a route for determining whether personal information exists from a grayscale image (S22 (or S26, S27) , S31, S33) and a route (S22, S32, S33) for determining whether personal information exists from the color image, and personal information (characters) is stored in a memory (not shown) in each route. Then, the personal information (characters) stored in each route is integrated to obtain the personal information contained in the color image. However, the present invention is not limited to this, and at least one route out of these three routes may be applied, or two routes out of the three routes may be selected and integrated.

 ステップS33で個人情報が特定されると、個人情報消去部112は所定サイズのカラー画像から判定された個人情報を、その周辺画像で違和感なく消去する(S34)。図3(C)は、所定サイズのカラー画像に個人情報(図3(C)では線状のキズ)が描かれているカラー写真(C-1)と。黒色背景にステップS33で判定された個人情報を白抜きしたデータ(C-2)と、個人情報にその周囲の画素で描き入れていて(inpainting)したカラー画像(C-3)とを示した図である。別の言い方をすると、個人情報消去部112は、所定サイズのカラー画像(C-1)と同じサイズの黒色背景に個人情報を白抜きした画像(C-2)とを用意し、個人情報の境界から内側に向かって徐々に個人情報に周囲の画素で描き入れていていくことで(Inpainting)、個人情報が消去される(C-3)。 When the personal information is specified in step S33, the personal information erasing unit 112 erases the personal information determined from the color image of the predetermined size with the peripheral image without discomfort (S34). FIG. 3(C) is a color photograph (C-1) in which personal information (a linear flaw in FIG. 3(C)) is drawn on a color image of a predetermined size. Data (C-2) in which the personal information determined in step S33 is outlined on a black background, and a color image (C-3) in which the personal information is inpainted with surrounding pixels are shown. It is a diagram. In other words, the personal information erasing unit 112 prepares a color image (C-1) of a predetermined size and an image (C-2) of the same size in which the personal information is outlined against a black background and erases the personal information. The personal information is erased (C-3) by gradually drawing in the personal information with surrounding pixels from the boundary toward the inside (Inpainting).

 個人情報にその周囲の画素で描き入れていていく(inpainting)処理は、計算量が多く、プロセッサに負担がかかる又は時間がかかるため。ステップS34では、計算量を削減する工夫をすることが好ましい。具体的には、個人情報消去部112は所定サイズ(例えばFHD)から縮小サイズ画像(例えばHD)に変換する。この縮小サイズでカラー写真(C-1)と白抜きデータと(C-2)とに基づいて、個人情報の周囲の画素で描き入れる(inpainting)。そして個人情報消去部112は、縮小サイズの個人情報が消去された画像(C-3)を作成する。そして縮小サイズの個人情報が消去された画像(C-3)を所定サイズに戻し、所定サイズのカラー画像(C-1)に、所定サイズに戻された画像(C-3)で白抜きした位置(C-2)の画像を上書きすればよい。このような処理により計算量を格段に減らすことができる。 This is because the process of drawing (inpainting) personal information with the surrounding pixels requires a large amount of calculations, which puts a heavy burden on the processor or takes time. In step S34, it is preferable to devise ways to reduce the amount of calculation. Specifically, the personal information erasing unit 112 converts an image of a predetermined size (eg, FHD) into a reduced size image (eg, HD). Based on the color photograph (C-1), the white data (C-2), and the reduced size, pixels surrounding the personal information are drawn in (inpainting). Then, the personal information erasing unit 112 creates a reduced-size image (C-3) from which the personal information has been erased. Then, the reduced size image (C-3) from which the personal information has been deleted is returned to a predetermined size, and the color image (C-1) of the predetermined size is blanked with the image (C-3) restored to the predetermined size. The image at position (C-2) should be overwritten. Such processing can significantly reduce the amount of calculation.

 より詳しく説明すると、個人情報の境界から内側に向かって徐々に個人情報に周囲の画素で描き入れていく(Inpainting)手法には2つある。1つは、消去する近傍領域上のある一画素の値を、その周囲の画素の中で画素値が既に分かっている画素の画素値の重み付き和で置換し、ある画素を描き入れたら、FMM(Fast Marching Method)を使って次の最近傍点に移動し、個人情報がないもしくは消去済みの画素に近い画素から順番に描き入れていく。より詳細は、論文An image inpainting technique based on the fast marching method; Alexandru Telea著(January 2004 Journal of Graphics Tools)に開示されている。もう一つは、個人情報の周囲の既知の領域の画素値から個人情報の領域へエッジに沿って探索し画素を描き入れていく手法もある。より詳細は、論文Navier-stokes, fluid dynamics, and image and video inpainting;Marcelo Bertalmio et al著(Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.)に開示されている。 To explain in more detail, there are two methods of inpainting (Inpainting) in which the surrounding pixels are gradually drawn into the personal information from the boundary of the personal information toward the inside. One is to replace the value of a pixel in the neighboring area to be erased with the weighted sum of the pixel values of pixels whose pixel values are already known among the surrounding pixels, and draw a certain pixel, FMM (Fast Marching Method) is used to move to the next nearest neighbor point, and the pixels close to the pixels without personal information or erased are drawn in order. More details are disclosed in the article An image inpainting technique based on the fast marching method; by Alexandru Telea (January 2004 Journal of Graphics Tools). Another method is to draw pixels by searching along edges from pixel values in a known area around the personal information to the area of the personal information. More details are disclosed in the paper Navier-stokes, fluid dynamics, and image and video inpainting; by Marcelo Bertalmio et al (Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.).

 次に、保護領域検出部103で保護領域を検出した場合、保護領域上書き部113は、保存しておいた保護領域を、個人情報が消去されたカラー画像中に上書きする(S35)。領域検出部110は個人情報が存在しそうな領域には、例えば眼鏡のテンプルや鼻の影等が含まれることがあり、個人情報消去部112でテンプルや鼻の影を消去してしまうと保護領域が乱れてしまう。このため、保存しておいた保護領域を個人情報が消去されたカラー画像に上書きすることで保護領域を保護する。ステップS23で保護領域が検出されなかった場合には、フラグに基づいてステップS35をスキップしても良い。 Next, when the protected area detection unit 103 detects a protected area, the protected area overwrite unit 113 overwrites the saved protected area in the color image from which the personal information has been deleted (S35). The area detection unit 110 detects that areas in which personal information is likely to exist include, for example, the temples of eyeglasses and the shadow of the nose. is disturbed. Therefore, the protected area is protected by overwriting the saved protected area with the color image from which the personal information has been deleted. If no protected area is detected in step S23, step S35 may be skipped based on the flag.

 そして、保護領域を上書きしたカラー写真に対して、仕上げ部114が仕上げ処理することが好ましい(S36)。具体的には、境界をまたいでの輝度値の平滑を抑制するパターン保存フィルター(ノンローカルミーンフィルター)又はガウシアンぼかし処理をカラー画像に対して使用する。上書きした保護領域との境界を目立たなくするとともに、消去できなかった微小な個人情報(文字)を判読不可能にすることができる。仕上げ処理がなくても問題はない。 Then, it is preferable that the finishing section 114 performs finishing processing on the color photograph in which the protected area has been overwritten (S36). Specifically, a pattern preserving filter (non-local mean filter) that suppresses smoothing of luminance values across boundaries or Gaussian blurring is used for color images. The boundary with the overwritten protected area can be made inconspicuous, and minute personal information (characters) that could not be erased can be made illegible. There is no problem even if there is no finishing treatment.

 最後に、カラー画像出力部115が、仕上げ処理されたカラー画像を、パーソナルコンピュータ23もしくは携帯情報端末25に出力する(S37)。 Finally, the color image output unit 115 outputs the finished color image to the personal computer 23 or the mobile information terminal 25 (S37).

<保護領域の検出>
 ここで、保護領域検出部103が保護領域を検出する機能(S23)について、図3及び図4を使って説明する。
<Detection of protection area>
Here, the function (S23) of detecting the protected area by the protected area detection unit 103 will be described with reference to FIGS. 3 and 4. FIG.

 まず、ステップS231では、保護領域検出部103は、ディープラーニングを行えるAI等を使用し、所定サイズに変更されたカラー画像に基づいて、目や鼻、口の位置といった人間の顔が持つ特徴から、顔領域を検出したりする。顔領域は、個人情報(文字)を含まない保護領域であり、ステップS235で顔領域が保存され、S35で説明したように、個人情報が消去されたカラー画像中に上書きされる。 First, in step S231, the protection area detection unit 103 uses an AI capable of deep learning, etc., based on the color image that has been changed to a predetermined size, from the features of the human face such as the positions of the eyes, nose, and mouth. , to detect the face area. The face area is a protected area that does not contain personal information (characters). The face area is saved in step S235 and overwritten in the color image from which the personal information has been erased as described in S35.

 次に、ステップS232において、保護領域検出部103は、AI等を使用し、カラー画像に含まれる物体をラベル化する。このため、カラー画像に時計があったり、カレンダーがあったり、又はテーブルがあったりすると、保護領域検出部103は、“clock”、“wall calendar”又は“table”等のラベルを出力する。例えば、時計やカレンダーには、時計やカレンダーに時間や日付の数字(文字)が描かれているが、文字が含む物体であるが個人情報を含むものではない物体、時計等を予め登録しておく。時計等は保護領域として検出され、ステップS235で時計等の保護領域が保存される。 Next, in step S232, the protected area detection unit 103 uses AI or the like to label the objects included in the color image. Therefore, when a color image includes a clock, calendar, or table, the protected area detection unit 103 outputs a label such as "clock", "wall calendar", or "table". For example, clocks and calendars have time and date numbers (characters) drawn on them, but objects that contain characters but do not contain personal information, clocks, etc., are registered in advance. back. A clock or the like is detected as a protected area, and the protected area of the clock or the like is saved in step S235.

 次に、ステップS233において、保護領域検出部103は、所定サイズに変更されたカラー画像に皮膚の色があるかを検出する。例えば指と指との間の空間(影)が一本の線の文字(例えばアルファベットの“I”)と認識されることを防ぐ。皮膚領域は、個人情報を含まない保護領域であり、ステップS235で皮膚領域が保存される。 Next, in step S233, the protected area detection unit 103 detects whether the color image resized to the predetermined size has skin color. For example, the spaces (shadows) between the fingers are prevented from being recognized as a single line character (for example, "I" of the alphabet). The skin area is a protected area that does not contain personal information, and the skin area is saved in step S235.

 保護領域検出部103は、所定サイズのカラー画像(RGB)に例えば、カラー画像の表示方式RGB、HSV及びYCbCrを使って皮膚の色を検出しても良い。より詳細は、論文Human Skin Detection Using RGB, HSV and YCbCr Color Models;S. Kolkur et al著(ICCASP/ICMMD-2016. Published by Atlantic Press)に開示されている。本実施形態では、カラー画像(RGB)をカラー画像(HSV)に変換するとともに、色相(hue)、彩度(Saturation)及び明度(Value)を調整して皮膚の色の領域を検出する。例えば8ビット(0-255)で調整するならば、H=16~40(赤っぽい~黄色っぽい)S=20~200、V=128~255を範囲としておけば、カラー画像中の皮膚の色が存在する領域を検出できる。 The protected area detection unit 103 may detect the color of the skin using, for example, color image display methods RGB, HSV, and YCbCr for a color image (RGB) of a predetermined size. More details are disclosed in the paper Human Skin Detection Using RGB, HSV and YCbCr Color Models; by S. Kolkur et al (ICCASP/ICMMD-2016. Published by Atlantic Press). In this embodiment, a color image (RGB) is converted into a color image (HSV), and hue, saturation, and value are adjusted to detect a skin color area. For example, if adjusting with 8 bits (0-255), H = 16 to 40 (reddish to yellowish) S = 20 to 200, V = 128 to 255, the skin in the color image It can detect areas where color exists.

 次に、ステップS234において、保護領域検出部103は、ユーザHにパーソナルコンピュータ23又は携帯情報端末25の画面上で保護領域を設定できるように、保護枠を表示させる。図5は、アップロードされたカラー画像51が携帯情報端末25の画面に表示された一例であり、カラー画像51は学校の教室の一部である。教室内には、2つの机と、1つの棚があり、机には文字情報が映った状態のパーソナルコンピュータがあり、棚には複数の本やファイルがある。教室の壁には、時計とプラカードと窓とがある。時計には文字盤があり、プラカードには「優勝おめでとう」の文字がある。 Next, in step S234, the protected area detection unit 103 displays a protected frame so that the user H can set a protected area on the screen of the personal computer 23 or the mobile information terminal 25. FIG. 5 is an example of an uploaded color image 51 displayed on the screen of the mobile information terminal 25, and the color image 51 is part of a school classroom. In the classroom, there are two desks and one shelf, the desk has a personal computer on which character information is displayed, and the shelf has a plurality of books and files. There are clocks, placards and windows on the walls of the classroom. The clock has a dial and the placard has the words "Congratulations on winning".

 保護領域検出部103は、図5の携帯情報端末25に「画像中で残したい文字を枠で囲ってください。」の表示53をさせる。ユーザHが枠57をクリックすると、枠57が画面上に表示される。ユーザHは、その枠57の大きさと位置を変えて、例えばプラカードの周囲に枠57aを設定し、棚の一番下段に枠57bを設定する。ユーザHは枠57の設定が終了すると完了ボタン54をクリックする。このように枠57a又は57bで設定された領域が保護領域として検出され、ステップS235で保護領域が保存される。S35で説明したように、個人情報が消去されたカラー画像中に保護領域が上書きされる。なお、図5の教室の壁の時計は、ステップS234で保護領域として囲われなくても、ステップ232が実行されれば時計は保護領域とされる。図5では静止画を前提に説明したが動画でも保護領域の設定は可能である。例えば、ユーザHが指定した枠領域をテンプレート画像として保持し、テンプレートマッチング処理により、動画内にテンプレート画像とマッチする領域をチェックすることで対応できる。 The protected area detection unit 103 causes the portable information terminal 25 of FIG. When the user H clicks on the frame 57, the frame 57 is displayed on the screen. The user H changes the size and position of the frame 57, for example, sets a frame 57a around the placard and sets a frame 57b at the bottom of the shelf. User H clicks the completion button 54 when the setting of the frame 57 is completed. The area set by the frame 57a or 57b in this manner is detected as a protected area, and the protected area is saved in step S235. As described in S35, the protected area is overwritten in the color image from which the personal information has been erased. Note that even if the clock on the wall of the classroom in FIG. 5 is not enclosed as a protected area in step S234, the clock is made a protected area when step 232 is executed. Although FIG. 5 has been described on the premise of a still image, the protection area can also be set for a moving image. For example, a frame area designated by the user H is held as a template image, and template matching processing is performed to check areas in the moving image that match the template image.

 すべての保護領域をステップS231からステップS234の順番が入れ替わっても良く、また例えば顔領域のみを検出する等、いずれか一つのステップのみを実行してもよい。また、ステップS32等において、個人情報の領域検出部110で個人情報の領域を検出するとともに、その個人情報検出部111で検出された検出領域に対して、顔領域もしくは皮膚領域等の保護領域が存在するかを検出しても良い。 The order of steps S231 to S234 may be changed for all protected areas, or only one of the steps may be executed, such as detecting only the face area. In step S32 and the like, the personal information area detection unit 110 detects the area of personal information, and the detection area detected by the personal information detection unit 111 includes a protection area such as a face area or a skin area. You may detect whether it exists.

<個人情報(文字・数字)の判定>
 ここで、個人情報が存在しそうな領域に対して、個人情報判定部111が個人情報であるか否かを判定する手法(S33)について、図6から図8を使って説明する。なお、個人情報の判定手法では、個人情報が文字(数字)である場合について説明する。そのため、個人情報が存在しそうな領域にある個人情報を推定文字(Estimated letter)と呼ぶ。
<Determination of personal information (characters/numbers)>
Here, a method (S33) for determining whether personal information is personal information by the personal information determining unit 111 for an area in which personal information is likely to exist will be described with reference to FIGS. 6 to 8. FIG. In the personal information determination method, a case where the personal information is characters (numbers) will be described. Therefore, personal information in an area where personal information is likely to exist is called an estimated letter.

 まず、個人情報判定部111は、推定文字を回転させて元の推定文字に重ね合わせた面積と元の推定文字の面積との面積比が、ある閾値k1よりも大きいか否かを判定する(S331)。閾値k1より面積比が大きい場合には図形等(文字以外)であると判定してステップS336に進み、閾値k1より面積比が小さい場合には文字であると判定してステップS332に進む。 First, the personal information determination unit 111 determines whether or not the area ratio between the area of the estimated character rotated and superimposed on the original estimated character and the area of the original estimated character is greater than a certain threshold k1 ( S331). If the area ratio is larger than the threshold k1, it is determined to be a figure (other than a character) and the process proceeds to step S336.

 図7(A)に、推定文字を回転させた例を示す。図7(A)では、アルファベットの“C”、“O”及び“Z”並びに真円、正方形及び正三角形の図形を推定文字の例とする。また図7(A)に示した表では、左から0°(回転せず)、90°回転、0°の推定文字と90°回転した推定文字との重ね合わせ、180°回転、0°の推定文字と180°回転した推定文字との重ね合わせ、270°回転、0°の推定文字と270°回転した推定文字との重ね合わせが示されている。なお回転は、文字領域または図形領域の重心位置を中心に回転させている。 Fig. 7(A) shows an example of rotating the estimated characters. In FIG. 7A, alphabetic characters "C", "O" and "Z" and figures of perfect circles, squares and equilateral triangles are used as examples of estimated characters. Further, in the table shown in FIG. 7A, from the left, 0° (not rotated), 90° rotation, superposition of 0° estimated character and 90° rotated estimated character, 180° rotation, 0° rotation Superposition of the estimated character and the estimated character rotated by 180°, superposition of the estimated character rotated by 270°, and the estimated character rotated by 0° and the estimated character rotated by 270° are shown. Note that the rotation is performed around the position of the center of gravity of the character area or figure area.

 まず、0°と90°との重ね合わせでは、“C”、“O”及び“Z”は、重なり領域が50%以下である。一方、真円、正方形及び正三角形は、重なり領域が50%以上である。これらの文字及び図形は、0°と270°との重ね合わせにおいても同様である。 First, in the superimposition of 0° and 90°, the overlap region of "C", "O" and "Z" is 50% or less. On the other hand, a perfect circle, a square and an equilateral triangle have an overlapping area of 50% or more. These letters and figures are the same in superposition of 0° and 270°.

 次に0°と180°との重ね合わせでは、“O”及び“Z”は、それらの重なり領域が100%であり、“C”の重なり領域は約80%である。真円及び正方形は、重なり領域が100%であり、正三角形の重なり領域は約50%である。 Next, in the superposition of 0° and 180°, "O" and "Z" have 100% of their overlapping area, and "C" has about 80% of their overlapping area. A perfect circle and a square have an overlapping area of 100%, and an equilateral triangle has an overlapping area of about 50%.

 本実施形態では、個人情報判定部111は、アルファベットだけでなく多言語に対応するため、0°と90°との重ね合わせの閾値k11、0°と180°との重ね合わせの閾値k12、0°と270°との重ね合わせの閾値k13をそれぞれ有しており、その3つ回転角度の重ね合わせを組み合わせて、推定文字が文字であるかを判定する。個人情報判定部111が、ある特定言語の推定文字を文字判定するのであれば、例えば0°と90°との重ね合わせの閾値k11だけを使用してもよい。また、閾値k1は可変できるようにすることが好ましく、言語ごとに閾値k1値を可変できるようにすることが好ましい。 In this embodiment, the personal information determination unit 111 supports not only the alphabet but also multiple languages. It has a threshold value k13 for superposition of ° and 270°, respectively, and it is determined whether or not the estimated character is a character by combining the superposition of the three rotation angles. If the personal information determination unit 111 performs character determination on an estimated character of a certain specific language, for example, only the threshold value k11 for overlapping 0° and 90° may be used. Moreover, it is preferable to make the threshold k1 variable, and it is preferable to make the threshold k1 value variable for each language.

 上述したように、0°と180°との重ね合わせでは、“C”の重なり領域は約80%であり、正三角形の重なり領域は約50%であり、文字より図形の方が重なり領域が小さくなっている。これでは推定文字が文字であるかを判定し難いため、図7(B)に示すように、推定文字を回転させてもよい。図7(B)では、アルファベットの“C”、真円及び正三角形の図形を推定文字の例とする。また図7(B)に示した表では、左から0°(回転せず)、距離Sだけシフト、シフトした状態で180°回転、0°の推定文字とシフトして180°回転した推定文字との重ね合わせが示されている。なお回転は、文字領域または図形領域の重心位置を中心に回転させている。“C”の重なり領域は約10%であり、新円の重なり領域は約70%であり、正三角形の重なり領域は約50%である。0度の推定文字と距離Sのシフト且つ180°回転した推定文字との重なり領域で、閾値k12より大きいか否かで推定文字が文字であるかを判定できる。シフトして180°回転させる手法と、90°回転と、270°回転とを組み合わせてもよい。なお、図7ではアルファベットについて説明したが、算用数字や漢字でも同様である。 As described above, in the superimposition of 0° and 180°, the overlapping area of "C" is about 80%, and the overlapping area of the equilateral triangle is about 50%. It's getting smaller. Since it is difficult to determine whether the estimated character is a character, the estimated character may be rotated as shown in FIG. 7B. In FIG. 7B, an alphabet "C", a perfect circle, and an equilateral triangle are examples of estimated characters. Further, in the table shown in FIG. 7B, 0° from the left (not rotated), shifted by a distance S, rotated 180° with the shift, the estimated character at 0°, and the estimated character shifted and rotated 180°. A superposition with is shown. Note that the rotation is performed around the position of the center of gravity of the character area or figure area. The overlap area of the "C" is about 10%, the overlap area of the new circle is about 70%, and the overlap area of the equilateral triangle is about 50%. Whether or not the estimated character is a character can be determined by checking whether or not the overlap region of the estimated character at 0 degrees and the estimated character shifted by the distance S and rotated by 180 degrees is larger than the threshold value k12. A technique of shifting and rotating 180°, rotating 90°, and rotating 270° may be combined. Although the alphabet is explained in FIG. 7, the same applies to Arabic numerals and Chinese characters.

 ステップS332では、個人情報判定部111は、推定文字の中心線から推定文字の境界線までの距離の標準偏差を算出し、標準偏差と閾値k2とを比較して文字と判定する。閾値k2より標準偏差が大きい場合には図形などであると判定してステップS336に進み、閾値k2より標準偏差が小さい場合には文字であると判定してステップS333に進む。既にステップS331で文字であると判定しているが、ステップS332では別の手法により推定文字が文字であるかを判定してもよい。 In step S332, the personal information determination unit 111 calculates the standard deviation of the distance from the center line of the estimated character to the boundary line of the estimated character, compares the standard deviation with a threshold value k2, and determines that the character is a character. If the standard deviation is larger than the threshold k2, it is determined to be a figure or the like, and the process proceeds to step S336. Although it has already been determined in step S331 that the character is a character, in step S332 another method may be used to determine whether the estimated character is a character.

 文字は、文字の中心線(骨組み)から文字の境界線までの距離(複数個所)に大きな変化が少ないという特性を有している。このため、推定文字の中心線から推定文字の境界線までの距離の標準偏差を算出する。ただし、文字の大きさに比例して標準偏差も大きくなるため、中心線から境界線までの平均距離で割る、つまり閾値k2=距離の標準偏差/平均距離を計算することが好ましい。平均距離で割ることで閾値k2は、文字の大きさに影響を受けにくくなる。また中心線から境界線までの平均距離で割る代わりに、(距離/平均距離)の標準偏差で閾値k2を設定しても良い。また図形等と文字との判定基準を調整するため、閾値k2値を可変できるようにすることが好ましい。 Characters have the characteristic that the distance (multiple points) from the center line (framework) of the character to the boundary line of the character does not change significantly. Therefore, the standard deviation of the distance from the center line of the estimated character to the boundary line of the estimated character is calculated. However, since the standard deviation also increases in proportion to the character size, it is preferable to divide by the average distance from the center line to the boundary line, that is, to calculate threshold k2=standard deviation of distance/average distance. By dividing by the average distance, the threshold k2 becomes less susceptible to the character size. Also, instead of dividing by the average distance from the center line to the boundary line, the threshold value k2 may be set by the standard deviation of (distance/average distance). Also, it is preferable to make the threshold value k2 variable in order to adjust the criteria for judging graphics and the like and characters.

 図8(A)は、アルファベットの“J”とJに似た台形の図形を示した例である。“J”の中心線(骨組み)61からは境界線63までの距離L1は、ほとんど一定でありその標準偏差は小さい。図8(A)ではJはゴジック体(Gothic)であるが他の書体例えばニューセンチュリー(New Century )であっても標準偏差は小さい。台形の図形の中心線(骨組み)61からは境界線63までの距離L2は大きく変化しており、その標準偏差は大きくなる。例えば閾値k2=0.6と設定してもよい。 FIG. 8(A) is an example showing the alphabet "J" and a trapezoidal figure similar to J. The distance L1 from the center line (framework) 61 of "J" to the boundary line 63 is almost constant and its standard deviation is small. In FIG. 8A, J is Gothic, but the standard deviation is small even in other typefaces such as New Century. The distance L2 from the center line (framework) 61 of the trapezoidal figure to the boundary line 63 varies greatly, and its standard deviation increases. For example, the threshold k2 may be set to 0.6.

 ステップS333では、個人情報判定部111は、推定文字の文字自体の面積と推定文字を囲む領域全体の面積との面積比を算出し、閾値k3より小さければ文字と判定する。閾値k3より大きい場合には図形などであると判定してステップS336に進み、閾値k3より小さい場合には文字であると判定してステップS334に進む。既にステップS331及びS332で文字であると判定しているが、ステップS333では別の手法により推定文字が文字であるかを判定してもよい。 In step S333, the personal information determination unit 111 calculates the area ratio of the area of the estimated character itself and the area of the entire area surrounding the estimated character, and determines that it is a character if it is smaller than the threshold value k3. If it is larger than the threshold k3, it is determined to be a figure or the like and the process proceeds to step S336. If it is smaller than the threshold k3, it is determined to be a character and the process proceeds to step S334. Although it has already been determined that the character is a character in steps S331 and S332, another method may be used to determine whether the estimated character is a character in step S333.

 図8(B)は、ステップS333の具体例であり、アルファベットの“J”とJに似たスプーン形状の図形を示した例である。図8(B)ではJ文字自体の面積66とJを囲む領域全体の面積65との面積比は約60%であり、スプーン形状の図形の面積66とスプー形状の図形を囲む領域全体の面積65との面積比は約80%である。例えば閾値k3=70%と設定してもよい。 FIG. 8(B) is a specific example of step S333, showing an alphabet "J" and a spoon-shaped figure similar to J. In FIG. 8B, the area ratio of the area 66 of the J character itself to the area 65 of the entire area surrounding J is about 60%, and the area 66 of the spoon-shaped figure and the area of the entire area surrounding the spoon-shaped figure The area ratio with 65 is about 80%. For example, the threshold k3 may be set to 70%.

 ステップS331からステップS333の順番が入れ替わっても良く、またいずれか一つのステップのみで推定文字が文字であると判定してもよい。ステップS334及びS335では、ステップS331、S332又はS333で文字であることが判定されたので、その文字が個人情報であるか否かを判定する。  The order of steps S331 to S333 may be changed, or it may be determined that the estimated character is a character in only one of the steps. In steps S334 and S335, it is determined whether or not the character is personal information because it was determined in step S331, S332 or S333 that the character is a character.

 再び図6に戻り、ステップS334では、推定文字が所定サイズの画像に対してある閾値k4よりも大きいか否かを判定する。閾値k4より大きい場合には大きな文字であると判定してステップS338に進み、閾値k4より小さい場合には小さな文字であると判定してステップS336に進む。より具体的には、画像サイズがFHD(1920*1088)である場合、推定文字の少なくとも一方の縦方向のピクセル数もしくは横方向のピクセル数が、画像サイズの例えば5パーセント(96*54)以上であるか否を判定する。大きい場合にはステップS339に進む。画像サイズの何パーセント(閾値k4)であるかは可変できることが好ましい。ユーザが残したい大きさ文字を決定できるように、カラー画像をアップロードするweb画面等に、閾値k4を可変できる機能を表示されても良い。 Returning to FIG. 6 again, in step S334, it is determined whether or not the estimated character is larger than a certain threshold k4 for an image of a predetermined size. If it is larger than the threshold k4, it is determined that the character is large, and the process proceeds to step S338. If it is smaller than the threshold k4, it is determined that the character is small, and the process proceeds to step S336. More specifically, when the image size is FHD (1920*1088), at least one of the estimated characters has a vertical pixel count or a horizontal pixel count of, for example, 5 percent (96*54) or more of the image size. It is determined whether or not. If larger, the process proceeds to step S339. The percentage of the image size (threshold k4) is preferably variable. A function that can change the threshold value k4 may be displayed on a web screen or the like for uploading a color image so that the user can determine the size of characters that the user wants to leave.

 ステップS335では、文字がある閾値k5よりも大きな太い線幅で描かれているか否かを判定する。閾値k5より大きい場合には個人情報でないと判定してステップS338に進み、閾値k5より小さい場合には細い線幅の文字であると判定してステップS337に進む。閾値k5も可変できることが好ましい。ユーザが残したい太い線幅の文字を決定できるように、カラー画像をアップロードするweb画面等に、閾値k5を可変できる機能を表示されても良い。 In step S335, it is determined whether or not the character is drawn with a thick line width greater than a certain threshold value k5. If it is larger than the threshold k5, it is determined that the information is not personal information, and the process proceeds to step S338. Preferably, the threshold k5 is also variable. A function that can change the threshold value k5 may be displayed on a web screen or the like for uploading a color image so that the user can determine the characters with a thick line width that the user wants to leave.

 図8(C)は、ステップS335の具体例であり、アルファベットの太い線幅の“I”と細い線幅の“I”を示した例である。“I”文字自体68を構成する面積と“I”の3本の中心線(骨組み)67の面積とを計算し、その面積比が閾値k5より、例えば10よりも大きいかを判定する。 FIG. 8(C) is a specific example of step S335, showing a thick line width "I" and a thin line width "I" of the alphabet. The area of the "I" character itself 68 and the area of the three centerlines (framework) 67 of the "I" are calculated, and it is determined whether the ratio of the areas is greater than a threshold k5, for example 10.

 図6のステップS336において、ステップS331、S332又はS333で、文字ではないとして判定された推定文字は図形等と考えられるため、カラー画像の一部として残される。  In step S336 of FIG. 6, the estimated characters determined as not being characters in steps S331, S332, or S333 are considered to be graphics or the like, and are therefore left as part of the color image.

 ステップS337において、ステップS331又はS332で推定文字は文字と判定され、且つステップS334又はS335で大きな文字でもなく太い線幅の文字でもないと判定されたため、これらの文字は個人情報としてカラー画像から消去される対象となる。 In step S337, it is determined that the estimated characters are characters in step S331 or S332, and that they are neither large characters nor thick line width characters in step S334 or S335. Therefore, these characters are deleted from the color image as personal information. subject to

 ステップS338において、ステップS332又はS333で推定文字は文字と判定され、且つステップS334又はS335で大きな文字又は太い線幅の文字と判断されたため、これらの文字は個人情報ではなく、例えばカラー画面に大きく写った学校の入り口に吊るされた「卒業式」の看板や、太い線幅でTシャツに描かれた文字は、カラー画像に残される。なお、大きな文字又は太い線幅の文字であれば、名前、住所、生年月日、性別、ナンバープレートの数字等であっても、本実施形態では個人情報として扱わない。 In step S338, the estimated characters are determined to be characters in step S332 or S333, and are determined to be large characters or characters with a thick line width in step S334 or S335. The "Graduation Ceremony" signboard hanging at the entrance of the school and the letters drawn in thick lines on the T-shirt are preserved in the color image. In the present embodiment, large characters or characters with a thick line width are not treated as personal information, even if they are names, addresses, dates of birth, sex, numbers on license plates, and the like.

 図9は、パーソナルコンピュータ23から個人情報マスキング装置100に取得されるカラー画像である。また、図10、図11、図12及び図13は、個人情報マスキング装置100がパーソナルコンピュータ23に出力したカラー画像である。 9 is a color image acquired by the personal information masking device 100 from the personal computer 23. FIG. 10, 11, 12 and 13 are color images output to the personal computer 23 by the personal information masking device 100. FIG.

 図9に示されたカラー画像には、10台以上の自動車のナンバープレート、自動車に描かれた8種類(英語、タイ語、アラビア語、キリル語、日本語、グルジア語、ハングル語、ミャンマー語)の言語の文字、ニュースタイトル等の英語の文字、レポータの顔などが含まれている。 The color image shown in FIG. 9 includes license plates of more than 10 cars, 8 languages (English, Thai, Arabic, Cyrillic, Japanese, Georgian, Korean, Burmese) drawn on the car. ) language characters, English characters such as news titles, and reporter faces.

 図10は実施例1の出力例であり、第1実施例は、図2のステップS31及びS32を使用せずに、カラー画像を処理した例である。つまり図10は、図2のステップS21-S30、S33-S37を実行したカラー画像である。なおステップS23は、図3のS231(顔領域の保護)のみ実行されている。 FIG. 10 is an output example of Example 1, and Example 1 is an example of processing a color image without using steps S31 and S32 of FIG. That is, FIG. 10 is a color image obtained by executing steps S21-S30 and S33-S37 in FIG. In step S23, only S231 (face area protection) in FIG. 3 is executed.

 図10に示されたカラー画像には、タイ語、日本語、ハングル語及びキリル語が消去されている。しかし英語(アルファベット)、ミャンマー語及びアラビア語の一部や数字が残っている。多くの白線は図形として判定されたため、多くの道路の白線が残っている。レポータの名札の名前、レポ―タが持つマイクや放送局名も消去されている。LIVE NEWSの文字や、ニュースタイトルはしっかりと残っている。ステップS231)及びS36の実行によりレポータの顔ははっきりしている。 Thai, Japanese, Hangul, and Cyrillic are deleted from the color image shown in FIG. However, English (alphabet), Burmese and Arabic parts and numbers remain. Since many white lines were determined as figures, many white lines of the road remain. The name on the reporter's name tag, the microphone held by the reporter, and the name of the broadcasting station have also been erased. The words LIVE NEWS and the news titles are still there. The reporter's face has been clarified by the execution of steps S231) and S36.

 図11は実施例2の出力例であり、実施例2では、図2のステップS24-30及びS32が実行されていない。つまり、図11は、ステップS22で変換されたグレースケール画像をステップS31で実行したカラー画像である。なおステップS23は、図3のS231(顔領域の保護)のみ実行されている。 FIG. 11 is an output example of Example 2, in which Steps S24-30 and S32 of FIG. 2 are not executed. That is, FIG. 11 is a color image obtained by performing step S31 on the grayscale image converted in step S22. In step S23, only S231 (face area protection) in FIG. 3 is executed.

 図11に示されたカラー画像には、遠くの自動車のナンバープレートも含めてすべてのナンバープレートの文字及び数字が消去されている。また、自動車に描かれた8種類の言語すべてが消去されている。レポータの名札の名前、レポ―タが持つマイクや放送局名も消去されている。また道路の白線も文字として判定されて消去されている。LIVE NEWSの文字や、ニュースタイトルはボケている箇所もあるが消去されていない。なお、図10と図11とを比べても分かりにくいが、グレースケール画像をステップS31で実行したカラー画像では、特に赤地に黒字などは消去されにくいことがある。ステップS231及びS36の実行によりレポータの顔ははっきりしている。 In the color image shown in Fig. 11, the letters and numbers on all license plates, including those of distant cars, have been erased. Also, all eight languages painted on the car have been erased. The name on the reporter's name tag, the microphone held by the reporter, and the name of the broadcasting station have also been erased. The white lines on the road are also determined as characters and deleted. Some parts of the LIVE NEWS text and news titles are blurred, but they have not been erased. Although it is difficult to understand by comparing FIG. 10 and FIG. 11, in the color image obtained by processing the grayscale image in step S31, black characters on a red background may be particularly difficult to erase. The reporter's face is clear due to the execution of steps S231 and S36.

 図12は実施例3の出力例であり、実施例3では、ステップS32を除き、図2のフローチャートがすべて実行されている。このため図12は、図10及び図11を重ね合わせた写真と同等になっている。なおステップS23は、図3のS231(顔領域の保護)のみ実行されている。 FIG. 12 is an output example of Example 3. In Example 3, all the flowcharts of FIG. 2 are executed except for step S32. Therefore, FIG. 12 is equivalent to the photograph in which FIGS. 10 and 11 are superimposed. In step S23, only S231 (face area protection) in FIG. 3 is executed.

 図12に示されたカラー画像には、遠くの自動車のナンバープレートも含めてすべてのナンバープレートの文字及び数字が消去されている一方で、図11では消去されていた、白線の一部や自動車の車体の一部が描かれている。ステップS231及びS36の実行によりレポータの顔ははっきりしている。 In the color image shown in FIG. 12, the letters and numbers of all the license plates, including the license plate of the distant car, have been erased, while some of the white lines and the car have been erased in FIG. A part of the car body of is drawn. The reporter's face is clear due to the execution of steps S231 and S36.

 図13は実施例4の出力例であり、実施例4では、実施例2と同様にステップS24-30及びS32が実行されておらず、ステップS22で変換されたグレースケール画像をステップS31で実行したカラー画像である。但し、ステップS23は、図3のS234(保護領域の設定)のみ実行されており、ユーザHに設定された保護領域は、レポータの写っているレポータ枠である。 FIG. 13 is an output example of Example 4. In Example 4, steps S24-30 and S32 are not executed as in Example 2, and the grayscale image converted in step S22 is executed in step S31. This is a color image. However, in step S23, only S234 (setting of protected area) in FIG. 3 is executed, and the protected area set by user H is the reporter frame in which the reporter is captured.

 図13に示されたカラー画像には、遠くの自動車のナンバープレートも含めてすべてのナンバープレートの文字及び数字が消去されている。また、自動車に描かれた8種類の言語すべてが消去されている。一方、レポータの名札の名前、レポ―タが持つマイクの文字、並びにレポータ枠の左上の放送局名及び日付も元のカラー画像そのままに維持されている。もちろん、図13に示されたカラー画像では、レポータ枠内にあるレポータの顔及び指先もはっきりしている。 In the color image shown in Fig. 13, the letters and numbers on all license plates, including those of distant cars, have been erased. Also, all eight languages painted on the car have been erased. On the other hand, the name of the reporter's name tag, the characters on the microphone held by the reporter, and the broadcasting station name and date on the upper left of the reporter's frame are also maintained as they are in the original color image. Of course, the color image shown in FIG. 13 also clearly shows the reporter's face and fingertips within the reporter's frame.

 本実施形態の個人情報マスキング装置は、カラー画像の取得から個人情報を消去したカラー画像の出力までをすべて処理している。しかしながら一部の処理をパーソナルコンピュータ23もしくは携帯情報端末25に担わせても良い。例えば、パーソナルコンピュータ23もしくは携帯情報端末25は、クラウドサーバーからアプリを事前にダウンロードしておき、パーソナルコンピュータ23もしくは携帯情報端末25で、カラー画像を所定のサイズに変換し、保護領域例えば顔領域等を検出し保存してから、個人情報マスキング装置100にアップロードしてもよい。そして個人情報が消去された後に、個人情報マスキング装置100からカラー画像を出力し、パーソナルコンピュータ23もしくは携帯情報端末25で保護領域を上書きしてもよい。 The personal information masking device of this embodiment processes everything from acquiring a color image to outputting a color image from which personal information has been erased. However, the personal computer 23 or the mobile information terminal 25 may be responsible for part of the processing. For example, the personal computer 23 or the mobile information terminal 25 downloads an application from the cloud server in advance, and the personal computer 23 or the mobile information terminal 25 converts the color image into a predetermined size and protects the protected area such as the face area. may be detected and stored before being uploaded to the personal information masking device 100 . After the personal information is erased, a color image may be output from the personal information masking apparatus 100 and the protected area may be overwritten by the personal computer 23 or the portable information terminal 25 .

 また、パーソナルコンピュータ23もしくは携帯情報端末25で、個人情報マスキング装置100のソフトウェアの処理をすべて担わせてもよい。  Also, the personal computer 23 or the portable information terminal 25 may be responsible for all software processing of the personal information masking device 100 . 

23 パーソナルコンピュータ
25 携帯情報端末
100 個人情報マスキング装置
101 カラー画像取得部
102 画像サイズ変換部
103 保護領域検出部と、
104 ノイズ除去部(カラー)
105 色空間変換部
106 グレースケール変換部
107 局所的ヒストグラム平坦化部
108 二値変換部
109 ノイズ除去部(二値)
110 領域検出部
111 個人情報判定部
112 個人情報消去部
113 保護領域上書き部
114 仕上げ部
115 カラー画像出力部
116 個人情報の検出部

 
23 personal computer 25 portable information terminal 100 personal information masking device 101 color image acquisition unit 102 image size conversion unit 103 protected area detection unit;
104 Noise remover (color)
105 color space converter 106 grayscale converter 107 local histogram equalizer 108 binary converter 109 noise remover (binary)
110 Area detection unit 111 Personal information determination unit 112 Personal information deletion unit 113 Protected area overwrite unit 114 Finishing unit 115 Color image output unit 116 Personal information detection unit

Claims (12)

 カラー画像を取得する工程と、
 取得した前記カラー画像を所定サイズの第1カラー画像に変換するサイズ変換工程と、
 前記第1カラー画像を二値画像に変換する工程と、
 前記二値画像に対して、個人情報が存在する可能性がある領域を検出する領域検出工程と、
 前記個人情報が存在しそうな領域に対して個人情報の有無を判定する判定工程と、
 前記判定工程で判定された個人情報と前記第1カラー画像とに基づいて、前記個人情報の周辺画素を前記第1カラー画像の前記個人情報に描き入れ前記個人情報を消去する消去工程と、
 前記消去工程後のカラー画像を出力する工程と、
 を備える、個人情報マスキング方法。
obtaining a color image;
a size conversion step of converting the obtained color image into a first color image of a predetermined size;
converting the first color image to a binary image;
a region detection step of detecting a region in which personal information may exist in the binary image;
a determination step of determining the presence or absence of personal information in the area where the personal information is likely to exist;
an erasing step of drawing peripheral pixels of the personal information into the personal information of the first color image and erasing the personal information based on the personal information determined in the determining step and the first color image;
a step of outputting a color image after the erasing step;
A method of masking personal information, comprising:
 前記サイズ変換工程は、取得した前記カラー画像を前記所定サイズの第1グレースケール画像に変換し、
 前記領域検出工程は、前記第1グレースケール画像に対して個人情報が存在する可能性のある領域を検出する、
 請求項1に記載の個人情報マスキング方法。
The size conversion step converts the obtained color image into a first grayscale image of the predetermined size,
The region detection step detects regions in which personal information may exist in the first grayscale image.
The personal information masking method according to claim 1.
 カラー画像を取得する工程と、
 取得した前記カラー画像を所定サイズの第1カラー画像及び第1グレースケール画像に変換するサイズ変換工程と、
 前記第1グレースケール画像に対して個人情報が存在する可能性のある領域を検出する領域検出工程と、
 前記個人情報が存在しそうな領域に対して個人情報の有無を判定する判定工程と、
 前記判定工程で判定された個人情報と前記第1カラー画像とに基づいて、前記個人情報の周辺画素を前記第1カラー画像の前記個人情報に描き入れ前記個人情報を消去する消去工程と、
 前記消去工程後のカラー画像を出力する工程と、
 を備える、個人情報マスキング方法。
obtaining a color image;
a size conversion step of converting the obtained color image into a first color image and a first grayscale image of a predetermined size;
an area detection step of detecting an area in which personal information may exist in the first grayscale image;
a determination step of determining the presence or absence of personal information in the area where the personal information is likely to exist;
an erasing step of drawing peripheral pixels of the personal information into the personal information of the first color image and erasing the personal information based on the personal information determined in the determining step and the first color image;
a step of outputting a color image after the erasing step;
A method of masking personal information, comprising:
 前記第1カラー画像に対して人物の保護領域を検出する工程と、
 前記保護領域を検出した場合に前記保護領域を保存する工程と、
 前記消去工程後の前記第1カラー画像に、前記保護領域を上書きする上書き工程と、
 を備える請求項1から請求項3のいずれか一項に記載の個人情報マスキング方法。
detecting protected areas of a person for the first color image;
storing the protected area when the protected area is detected;
an overwriting step of overwriting the protected area on the first color image after the erasing step;
A personal information masking method according to any one of claims 1 to 3, comprising:
 前記第1カラー画像に対してノイズ除去する工程と、
 前記第1カラー画像に対して色空間変換を処理する工程と、
 前記ノイズ除去する工程及び色空間変換を処理する工程の後に、前記第1カラー画像を第2グレースケール画像に変換する変換工程と、
を備える請求項1から請求項4のいずれか一項に記載の個人情報マスキング方法。
denoising the first color image;
performing a color space transformation on the first color image;
a converting step of converting the first color image into a second grayscale image after the steps of denoising and processing color space conversion;
A personal information masking method according to any one of claims 1 to 4, comprising:
 前記第2グレースケール画像に対して局所的ヒストグラム平坦化を処理し、第3グレースケール画像に変換する工程、を備え、
 前記領域検出工程は、前記前記第2グレースケール画像又は前記第3グレースケール画像に対して個人情報が存在する可能性のある領域を検出する、請求項5に記載の個人情報マスキング方法。
processing a local histogram equalization on the second grayscale image to convert to a third grayscale image;
6. The personal information masking method according to claim 5, wherein said area detection step detects areas in which personal information may exist in said second grayscale image or said third grayscale image.
 前記判定工程は、
(a) 前記個人情報が存在しそうな領域とかかる領域を所定角度回転させた領域との論理積が第1閾値より大きいかに基づいて、個人情報の有無を判定するフィルター、
(b) 前記個人情報が存在しそうな領域の骨組みとなる細線から前記領域の境界線までの距離の標準偏差が第2閾値より大きいかに基づいて、個人情報の有無を判定するフィルター、
(c) 前記個人情報が存在しそうな領域の面積と前記個人情報が存在しそうな領域を取り囲む面積との面積比が第3閾値より大きいかに基づいて、個人情報の有無を判定するフィルター、
(d) 前記個人情報が存在しそうな領域の縦長さ又は横長さが前記所定サイズに対する第4閾値より大きいかに基づいて、個人情報の有無を判定するフィルター、
(e) 前記個人情報が存在しそうな領域の骨組みとなる細線の面積と前記個人情報が存在しそうな領域の面積との比が第5閾値より大きいかに基づいて、個人情報の有無を判定するフィルター、
 の少なくとも1つのフィリター処理を適用する、請求項1から請求項6のいずれか一項に記載の個人情報マスキング方法。
The determination step includes
(a) a filter for determining the presence or absence of personal information based on whether the logical product of the area where the personal information is likely to exist and the area obtained by rotating the area by a predetermined angle is greater than a first threshold;
(b) a filter that determines the presence or absence of personal information based on whether the standard deviation of the distance from the thin line that forms the framework of the area where the personal information is likely to exist to the boundary line of the area is greater than a second threshold;
(c) a filter for determining the presence or absence of personal information based on whether the area ratio of the area where the personal information is likely to exist and the area surrounding the area where the personal information is likely to exist is greater than a third threshold;
(d) a filter for determining the presence or absence of personal information based on whether the vertical or horizontal length of the area where the personal information is likely to exist is greater than a fourth threshold for the predetermined size;
(e) Determining the presence or absence of personal information based on whether the ratio of the area of the thin line that forms the framework of the area where the personal information is likely to exist and the area of the area where the personal information is likely to exist is greater than a fifth threshold. filter,
7. The method of masking personal information according to any one of claims 1 to 6, applying at least one filter process of .
 前記第1カラー画像に対して個人情報を検出する個人情報検出工程を、備える、請求項1から請求項7のいずれか一項に記載の個人情報マスキング方法。 The personal information masking method according to any one of claims 1 to 7, comprising a personal information detection step of detecting personal information for the first color image.  前記上書き工程後に、エッジ保存フィルターにより仕上げ処理する仕上げ工程、
 を備える請求項4に記載の個人情報マスキング方法。
After the overwriting step, a finishing step of finishing with an edge preserving filter;
5. The personal information masking method of claim 4, comprising:
 所定サイズの第1カラー画像を取得する取得部と、
 前記第1カラー画像を二値画像に変換する二値変換部と、
 前記二値画像に対して、個人情報が存在する可能性がある領域を検出する領域検出工程と、
 前記個人情報が存在しそうな領域に対して個人情報の有無を判定する判定部と、
 前記判定部で判定された個人情報と前記第1カラー画像とに基づいて、前記個人情報の周辺画素を前記第1カラー画像の前記個人情報に描き入れ前記個人情報を消去する消去部と、
 前記消去後のカラー画像を出力する出力部と、
 を備える、個人情報マスキング装置。
an acquisition unit that acquires a first color image of a predetermined size;
a binary conversion unit that converts the first color image into a binary image;
a region detection step of detecting a region in which personal information may exist in the binary image;
a determination unit that determines the presence or absence of personal information in the area where the personal information is likely to exist;
an erasing unit that draws peripheral pixels of the personal information into the personal information of the first color image and erases the personal information based on the personal information determined by the determining unit and the first color image;
an output unit that outputs the erased color image;
A personal information masking device comprising:
 前記判定部は、
(a) 前記個人情報が存在しそうな領域とかかる領域を所定角度回転させた領域との論理積が第1閾値より大きいかに基づいて、個人情報の有無を判定するフィルター、
(b) 前記個人情報が存在しそうな領域の骨組みとなる細線から前記領域の境界線までの距離の標準偏差が第2閾値より大きいかに基づいて、個人情報の有無を判定するフィルター、
(c) 前記個人情報が存在しそうな領域の面積と前記個人情報が存在しそうな領域を取り囲む面積との面積比が第3閾値より大きいかに基づいて、個人情報の有無を判定するフィルター、
(d) 前記個人情報が存在しそうな領域の縦長さ又は横長さが前記所定サイズに対する第4閾値より大きいかに基づいて、個人情報の有無を判定するフィルター、
(e) 前記個人情報が存在しそうな領域の骨組みとなる細線の面積と前記個人情報が存在しそうな領域の面積との比が第5閾値より大きいかに基づいて、個人情報の有無を判定するフィルター、
 の少なくとも1つのフィリターを適用する、請求項9に記載の個人情報マスキング装置。
The determination unit
(a) a filter for determining the presence or absence of personal information based on whether the logical product of the area where the personal information is likely to exist and the area obtained by rotating the area by a predetermined angle is greater than a first threshold;
(b) a filter that determines the presence or absence of personal information based on whether the standard deviation of the distance from the thin line that forms the framework of the area where the personal information is likely to exist to the boundary line of the area is greater than a second threshold;
(c) a filter for determining the presence or absence of personal information based on whether the area ratio of the area where the personal information is likely to exist and the area surrounding the area where the personal information is likely to exist is greater than a third threshold;
(d) a filter for determining the presence or absence of personal information based on whether the vertical or horizontal length of the area where the personal information is likely to exist is greater than a fourth threshold for the predetermined size;
(e) Determining the presence or absence of personal information based on whether the ratio of the area of the thin line that forms the framework of the area where the personal information is likely to exist and the area of the area where the personal information is likely to exist is greater than a fifth threshold. filter,
10. The personal information masking device of claim 9, applying at least one filter of .
 前記第1カラー画像に対して人物の保護領域を検出する保護領域検出部をさらに備え、該保護領域検出部は、
  i)ディープラーニングを使用した顔領域の検出する機能、
 ii)ディープラーニングを使用した特定物体の検出する機能、
iii)前記第1カラー画像にある皮膚の色による皮膚領域の検出する機能、
 iv)前記第1カラー画像に保護すべき保護枠を設定させて、その保護枠内の領域を検出する機能、
 の少なくとも1つの機能を適用する、請求項10又は請求項11に記載の個人情報マスキング装置。
 

 
Further comprising a protected area detection unit for detecting a protected area of a person in the first color image, the protected area detection unit comprising:
i) the ability to detect facial regions using deep learning;
ii) the ability to detect specific objects using deep learning;
iii) detecting skin areas by skin color in said first color image;
iv) a function of setting a protective frame to be protected on the first color image and detecting an area within the protective frame;
12. Personal information masking device according to claim 10 or claim 11, applying at least one function of


PCT/JP2022/023832 2021-07-06 2022-06-14 Personal information masking method, and personal information masking device Ceased WO2023281995A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2021112125A JP7156771B1 (en) 2021-07-06 2021-07-06 Personal information masking method and personal information masking device
JP2021-112125 2021-07-06

Publications (1)

Publication Number Publication Date
WO2023281995A1 true WO2023281995A1 (en) 2023-01-12

Family

ID=83688424

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2022/023832 Ceased WO2023281995A1 (en) 2021-07-06 2022-06-14 Personal information masking method, and personal information masking device

Country Status (2)

Country Link
JP (1) JP7156771B1 (en)
WO (1) WO2023281995A1 (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2023128983A (en) * 2022-03-04 2023-09-14 パイオニア株式会社 Image processing device, image processing method, image processing program, and recording medium
KR102737592B1 (en) * 2023-03-27 2024-12-03 홍익대학교 산학협력단 System and method for masking personal information in real estate property image

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2000324445A (en) * 1999-05-10 2000-11-24 Nri & Ncc Co Ltd Customer behavior image data recording device and customer behavior image data recording method
JP2009015478A (en) * 2007-07-03 2009-01-22 Omron Corp License plate reader
US20140369567A1 (en) * 2006-04-04 2014-12-18 Cyclops Technologies, Inc. Authorized Access Using Image Capture and Recognition System
JP2018061214A (en) * 2016-10-07 2018-04-12 パナソニックIpマネジメント株式会社 Surveillance video analysis system and surveillance video analysis method

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2000324445A (en) * 1999-05-10 2000-11-24 Nri & Ncc Co Ltd Customer behavior image data recording device and customer behavior image data recording method
US20140369567A1 (en) * 2006-04-04 2014-12-18 Cyclops Technologies, Inc. Authorized Access Using Image Capture and Recognition System
JP2009015478A (en) * 2007-07-03 2009-01-22 Omron Corp License plate reader
JP2018061214A (en) * 2016-10-07 2018-04-12 パナソニックIpマネジメント株式会社 Surveillance video analysis system and surveillance video analysis method

Also Published As

Publication number Publication date
JP2023008507A (en) 2023-01-19
JP7156771B1 (en) 2022-10-19

Similar Documents

Publication Publication Date Title
CN106446873B (en) Face detection method and device
JP6956252B2 (en) Facial expression synthesis methods, devices, electronic devices and computer programs
CN103914676B (en) A kind of method and apparatus used in recognition of face
US11023708B2 (en) Within document face verification
US11823358B2 (en) Handwritten content removing method and device and storage medium
Lin Face detection in complicated backgrounds and different illumination conditions by using YCbCr color space and neural network
Nakamura et al. Scene text eraser
US10902283B2 (en) Method and device for determining handwriting similarity
CN108230252B (en) Image processing method and device and electronic equipment
Marciniak et al. Influence of low resolution of images on reliability of face detection and recognition
CN109948566B (en) Double-flow face anti-fraud detection method based on weight fusion and feature selection
EP3035236A1 (en) Four dimensional code, and four dimensional code-based image recognition system and method, and retrieval system and method
WO2019061658A1 (en) Method and device for positioning eyeglass, and storage medium
Song et al. Stylizing face images via multiple exemplars
CN104794693A (en) Human image optimization method capable of automatically detecting mask in human face key areas
JP7156771B1 (en) Personal information masking method and personal information masking device
Rosin et al. Artistic minimal rendering with lines and blocks
Dlshad et al. Kurdish sign language recognition system
CN111666911A (en) Micro-expression data expansion method and device
CN110766001A (en) Bank card number positioning and end-to-end identification method based on CNN and RNN
Sujithra et al. Aadhar Card Based Banking Security System To Protect Unauthorized Access
CN108509885A (en) A kind of efficient identity card picture screening technique
Zhang A selection of image processing techniques: from fundamentals to research front
Bloomberg et al. Document image applications
Menaga et al. Identification of Facial Retouching Using Supervised Deep Learning Algorithm

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22837413

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 22837413

Country of ref document: EP

Kind code of ref document: A1