[go: up one dir, main page]

US20130002885A1 - Image pick-up apparatus and tracking method therefor - Google Patents

Image pick-up apparatus and tracking method therefor Download PDF

Info

Publication number
US20130002885A1
US20130002885A1 US13/608,654 US201213608654A US2013002885A1 US 20130002885 A1 US20130002885 A1 US 20130002885A1 US 201213608654 A US201213608654 A US 201213608654A US 2013002885 A1 US2013002885 A1 US 2013002885A1
Authority
US
United States
Prior art keywords
subject
tracking
holding time
display
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US13/608,654
Inventor
Takaaki Fukui
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Canon Inc
Original Assignee
Canon Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Canon Inc filed Critical Canon Inc
Priority to US13/608,654 priority Critical patent/US20130002885A1/en
Publication of US20130002885A1 publication Critical patent/US20130002885A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • G06V40/162Detection; Localisation; Normalisation using pixel segmentation or colour matching
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/61Control of cameras or camera modules based on recognised objects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/61Control of cameras or camera modules based on recognised objects
    • H04N23/611Control of cameras or camera modules based on recognised objects where the recognised objects include parts of the human body
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/63Control of cameras or camera modules by using electronic viewfinders
    • H04N23/633Control of cameras or camera modules by using electronic viewfinders for displaying additional information relating to control or operation of the camera
    • H04N23/635Region indicators; Field of view indicators
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/695Control of camera direction for changing a field of view, e.g. pan, tilt or based on tracking of objects

Definitions

  • the present invention relates to an image pick-up apparatus that generates a captured image, and a tracking method therefor.
  • an image pick-up system that tracks a subject has been known as a monitoring apparatus using a digital camera or a surveillance camera.
  • JP06-054326A discloses a system that calculates a motion vector of a subject based on pieces of image data and rotates a camera so as to track the motion of the subject.
  • this system rotates the camera at high speed when a zoom position is telephoto and rotates the camera at low speed when a zoom position is wide.
  • JP2007-104200A discloses that controlling a WB (white balance) and an exposure within a tracking frame that is detected as a flesh color region optimizes performance of controls of a white balance and an exposure.
  • a tracking method using correlation between pieces of frame image data by which a motion vector is calculated, a subject tracking method performed based on image characteristic quantity, etc. are general.
  • the above-mentioned conventional image pick-up apparatuses have the following problems. That is, in the tracking method that uses correlation between pieces of image data as the above-mentioned tracking method, when a moving subject being different from a main subject crosses in front of the main subject, for example, the correlation between pieces of the image data of the main subject decreases, there is high possibility of tracking the moving subject that crossed in front of the main subject.
  • JP2007-104200A discloses a condition for finishing subject tracking. That is, the subject is tracked by using data of an angular velocity sensor and AF (auto-focusing) data, and it is determined that the subject is lost when difference between characteristic quantity of the subject within a tracking region and that at the time of a tracking start becomes large.
  • the system can be controlled without loosing the subject by fixing the tracking region temporarily and by restarting the tracking again when a subject that has high correlation with the original subject is detected again after the lapse of time. This enables to improve the ease-of-use of the apparatus.
  • the image pick-up apparatus provided with such a subject tracking function has a function that generates an image on which a frame (a tracking frame) is superimposed on the tracking region and displays it on a monitoring device, in order that a user can check easily whether the subject tracking is succeeding or not.
  • the tracking frame When a main subject is lost, the tracking frame should be eliminated. However, if the tracking frame is immediately eliminated after loosing, the tracking frame returns immediately when the main subject is found immediately after that, which may serve as rather unseemly display.
  • the present invention provides an image pick-up apparatus and a tracking method therefor that are capable of appropriately setting a holding time in which there is a possibility of returning to a state that allows tracking when a subject cannot be tracked temporarily during the subject tracking to improve the ease-of-use for a user.
  • a first aspect of the present invention provides an image pick-up apparatus that generates a captured image comprising a specifying unit adapted to specify a subject included in the captured image, a display unit adapted to display the captured image on a screen and display identification information showing that the specified subject is tracked, a tracking unit adapted to track the subject, a setting unit adapted to set a holding time in which the display of the identification information is held according to at least one of a focal length of the image pick-up apparatus and a subject distance, and an elimination unit adapted to eliminate a display of the identification information when the holding time has passed after the tracking unit lost the subject.
  • a second aspect of the present invention provides a tracking method for an image pick-up apparatus that generates a captured image, the method comprising a specifying step of the image pick-up apparatus specifying a subject included in the captured image, a display step of the image pick-up apparatus displaying the captured image on a screen and display identification information showing that the specified subject is tracked, a tracking step of the image pick-up apparatus tracking the subject, a setting step of the image pick-up apparatus setting a holding time in which the display of the identification information is held according to at least one of a focal length of the image pick-up apparatus and a subject distance, and an elimination step of the image pick-up apparatus eliminating a display of the identification information when the holding time has passed after the subject has been lost in the tracking step.
  • the holding time to hold the display of identification information showing that the subject is being tracked can be established according to at least one of the focal length and the subject distance, when controlling the display and the elimination of the identification information.
  • the image pick-up apparatus finishes the display of the identification information. Therefore, a holding time in which there is a possibility of returning to a state that allows tracking when a subject cannot be tracked temporarily during the subject tracking can be appropriately established, which improves the ease-of-use for a user.
  • FIG. 1 is a block diagram schematically showing a configuration of an image pick-up apparatus according to an embodiment of the present invention.
  • FIG. 2 is a flowchart showing the first half of a subject tracking operation of the image pick-up apparatus of FIG. 1 .
  • FIG. 3 is a flowchart showing the second half of the subject tracking operation of the image pick-up apparatus of FIG. 1 .
  • FIG. 4 is a flowchart showing a face detection process of the image pick-up apparatus of FIG. 1 .
  • FIG. 5A is a view showing original image data that is a target of the face detection process of the image pick-up apparatus of FIG. 1 .
  • FIG. 5B is a view showing image data of a flesh color region that is extracted from the original image data by the face detection process of the image pick-up apparatus of FIG. 1 .
  • FIG. 5C is a view showing image data that is obtained by applying a high-pass filter to the image data of the flesh color region by the face detection process of the image pick-up apparatus of FIG. 1 .
  • FIG. 5D is a view showing a template matching method with respect to the image data to which the high-pass filter has been applied by the face detection process of the image pick-up apparatus of FIG. 1 .
  • FIG. 6 is a chromaticity diagram showing representative colors in the CIE 1976 Lab color space of the CIELAB.
  • FIG. 7 is a view showing a matrix of an example of a two-dimensional high-pass filter in which factors used for a spatial-filtering process of the image pick-up apparatus of FIG. 1 are shown.
  • FIG. 8 is a flowchart showing procedures corresponding to steps S 4 and S 5 of FIG. 2 for calculating correlation between pieces of the image data.
  • FIG. 9A is a view showing image data of a tracking target region that is an origin for calculating the correlation in the tracking process of FIG. 8 .
  • FIG. 9B is a view showing the next frame image data that is a target for calculating the correlation in the tracking process of FIG. 8 .
  • FIG. 10A shows a table showing a holding time corresponding to a focal length and a subject distance that are bases for calculating the holding time in step S 12 of FIG. 3 .
  • FIG. 10B shows a graph showing a holding time corresponding to a focal length and a subject distance that are bases for calculating the holding time in step S 12 of FIG. 3 .
  • FIG. 1 is a block diagram schematically showing a configuration of an image pick-up apparatus according to the embodiment.
  • the image pick-up apparatus has an image sensor 101 , a buffer memory 102 , a user interface 103 , a compression circuit 104 , a recorder 105 , a recording medium 106 , a system controller 107 , a display control circuit 108 , a D/A converter 109 , and a monitoring device 110 .
  • the image pick-up apparatus has a signal processing circuit 111 , a lens barrel device 112 , a light emitting device 113 , a face detection unit 120 , and a subject tracking unit 121 .
  • a lens barrel device 112 In the lens barrel device 112 , an image stabilizer (IS) 112 a and an angular velocity sensor 112 b are mounted.
  • IS image stabilizer
  • the system controller 107 receives a shooting instruction from a user via the user interface 103 , the system controller 107 shoots by controlling a focal position and an aperture/mechanical shutter of the lens barrel device 112 , the image sensor 101 formed of a CCD, the light emitting device 113 when performing a flash shooting, etc.
  • the image sensor 101 After shooting, the image sensor 101 outputs an image signal, which is stored in the buffer memory 102 . Then, the signal processing circuit 111 applies a signal processing for generating an image to the image signal stored in the buffer memory 102 to generate YUV image data.
  • the generated image data is stored in the buffer memory 102 .
  • the compression circuit compression/expansion circuit
  • the compressed image data is recorded on the recording medium 106 by the recorder (recorder/reader) 105 .
  • the display control circuit 108 When the display control circuit 108 generates display image data (a shot image) based on the YUV image data stored in the buffer memory 102 , the shot image is converted by the D/A converter 109 and is displayed on the screen of the monitoring device 110 .
  • the face detection unit 120 When performing a face detection from the image data, the face detection unit 120 performs the face detection from the YUV image data stored in the buffer memory 102 or CCD image data, and outputs a coordinate position (location) in the image data.
  • the system controller 107 instructs the display control circuit 108 to display a face frame at the face position. Then, the system controller 107 selects one of the detected faces as a tracking target and instructs the display control circuit 108 to display a tracking frame in place of the face frame at the position of the face selected as the tracking target.
  • the tracking frame is used as identification information by which a user identifies which subject is tracked.
  • the subject tracking unit 121 acquires the image data within the tracking target region, which includes a tracking subject, from the system controller 107 . And the subject tracking unit 121 acquires the position of the tracking target region in a new frame image data by calculating the correlation between pieces of the image data of the tracking target region and newly acquired frame image data.
  • the system controller 107 instructs the display control circuit 108 to change the position of the tracking frame to the position of the tracking target region that is newly acquired by the subject tracking unit 121 in the new frame image data.
  • the system controller 107 receives the face position detected by the face detection unit 120 as face detection information, and sets the image data of the face position as the image data of the tracking target region. And while the subject tracking unit 121 repeatedly detects image data having the highest correlation with the image data within the tracking target region, a subject is tracked. The subject tracking unit 121 does not use the face position that is newly detected by the face detection unit 120 , while the subject is tracked successfully. The subject tracking unit 121 sets the face position detected by the face detection unit 120 as the tracking target region again, when the tracking frame is eliminated due to a failure of the tracking.
  • the present embodiment is described with the assumption that the subject set as the tracking target is a human face, the subject is not limited to the human face. For example, a subject freely specified by a user via the user interface (control panel) 103 may be set as a tracking target.
  • FIG. 2 and FIG. 3 are flowcharts showing the subject tracking operation. The process is executed by the subject tracking unit 121 .
  • the subject tracking unit 121 acquires a tracking target region from the system controller 107 , and acquires the image data within the tracking target region (step S 1 ).
  • the system controller 107 sets the region of the face position detected by the face detection unit 120 as a first tracking target region.
  • the subject tracking unit 121 extracts the image characteristic quantity of the first tracking target region (step S 2 ).
  • the image characteristic quantity is color information such as brightness and a tint, for example.
  • the subject tracking unit 121 acquires next frame image data (step S 3 ).
  • the subject tracking unit 121 detects the region having the highest correlation with the tracking target region from the next frame image data (step S 4 ), and calculates a motion vector (step S 5 ). The details of the calculations of the correlation and the motion vector will be described later.
  • the subject tracking unit 121 moves the position of the tracking target region by the amount of the motion vector and finds the position of a tracking target candidate region, and acquires image data within the tracking target candidate region from the next frame image data (step S 6 ). That is, the subject tracking unit 121 acquires the image data within the tracking target candidate region having the highest correlation with the tracking target region from the next frame image data.
  • the subject tracking unit 121 extracts the image characteristic quantity of the tracking target candidate region in the same manner as step S 2 (step S 7 ).
  • the subject tracking unit 121 compares the image characteristic quantity of the tracking target region acquired in step S 2 with the image characteristic quantity of the tracking target candidate region acquired in step S 7 (step S 8 ). As a result of the comparison, the subject tracking unit 121 determines whether the difference of the image characteristic quantities falls within a predetermined range (step S 9 ). That is, the subject tracking unit 121 determines whether the subject in the tracking target candidate region acquired in step S 6 is the same as the subject in the first tracking target region set up in step S 1 .
  • the subject tracking unit 121 informs the system controller 107 about success of the subject tracking.
  • the system controller 107 updates the tracking target region by replacing with the tracking target candidate region acquired in step S 6 as a new tracking target region, and instructs the display control circuit 108 to update the display position of the tracking frame based on the position of the updated tracking target region (step S 10 ).
  • the system controller 107 makes the timer stop and resets the holding time.
  • the subject tracking unit 121 returns to the process in step S 3 , and newly acquires next frame image data.
  • the subject tracking unit 121 detects a tracking target candidate region that is most similar to the tracking target region updated in step S 10 from the next frame image data in steps S 4 through S 6 , and calculates the image characteristic quantity of the tracking target candidate region in step S 7 .
  • the subject tracking unit 121 compares the image characteristic quantity of the tracking target candidate region with the image characteristic quantity of the initial tracking target region calculated in step S 2 .
  • the system controller 107 updates the tracking target region again if the difference of the quantities falls within the predetermined range.
  • the subject tracking unit 121 detects a tracking target candidate region that has high correlation with the tracking target region from a new frame image data. And if the tracking target candidate region and the initial tracking target region is similar in the image characteristic quantity, the subject tracking unit 121 updates the tracking target region by replacing with the tracking target candidate region. The subject tracking unit 121 can continue the tracking by updating the tracking target region, even if a direction of a subject changes gradually or a shape of a subject changes gradually. However, if the difference of the image characteristic quantities between the tracking target candidate region and the initial tracking target region exceeds the predetermined range, there is a high possibility that a subject being different from the initial tracking subject has been detected as a tracking target candidate region. In such a case, it is determined that the tracking has failed.
  • step S 9 if the difference of the image characteristic quantities does not fall within the predetermined range, the subject tracking unit 121 determines that the tracking has failed, and proceeds with the process to a subject tracking failure routine in and after step S 11 .
  • the subject tracking unit 121 determines whether the tracking has succeeded in the previous frame image data (step S 11 ). If the tracking has succeeded in the previous frame image data, the subject tracking unit 121 determines (establishes) the holding time on the assumption that the tracking has failed for the first time in the frame of the present stage (step S 12 ).
  • the holding time is a time for holding the display of the tracking frame as-is even if a tracking fails. As mentioned below, the holding time is established by acquiring a focal length controlled in the lens barrel device 112 , a subject position and a subject distance detected before the previous frame image data, control information of the image stabilizer (IS) 112 a, an output of the angular velocity sensor 112 b, and the like.
  • the subject tracking unit 121 starts the built-in timer (step S 13 ), and proceeds with the process to step S 14 .
  • the subject tracking unit 121 determines whether the elapsed time of the timer exceeds the predetermined holding time (step S 14 ) on the assumption that the tracking has failed subsequent to the failure in the previous frame image data.
  • the subject tracking unit 121 acquires next frame image data (step S 16 ) while making the display control circuit 108 to continue the display of the tracking frame in the same position, and returns to the process in step S 6 .
  • the subject tracking unit 121 proceeds with the process from step S 16 to step S 6 , the finally updated tracking target region is set as the tracking target candidate region.
  • the subject tracking unit 121 determines to eliminate the tracking frame (step S 15 ), and finishes the process.
  • the determination of elimination of the tracking frame is transmitted to the system controller 107 from the subject tracking unit 121 .
  • the system controller 107 sends the instruction to eliminate the tracking frame to the display control circuit 108 according to the determination, and then, the tracking frame is eliminated. This finishes the tracking.
  • a face detection method by the face detection unit 120 will be described.
  • Many methods are proposed as a face detection method.
  • a method using a learning represented by the neural network a method using template matching of regions including a part characterized in physical shape such as an eye and a nose, a method using statistical analysis of detected image characteristic quantity such as a skin color and an eye shape.
  • a face detection method combines these methods.
  • FIG. 4 is a flowchart showing a face detection process as an example of them. The process is executed by the face detection unit 120 as mentioned above.
  • FIGS. 5A through 5D are views useful in explaining a face detecting method.
  • FIG. 5A shows target image data (an original image) from which a face is detected.
  • the face detection unit 120 performs a flesh color region extraction process, and extracts a flesh color region from the target image data (step S 21 ).
  • FIG. 6 is a chromaticity diagram showing representative colors in the CIE 1976 Lab color space of the CIELAB. The area shown by an ellipse E 1 has a high possibility of a flesh color region.
  • FIG. 5B shows the image data of a region in which chromaticity is the flesh color that has been extracted from the target image data by the process in step S 21 .
  • the face detection unit 120 performs a spatial-filtering process in which a high-pass filter is applied to the image data of the extracted flesh color region (step S 22 ).
  • FIG. 7 is a view showing a matrix of an example of a two-dimensional high-pass filter in which factors used for a spatial-filtering process are shown.
  • FIG. 5C is a view showing image data that is acquired by applying the high-pass filter of FIG. 7 to the image data of FIG. 5B . Applying the high-pass filter emphasizes outlines of parts having variations such as an eye or a nose among the image data.
  • FIG. 5D is a view showing a template matching method applied to the image data of FIG. 5C that shows a result of applying the high-pass filter.
  • the face detection unit 120 recognizes the face based on the positional relationship of the detected eye regions, and extracts characteristic quantities, such as a direction and a size (step S 24 ). Then, the face detection unit 120 finishes the process. It should be noted that the process may be modified by omitting the step S 21 , applying the high-pass filter to the entire region of the target image data in step S 22 , and recognizing the face by performing the template matching in step S 23 .
  • the face detection information that includes the characteristic quantities such as a direction and a size extracted in step S 24 is transmitted to the system controller 107 from the face detection unit 120 .
  • the system controller 107 determines the region (tracking frame) that serves as a tracking target from the face detection information.
  • FIG. 8 is a flowchart showing a process for calculating correlation between pieces of image data. This process corresponds to the process in steps S 4 and S 5 of the tracking process shown in FIG. 2 .
  • FIG. 9A and FIG. 9B are views showing the image data of the tracking target region in the previous frame image data, and the next frame image data, respectively.
  • the subject tracking unit 121 acquires image data 201 within the tracking target region that is an origin for calculating the correlation (step S 31 ). It should be noted that the tracking frame 210 is established so as to surround the image data 201 within the tracking target region as shown in FIG. 9A .
  • the subject tracking unit 121 acquires the position of a comparison region that exists at the same position as the image data 201 within the tracking target region in frame image data 300 that is acquired at the next timing of acquisition of the image data 201 within the tracking target region (step S 32 ).
  • the subject tracking unit 121 shifts the position of the comparison region in a vertical, horizontal, or oblique direction within a predetermined region of the frame image data 300 , and acquires the image data within the comparison region at the shifted position (step S 33 ).
  • the subject tracking unit 121 calculates the image difference between the image data within the comparison region acquired in step S 33 and the image data 201 within the tracking target region (step S 34 ).
  • the subject tracking unit 121 acquires the integral value of the absolute value of the image difference calculated in step S 34 (step S 35 ).
  • the subject tracking unit 121 determines whether the correlation between pieces of the image data is the maximum among the values that have been acquired till then (step S 36 ). When the correlation between pieces of the image data is the maximum, the subject tracking unit 121 saves the correlation and the position of the comparison region (step S 37 ). On the other hand, when correlation between pieces of the image data is not the maximum, the subject tracking unit 121 proceeds with the process to step S 38 as-is.
  • the subject tracking unit 121 determines whether the position of the comparison region has been moved in step S 33 to all the positions established beforehand, including the position acquired in step S 32 (step S 38 ). When the comparison region has not been moved to all the positions, the subject tracking unit 121 returns the process to step S 33 . On the other hand, when the comparison region has been moved to all the positions, the subject tracking unit 121 calculates the shift amount (motion vector) to the position with the highest correlation and the correlation (step S 39 ). In FIG. 9B , the image data 301 shows the image data within the comparison region that is determined to have the highest correlation with the image data 201 within the tracking target region. Then, the subject tracking unit 121 finishes the process. It should be noted that there are many methods of calculating the motion vector, and the method is not limited to the above-mentioned method.
  • FIGS. 10A and 10B show the table and graph that show the holding time corresponding to a focal length and a subject distance.
  • FIGS. 10A and 10B show the table and graph that show the holding time corresponding to a focal length and a subject distance.
  • the holding time is set based on a combination of a subject distance and a focal length.
  • the same effect is acquired by adjusting the holding time according to a size of a main subject within a field of a captured image instead of the subject distance.
  • the holding time may be determined according to the combination of the focal length and the position at which the subject tracking failed within a field of a captured image. That is, when the distance between a position of the subject detected before the previous frame and a center of the field of the captured image is longer (i.e., when the subject is located in periphery), the possibility that a main subject goes out from the field of view becomes higher. In such a case, the short holding time is preferably set.
  • the output from the angular velocity sensor or the IS is larger than a predetermined value, it is determined that the main subject goes out from the field of view, and thus the short holding time may be set.
  • the long holding time may be set. Whether the tripod is equipped or not is determined based on the output from the angular velocity sensor mounted on the image pick-up apparatus. When correlation between pieces of the image data in parts other than a main subject is extremely high, it may be determined that a tripod is equipped, and the long holding time may be set.
  • the holding time may be set not only according to a focal length or a subject distance but also according to subject brightness, a color temperature, or sensitivity of the image sensor. For example, when the sensitivity of the image sensor is set highly, a noise increases, and when the subject brightness is low, the correlation between the subjects decreases.
  • the holding time is set shorter.
  • a low-intensity light source such as an electric bulb, is used in general, and it is assumed that the subject brightness is low. Therefore, the holding time may be set according to the color temperature.
  • the image pick-up apparatus of the embodiment controls a display and an elimination of the tracking frame showing that a subject is being tracked, during the shoot setting and the signal processing optimal to the subject using the subject tracking function. Further, since the image pick-up apparatus controls the holding time during which the tracking frame is displayed in spite of failure of the subject tracking according to the focal length etc., the ease-of-use for a user can be improved.
  • the image pick-up apparatus sets the holding time according to the focal length when controlling the display and elimination of the tracking frame showing that the subject is being tracked.
  • the image pick-up apparatus finishes the display of the tracking frame. Therefore, a holding time in which there is a possibility of returning to a state that allows tracking when a subject cannot be tracked temporarily during the subject tracking can be appropriately established, which improves the ease-of-use for a user.
  • the image pick-up apparatus can counter the problem that a main subject easily goes out from a field of view as the subject distance becomes shorter. Further, the image pick-up apparatus can counter the problem that a main subject easily goes out from a field of view as the distance between the position at which the subject tracking has failed and a center of the field of the captured image becomes larger. Further, the image pick-up apparatus can counter the problem that a main subject easily goes out from a field of view as the shake of the apparatus becomes larger. Further, since a subject is specified automatically or specified by a user, it becomes easier to specify the subject. Further, since the tracking frame is displayed to indicate the subject of the tracking target, it is easy to identify the subject region.
  • the face detection unit 120 performs a face detection and specifies the region including the face position as a subject region in the above-mentioned embodiment
  • the user appointed region that is appointed by a user via the user interface 103 can be specified, as mentioned above.
  • the COD element is used as the image sensor in the above-mentioned embodiment, a CMOS device etc. may be used.
  • the monitoring (display) device is not limited to a liquid crystal display, but an organic electroluminescence display, a surface electric-field display (SED), a plasma display, etc. may be used.
  • the present invention is applicable to a compact type digital camera, a digital video camera, a digital SLR (single-lens reflex) camera, etc. as an image pick-up apparatus.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Theoretical Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Studio Devices (AREA)
  • Image Analysis (AREA)
  • Automatic Focus Adjustment (AREA)
  • Image Processing (AREA)
  • Focusing (AREA)

Abstract

An image pick-up apparatus that enables to appropriately set a holding time in which there is a possibility of returning to a state that allows tracking when a subject cannot be tracked temporarily during the subject tracking to improve the ease-of-use for a user. A specifying unit specifies a subject included in a captured image. A display unit displays the captured image on a screen and displays identification information showing that the specified subject is tracked. A tracking unit tracks the subject. A setting unit sets a holding time in which the display of the identification information is held according to at least one of a focal length of the image pick-up apparatus and a subject distance. An elimination unit eliminates a display of the identification information when the holding time has passed after the tracking unit lost the subject.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application is a continuation of application Ser. No. 12/547, 168, filed Aug. 25, 2009 the entire disclosure of which is hereby incorporated by reference.
  • BACKGROUND OF THE INVENTION
  • 1. Field of the Invention
  • The present invention relates to an image pick-up apparatus that generates a captured image, and a tracking method therefor.
  • 2. Description of the Related Art
  • Conventionally, an image pick-up system that tracks a subject has been known as a monitoring apparatus using a digital camera or a surveillance camera.
  • For example, Japanese laid-open patent publication (Kokai) No. H06-054326 (JP06-054326A) discloses a system that calculates a motion vector of a subject based on pieces of image data and rotates a camera so as to track the motion of the subject. In order to track and shoot the subject effectively, this system rotates the camera at high speed when a zoom position is telephoto and rotates the camera at low speed when a zoom position is wide.
  • Japanese laid-open patent publication (Kokai) No. 2007-104200 (JP2007-104200A) discloses that controlling a WB (white balance) and an exposure within a tracking frame that is detected as a flesh color region optimizes performance of controls of a white balance and an exposure.
  • Thus, various applications, such as the subject tracking system and the apparatus to control a white balance and an exposure using the subject tracking function, have been proposed.
  • As these tracking methods, a tracking method using correlation between pieces of frame image data by which a motion vector is calculated, a subject tracking method performed based on image characteristic quantity, etc. are general.
  • However, the above-mentioned conventional image pick-up apparatuses have the following problems. That is, in the tracking method that uses correlation between pieces of image data as the above-mentioned tracking method, when a moving subject being different from a main subject crosses in front of the main subject, for example, the correlation between pieces of the image data of the main subject decreases, there is high possibility of tracking the moving subject that crossed in front of the main subject.
  • In the tracking method based on image characteristic quantity such as a tint, when a subject similar to a main subject exists in the surroundings of the main subject, it is difficult to continue tracking the main subject.
  • Also in the tracking method that is a combination of these methods, when the weak conditions for the respective methods are overlapped, it is difficult to continue tracking the main subject.
  • Japanese laid-open patent publication (Kokai) No. 2007-104200 (JP2007-104200A) discloses a condition for finishing subject tracking. That is, the subject is tracked by using data of an angular velocity sensor and AF (auto-focusing) data, and it is determined that the subject is lost when difference between characteristic quantity of the subject within a tracking region and that at the time of a tracking start becomes large.
  • However, even when a subject cannot be tracked temporarily, for example, when a car or the like crosses in front of the subject, or when a detected face as the subject turns sideway temporarily, there is a sufficient possibility of returning to the state that allows the tracking within a predetermined time.
  • In such a case, if a time during which the subject cannot be tracked is temporarily short, the system can be controlled without loosing the subject by fixing the tracking region temporarily and by restarting the tracking again when a subject that has high correlation with the original subject is detected again after the lapse of time. This enables to improve the ease-of-use of the apparatus.
  • Here, the image pick-up apparatus provided with such a subject tracking function has a function that generates an image on which a frame (a tracking frame) is superimposed on the tracking region and displays it on a monitoring device, in order that a user can check easily whether the subject tracking is succeeding or not.
  • When a main subject is lost, the tracking frame should be eliminated. However, if the tracking frame is immediately eliminated after loosing, the tracking frame returns immediately when the main subject is found immediately after that, which may serve as rather unseemly display.
  • Therefore, when the main subject is lost temporarily, the display of the tracking frame should be continued, but when the main subject is completely lost, the display of the tracking frame should be eliminated. However, the decision thereof is extremely difficult.
  • SUMMARY OF THE INVENTION
  • The present invention provides an image pick-up apparatus and a tracking method therefor that are capable of appropriately setting a holding time in which there is a possibility of returning to a state that allows tracking when a subject cannot be tracked temporarily during the subject tracking to improve the ease-of-use for a user.
  • Accordingly, a first aspect of the present invention provides an image pick-up apparatus that generates a captured image comprising a specifying unit adapted to specify a subject included in the captured image, a display unit adapted to display the captured image on a screen and display identification information showing that the specified subject is tracked, a tracking unit adapted to track the subject, a setting unit adapted to set a holding time in which the display of the identification information is held according to at least one of a focal length of the image pick-up apparatus and a subject distance, and an elimination unit adapted to eliminate a display of the identification information when the holding time has passed after the tracking unit lost the subject.
  • Accordingly, a second aspect of the present invention provides a tracking method for an image pick-up apparatus that generates a captured image, the method comprising a specifying step of the image pick-up apparatus specifying a subject included in the captured image, a display step of the image pick-up apparatus displaying the captured image on a screen and display identification information showing that the specified subject is tracked, a tracking step of the image pick-up apparatus tracking the subject, a setting step of the image pick-up apparatus setting a holding time in which the display of the identification information is held according to at least one of a focal length of the image pick-up apparatus and a subject distance, and an elimination step of the image pick-up apparatus eliminating a display of the identification information when the holding time has passed after the subject has been lost in the tracking step.
  • According to the present invention, the holding time to hold the display of identification information showing that the subject is being tracked can be established according to at least one of the focal length and the subject distance, when controlling the display and the elimination of the identification information. When the holding time lapses after loosing the subject, the image pick-up apparatus finishes the display of the identification information. Therefore, a holding time in which there is a possibility of returning to a state that allows tracking when a subject cannot be tracked temporarily during the subject tracking can be appropriately established, which improves the ease-of-use for a user.
  • The features and advantages of the invention will become more apparent from the following detailed description taken in conjunction with the accompanying drawings.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a block diagram schematically showing a configuration of an image pick-up apparatus according to an embodiment of the present invention.
  • FIG. 2 is a flowchart showing the first half of a subject tracking operation of the image pick-up apparatus of FIG. 1.
  • FIG. 3 is a flowchart showing the second half of the subject tracking operation of the image pick-up apparatus of FIG. 1.
  • FIG. 4 is a flowchart showing a face detection process of the image pick-up apparatus of FIG. 1.
  • FIG. 5A is a view showing original image data that is a target of the face detection process of the image pick-up apparatus of FIG. 1.
  • FIG. 5B is a view showing image data of a flesh color region that is extracted from the original image data by the face detection process of the image pick-up apparatus of FIG. 1.
  • FIG. 5C is a view showing image data that is obtained by applying a high-pass filter to the image data of the flesh color region by the face detection process of the image pick-up apparatus of FIG. 1.
  • FIG. 5D is a view showing a template matching method with respect to the image data to which the high-pass filter has been applied by the face detection process of the image pick-up apparatus of FIG. 1.
  • FIG. 6 is a chromaticity diagram showing representative colors in the CIE 1976 Lab color space of the CIELAB.
  • FIG. 7 is a view showing a matrix of an example of a two-dimensional high-pass filter in which factors used for a spatial-filtering process of the image pick-up apparatus of FIG. 1 are shown.
  • FIG. 8 is a flowchart showing procedures corresponding to steps S4 and S5 of FIG. 2 for calculating correlation between pieces of the image data.
  • FIG. 9A is a view showing image data of a tracking target region that is an origin for calculating the correlation in the tracking process of FIG. 8.
  • FIG. 9B is a view showing the next frame image data that is a target for calculating the correlation in the tracking process of FIG. 8.
  • FIG. 10A shows a table showing a holding time corresponding to a focal length and a subject distance that are bases for calculating the holding time in step S12 of FIG. 3.
  • FIG. 10B shows a graph showing a holding time corresponding to a focal length and a subject distance that are bases for calculating the holding time in step S12 of FIG. 3.
  • DESCRIPTION OF THE EMBODIMENTS
  • Hereafter, an embodiment of an image pick-up apparatus and a tracking method thereof according to the present invention will be described in detail with reference to the attached drawings. FIG. 1 is a block diagram schematically showing a configuration of an image pick-up apparatus according to the embodiment. The image pick-up apparatus has an image sensor 101, a buffer memory 102, a user interface 103, a compression circuit 104, a recorder 105, a recording medium 106, a system controller 107, a display control circuit 108, a D/A converter 109, and a monitoring device 110. Further, the image pick-up apparatus has a signal processing circuit 111, a lens barrel device 112, a light emitting device 113, a face detection unit 120, and a subject tracking unit 121. In the lens barrel device 112, an image stabilizer (IS) 112 a and an angular velocity sensor 112 b are mounted.
  • Receiving a shooting instruction from a user via the user interface 103, the system controller 107 shoots by controlling a focal position and an aperture/mechanical shutter of the lens barrel device 112, the image sensor 101 formed of a CCD, the light emitting device 113 when performing a flash shooting, etc.
  • After shooting, the image sensor 101 outputs an image signal, which is stored in the buffer memory 102. Then, the signal processing circuit 111 applies a signal processing for generating an image to the image signal stored in the buffer memory 102 to generate YUV image data.
  • The generated image data is stored in the buffer memory 102. When recording the image data, it is transmitted from the buffer memory 102 to the compression circuit (compression/expansion circuit) 104 and is compressed as a JPEG file. The compressed image data is recorded on the recording medium 106 by the recorder (recorder/reader) 105.
  • When the display control circuit 108 generates display image data (a shot image) based on the YUV image data stored in the buffer memory 102, the shot image is converted by the D/A converter 109 and is displayed on the screen of the monitoring device 110.
  • When performing a face detection from the image data, the face detection unit 120 performs the face detection from the YUV image data stored in the buffer memory 102 or CCD image data, and outputs a coordinate position (location) in the image data.
  • When the face detection unit 120 detects a face, the system controller 107 instructs the display control circuit 108 to display a face frame at the face position. Then, the system controller 107 selects one of the detected faces as a tracking target and instructs the display control circuit 108 to display a tracking frame in place of the face frame at the position of the face selected as the tracking target. The tracking frame is used as identification information by which a user identifies which subject is tracked.
  • When the subject tracking is performed while capturing moving images or using an EVF (Electric View Finder), the subject tracking unit 121 acquires the image data within the tracking target region, which includes a tracking subject, from the system controller 107. And the subject tracking unit 121 acquires the position of the tracking target region in a new frame image data by calculating the correlation between pieces of the image data of the tracking target region and newly acquired frame image data. The system controller 107 instructs the display control circuit 108 to change the position of the tracking frame to the position of the tracking target region that is newly acquired by the subject tracking unit 121 in the new frame image data.
  • According to the embodiment, the system controller 107 receives the face position detected by the face detection unit 120 as face detection information, and sets the image data of the face position as the image data of the tracking target region. And while the subject tracking unit 121 repeatedly detects image data having the highest correlation with the image data within the tracking target region, a subject is tracked. The subject tracking unit 121 does not use the face position that is newly detected by the face detection unit 120, while the subject is tracked successfully. The subject tracking unit 121 sets the face position detected by the face detection unit 120 as the tracking target region again, when the tracking frame is eliminated due to a failure of the tracking. Although the present embodiment is described with the assumption that the subject set as the tracking target is a human face, the subject is not limited to the human face. For example, a subject freely specified by a user via the user interface (control panel) 103 may be set as a tracking target.
  • In the image pick-up apparatus that has the above-mentioned configuration, the method of tracking the face detected by the face detection unit 120 will be described. FIG. 2 and FIG. 3 are flowcharts showing the subject tracking operation. The process is executed by the subject tracking unit 121.
  • The subject tracking unit 121 acquires a tracking target region from the system controller 107, and acquires the image data within the tracking target region (step S1). As mentioned above, in this embodiment, the system controller 107 sets the region of the face position detected by the face detection unit 120 as a first tracking target region.
  • The subject tracking unit 121 extracts the image characteristic quantity of the first tracking target region (step S2). The image characteristic quantity is color information such as brightness and a tint, for example.
  • The subject tracking unit 121 acquires next frame image data (step S3). The subject tracking unit 121 detects the region having the highest correlation with the tracking target region from the next frame image data (step S4), and calculates a motion vector (step S5). The details of the calculations of the correlation and the motion vector will be described later.
  • The subject tracking unit 121 moves the position of the tracking target region by the amount of the motion vector and finds the position of a tracking target candidate region, and acquires image data within the tracking target candidate region from the next frame image data (step S6). That is, the subject tracking unit 121 acquires the image data within the tracking target candidate region having the highest correlation with the tracking target region from the next frame image data. The subject tracking unit 121 extracts the image characteristic quantity of the tracking target candidate region in the same manner as step S2 (step S7).
  • The subject tracking unit 121 compares the image characteristic quantity of the tracking target region acquired in step S2 with the image characteristic quantity of the tracking target candidate region acquired in step S7 (step S8). As a result of the comparison, the subject tracking unit 121 determines whether the difference of the image characteristic quantities falls within a predetermined range (step S9). That is, the subject tracking unit 121 determines whether the subject in the tracking target candidate region acquired in step S6 is the same as the subject in the first tracking target region set up in step S1.
  • When the difference of the image characteristic quantities falls within the predetermined range, the subject tracking unit 121 informs the system controller 107 about success of the subject tracking. The system controller 107 updates the tracking target region by replacing with the tracking target candidate region acquired in step S6 as a new tracking target region, and instructs the display control circuit 108 to update the display position of the tracking frame based on the position of the updated tracking target region (step S10). When a timer has started in step S13, mentioned later, the system controller 107 makes the timer stop and resets the holding time.
  • Then, the subject tracking unit 121 returns to the process in step S3, and newly acquires next frame image data. The subject tracking unit 121 detects a tracking target candidate region that is most similar to the tracking target region updated in step S10 from the next frame image data in steps S4 through S6, and calculates the image characteristic quantity of the tracking target candidate region in step S7. The subject tracking unit 121 compares the image characteristic quantity of the tracking target candidate region with the image characteristic quantity of the initial tracking target region calculated in step S2. The system controller 107 updates the tracking target region again if the difference of the quantities falls within the predetermined range.
  • Thus, the subject tracking unit 121 detects a tracking target candidate region that has high correlation with the tracking target region from a new frame image data. And if the tracking target candidate region and the initial tracking target region is similar in the image characteristic quantity, the subject tracking unit 121 updates the tracking target region by replacing with the tracking target candidate region. The subject tracking unit 121 can continue the tracking by updating the tracking target region, even if a direction of a subject changes gradually or a shape of a subject changes gradually. However, if the difference of the image characteristic quantities between the tracking target candidate region and the initial tracking target region exceeds the predetermined range, there is a high possibility that a subject being different from the initial tracking subject has been detected as a tracking target candidate region. In such a case, it is determined that the tracking has failed.
  • On the other hand, returning to step S9, if the difference of the image characteristic quantities does not fall within the predetermined range, the subject tracking unit 121 determines that the tracking has failed, and proceeds with the process to a subject tracking failure routine in and after step S11.
  • The subject tracking unit 121 determines whether the tracking has succeeded in the previous frame image data (step S11). If the tracking has succeeded in the previous frame image data, the subject tracking unit 121 determines (establishes) the holding time on the assumption that the tracking has failed for the first time in the frame of the present stage (step S12). The holding time is a time for holding the display of the tracking frame as-is even if a tracking fails. As mentioned below, the holding time is established by acquiring a focal length controlled in the lens barrel device 112, a subject position and a subject distance detected before the previous frame image data, control information of the image stabilizer (IS) 112 a, an output of the angular velocity sensor 112 b, and the like.
  • The subject tracking unit 121 starts the built-in timer (step S13), and proceeds with the process to step S14.
  • On the other hand, when it is determined that the tracking has failed also in the previous frame image data in step S11, the subject tracking unit 121 determines whether the elapsed time of the timer exceeds the predetermined holding time (step S14) on the assumption that the tracking has failed subsequent to the failure in the previous frame image data.
  • When the elapsed time does not exceed the predetermined holding time, the subject tracking unit 121 acquires next frame image data (step S16) while making the display control circuit 108 to continue the display of the tracking frame in the same position, and returns to the process in step S6. When the subject tracking unit 121 proceeds with the process from step S16 to step S6, the finally updated tracking target region is set as the tracking target candidate region.
  • On the other hand, when the predetermined holding time is exceeded (elapsed), the subject tracking unit 121 determines to eliminate the tracking frame (step S15), and finishes the process. The determination of elimination of the tracking frame is transmitted to the system controller 107 from the subject tracking unit 121. The system controller 107 sends the instruction to eliminate the tracking frame to the display control circuit 108 according to the determination, and then, the tracking frame is eliminated. This finishes the tracking.
  • Next, the face detection method by the face detection unit 120 will be described. Many methods are proposed as a face detection method. Generally, there are a method using a learning represented by the neural network, a method using template matching of regions including a part characterized in physical shape such as an eye and a nose, a method using statistical analysis of detected image characteristic quantity such as a skin color and an eye shape. In general, a face detection method combines these methods.
  • Many face detection methods are proposed at the present time. For example, there are a method for detecting a face using the wavelet transform and image characteristic quantity, a method further combining template matching etc. FIG. 4 is a flowchart showing a face detection process as an example of them. The process is executed by the face detection unit 120 as mentioned above. FIGS. 5A through 5D are views useful in explaining a face detecting method. FIG. 5A shows target image data (an original image) from which a face is detected.
  • First, the face detection unit 120 performs a flesh color region extraction process, and extracts a flesh color region from the target image data (step S21). FIG. 6 is a chromaticity diagram showing representative colors in the CIE 1976 Lab color space of the CIELAB. The area shown by an ellipse E1 has a high possibility of a flesh color region. FIG. 5B shows the image data of a region in which chromaticity is the flesh color that has been extracted from the target image data by the process in step S21.
  • The face detection unit 120 performs a spatial-filtering process in which a high-pass filter is applied to the image data of the extracted flesh color region (step S22).
  • FIG. 7 is a view showing a matrix of an example of a two-dimensional high-pass filter in which factors used for a spatial-filtering process are shown. FIG. 5C is a view showing image data that is acquired by applying the high-pass filter of FIG. 7 to the image data of FIG. 5B. Applying the high-pass filter emphasizes outlines of parts having variations such as an eye or a nose among the image data.
  • Further, the face detection unit 120 performs template matching, and detects the eye in the image data (step S23). FIG. 5D is a view showing a template matching method applied to the image data of FIG. 5C that shows a result of applying the high-pass filter. The face detection unit 120 recognizes the face based on the positional relationship of the detected eye regions, and extracts characteristic quantities, such as a direction and a size (step S24). Then, the face detection unit 120 finishes the process. It should be noted that the process may be modified by omitting the step S21, applying the high-pass filter to the entire region of the target image data in step S22, and recognizing the face by performing the template matching in step S23.
  • The face detection information that includes the characteristic quantities such as a direction and a size extracted in step S24 is transmitted to the system controller 107 from the face detection unit 120. The system controller 107 determines the region (tracking frame) that serves as a tracking target from the face detection information.
  • Next, the method for tracking a subject will be described. Many methods of detecting correlation and a motion vector of a subject are proposed. Here, a comparatively simple method will be described.
  • FIG. 8 is a flowchart showing a process for calculating correlation between pieces of image data. This process corresponds to the process in steps S4 and S5 of the tracking process shown in FIG. 2.
  • FIG. 9A and FIG. 9B are views showing the image data of the tracking target region in the previous frame image data, and the next frame image data, respectively. The subject tracking unit 121 acquires image data 201 within the tracking target region that is an origin for calculating the correlation (step S31). It should be noted that the tracking frame 210 is established so as to surround the image data 201 within the tracking target region as shown in FIG. 9A.
  • Then, the subject tracking unit 121 acquires the position of a comparison region that exists at the same position as the image data 201 within the tracking target region in frame image data 300 that is acquired at the next timing of acquisition of the image data 201 within the tracking target region (step S32).
  • The subject tracking unit 121 shifts the position of the comparison region in a vertical, horizontal, or oblique direction within a predetermined region of the frame image data 300, and acquires the image data within the comparison region at the shifted position (step S33).
  • The subject tracking unit 121 calculates the image difference between the image data within the comparison region acquired in step S33 and the image data 201 within the tracking target region (step S34). The subject tracking unit 121 acquires the integral value of the absolute value of the image difference calculated in step S34 (step S35). Here, the smaller the integral value of the absolute value of the image difference is, the higher the correlation between pieces of the image data is (the correlation between pieces of the image data becomes higher).
  • The subject tracking unit 121 determines whether the correlation between pieces of the image data is the maximum among the values that have been acquired till then (step S36). When the correlation between pieces of the image data is the maximum, the subject tracking unit 121 saves the correlation and the position of the comparison region (step S37). On the other hand, when correlation between pieces of the image data is not the maximum, the subject tracking unit 121 proceeds with the process to step S38 as-is.
  • The subject tracking unit 121 determines whether the position of the comparison region has been moved in step S33 to all the positions established beforehand, including the position acquired in step S32 (step S38). When the comparison region has not been moved to all the positions, the subject tracking unit 121 returns the process to step S33. On the other hand, when the comparison region has been moved to all the positions, the subject tracking unit 121 calculates the shift amount (motion vector) to the position with the highest correlation and the correlation (step S39). In FIG. 9B, the image data 301 shows the image data within the comparison region that is determined to have the highest correlation with the image data 201 within the tracking target region. Then, the subject tracking unit 121 finishes the process. It should be noted that there are many methods of calculating the motion vector, and the method is not limited to the above-mentioned method.
  • Next, the setting of the holding time in step S12 will be described. FIGS. 10A and 10B show the table and graph that show the holding time corresponding to a focal length and a subject distance. In the table in FIG. 10A and the graph in
  • FIG. 10B, the holding time is set based on a combination of a subject distance and a focal length.
  • Here, the longer the focal length is, the shorter the holding time is set. Namely, when zooming up (i.e., when the focal length is long), since there is a high possibility that a subject as a tracking target goes out from a field of view because of a variation of a field of view due to a hand shake etc. or a movement of the subject, the short holding time is set in such a case. Accordingly, if the tracking fails, a time required to shift to a detection of a next subject becomes shorter. Similarly, the shorter the subject distance is, the shorter the holding time is set. That's because when the subject distance is shorter, the possibility that a main subject goes out from a field of view becomes higher. It should be noted that the same effect is acquired by adjusting the holding time according to a size of a main subject within a field of a captured image instead of the subject distance. In such a case, the larger the size of a main subject is, the longer the holding time is.
  • Similarly, the holding time may be determined according to the combination of the focal length and the position at which the subject tracking failed within a field of a captured image. That is, when the distance between a position of the subject detected before the previous frame and a center of the field of the captured image is longer (i.e., when the subject is located in periphery), the possibility that a main subject goes out from the field of view becomes higher. In such a case, the short holding time is preferably set.
  • It is also possible to detect that a main subject goes out from a field of view based on at least one of the focal length, the control information for the image stabilizer (IS), the output of the angular velocity sensor, and the subject position. When the output from the angular velocity sensor or the IS is larger than a predetermined value, it is determined that the main subject goes out from the field of view, and thus the short holding time may be set.
  • When it is determined that a tripod is equipped with the image pick-up apparatus, the long holding time may be set. Whether the tripod is equipped or not is determined based on the output from the angular velocity sensor mounted on the image pick-up apparatus. When correlation between pieces of the image data in parts other than a main subject is extremely high, it may be determined that a tripod is equipped, and the long holding time may be set.
  • The holding time may be set not only according to a focal length or a subject distance but also according to subject brightness, a color temperature, or sensitivity of the image sensor. For example, when the sensitivity of the image sensor is set highly, a noise increases, and when the subject brightness is low, the correlation between the subjects decreases.
  • Therefore, the accuracy of the face detection process or the tracking process decreases. In such a case, since a possibility that the tracking frame is displayed at an incorrect position increases, the holding time is set shorter. When the color temperature is low, a low-intensity light source, such as an electric bulb, is used in general, and it is assumed that the subject brightness is low. Therefore, the holding time may be set according to the color temperature.
  • The image pick-up apparatus of the embodiment controls a display and an elimination of the tracking frame showing that a subject is being tracked, during the shoot setting and the signal processing optimal to the subject using the subject tracking function. Further, since the image pick-up apparatus controls the holding time during which the tracking frame is displayed in spite of failure of the subject tracking according to the focal length etc., the ease-of-use for a user can be improved.
  • Thus, the image pick-up apparatus sets the holding time according to the focal length when controlling the display and elimination of the tracking frame showing that the subject is being tracked. When the established holding time lapses after loosing the subject, the image pick-up apparatus finishes the display of the tracking frame. Therefore, a holding time in which there is a possibility of returning to a state that allows tracking when a subject cannot be tracked temporarily during the subject tracking can be appropriately established, which improves the ease-of-use for a user.
  • Even if the subject is lost temporarily, it can return to tracking state promptly as long as within the holding time.
  • When zooming up (i.e., when the focal length is long), there is a high possibility that a tracking subject goes out from a field of view because of a variation of a field of view due to a hand shake etc. or a movement of the subject. However, when the focal length is longer, the holding time is set shorter, which shortens the time for performing the tracking process. As a result, it is easy to shift to a detection of a next subject.
  • Further, the image pick-up apparatus can counter the problem that a main subject easily goes out from a field of view as the subject distance becomes shorter. Further, the image pick-up apparatus can counter the problem that a main subject easily goes out from a field of view as the distance between the position at which the subject tracking has failed and a center of the field of the captured image becomes larger. Further, the image pick-up apparatus can counter the problem that a main subject easily goes out from a field of view as the shake of the apparatus becomes larger. Further, since a subject is specified automatically or specified by a user, it becomes easier to specify the subject. Further, since the tracking frame is displayed to indicate the subject of the tracking target, it is easy to identify the subject region.
  • While the present invention has been described with reference to exemplary embodiments and it is to be understood that the invention is not limited to the disclosed exemplary embodiments. The scope of the following claims is to be accorded the broadest interpretation so as to encompass all such modifications and equivalent structures and functions.
  • For example, although the face detection unit 120 performs a face detection and specifies the region including the face position as a subject region in the above-mentioned embodiment, the user appointed region that is appointed by a user via the user interface 103 can be specified, as mentioned above.
  • Although the COD element is used as the image sensor in the above-mentioned embodiment, a CMOS device etc. may be used.
  • The monitoring (display) device is not limited to a liquid crystal display, but an organic electroluminescence display, a surface electric-field display (SED), a plasma display, etc. may be used.
  • The present invention is applicable to a compact type digital camera, a digital video camera, a digital SLR (single-lens reflex) camera, etc. as an image pick-up apparatus.
  • This application claims priority from Japanese Patent Application No. 2008-221790, filed on Aug. 29, 2008, and which is hereby incorporated by reference herein in its entirety.

Claims (5)

1. An image pick-up apparatus that generates captured images in series, comprising:
a tracking unit adapted to track a specified subject by extracting a region of the specified subject in the presently generated captured image having high correlation with a region of the specified subject in the precedently generated captured image;
a display unit adapted to display the captured image on a screen and display identification information showing that the specified subject is tracked; and
a setting unit adapted to set a holding time in which the display of the identification information is held,
wherein said display unit continues to display the identification information when the holding time does not lapse after said tracking unit lost the subject, and eliminates a display of the identification information when the holding time has passed after said tracking unit lost the subject, and
wherein said setting unit changes the holding time according to a sensitivity and at least one of control information for an image stabilizer and an output from an angle velocity sensor of the image pick-up apparatus.
2. The image pick-up apparatus according to claim 1, wherein said setting unit sets the shorter holding time when it is determined that the image pick-up apparatus shakes according to at least one of the control information and the output from the angle velocity sensor, as compared with the holding time when the image pick-up apparatus does not shake.
3. The image pick-up apparatus according to claim 1, wherein said setting unit sets the holding time according to the position at which the subject tracking has failed in the field of the captured image.
4. The image pick-up apparatus according to claim 3, wherein said setting unit sets the shorter holding time as the distance between the position at which the subject tracking has failed and a center of the field of the captured image becomes larger.
5. A tracking method for an image pick-up apparatus that generates captured images in series; said method comprising:
a tracking step of the image pick-up apparatus tracking a specified subject by extracting a region of the specified subject in the presently generated captured image having high correlation with a region of the specified subject in the precedently generated captured image;
a display step of the image pick-up apparatus displaying the captured image on a screen and display identification information showing that the specified subject is tracked; and
a setting step of the image pick-up apparatus setting a holding time in which the display of the identification information is held,
wherein said display step continues to display the identification information when the holding time does not lapse after said tracking step lost the subject, and eliminates a display of the identification information when the holding time has passed after the subject has been lost in said tracking step, and
wherein said setting step changes the holding time according to a sensitivity and at least one of control information for an image stabilizer and an output from an angle velocity sensor of the image pick-up apparatus.
US13/608,654 2008-08-29 2012-09-10 Image pick-up apparatus and tracking method therefor Abandoned US20130002885A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US13/608,654 US20130002885A1 (en) 2008-08-29 2012-09-10 Image pick-up apparatus and tracking method therefor

Applications Claiming Priority (6)

Application Number Priority Date Filing Date Title
JP2008221790 2008-08-29
JP2008-221790 2008-08-29
JP2009-190119 2009-08-19
JP2009190119A JP5409189B2 (en) 2008-08-29 2009-08-19 Imaging apparatus and control method thereof
US12/547,168 US8284257B2 (en) 2008-08-29 2009-08-25 Image pick-up apparatus and tracking method therefor
US13/608,654 US20130002885A1 (en) 2008-08-29 2012-09-10 Image pick-up apparatus and tracking method therefor

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US12/547,168 Continuation US8284257B2 (en) 2008-08-29 2009-08-25 Image pick-up apparatus and tracking method therefor

Publications (1)

Publication Number Publication Date
US20130002885A1 true US20130002885A1 (en) 2013-01-03

Family

ID=41724850

Family Applications (2)

Application Number Title Priority Date Filing Date
US12/547,168 Expired - Fee Related US8284257B2 (en) 2008-08-29 2009-08-25 Image pick-up apparatus and tracking method therefor
US13/608,654 Abandoned US20130002885A1 (en) 2008-08-29 2012-09-10 Image pick-up apparatus and tracking method therefor

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US12/547,168 Expired - Fee Related US8284257B2 (en) 2008-08-29 2009-08-25 Image pick-up apparatus and tracking method therefor

Country Status (3)

Country Link
US (2) US8284257B2 (en)
JP (1) JP5409189B2 (en)
CN (1) CN101662587B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11889180B2 (en) 2019-01-03 2024-01-30 Huawei Technologies Co., Ltd. Photographing method and electronic device

Families Citing this family (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2011148746A1 (en) * 2010-05-24 2011-12-01 富士フイルム株式会社 3d imaging device, face detection device, and operation control method therefor
CN201708684U (en) * 2010-06-07 2011-01-12 柳州五菱汽车有限责任公司 Heat dissipation system for driving controller of electric vehicle
JP5629533B2 (en) * 2010-08-30 2014-11-19 キヤノン株式会社 Imaging apparatus, control method therefor, program, and storage medium
JP2012109900A (en) * 2010-11-19 2012-06-07 Aof Imaging Technology Ltd Photographing device, photographing method and program
JP2013013050A (en) * 2011-05-27 2013-01-17 Ricoh Co Ltd Imaging apparatus and display method using imaging apparatus
JP5821457B2 (en) * 2011-09-20 2015-11-24 ソニー株式会社 Image processing apparatus, image processing apparatus control method, and program for causing computer to execute the method
US9275285B2 (en) * 2012-03-29 2016-03-01 The Nielsen Company (Us), Llc Methods and apparatus to count people in images
US8660307B2 (en) 2012-03-29 2014-02-25 The Nielsen Company (Us), Llc Methods and apparatus to count people in images
US9092675B2 (en) 2012-03-29 2015-07-28 The Nielsen Company (Us), Llc Methods and apparatus to count people in images
US8761442B2 (en) 2012-03-29 2014-06-24 The Nielsen Company (Us), Llc Methods and apparatus to count people in images
JP5873378B2 (en) * 2012-04-10 2016-03-01 キヤノン株式会社 Imaging apparatus and control method thereof
JP2014123815A (en) * 2012-12-20 2014-07-03 Sony Corp Image processing device, image processing method, and recording medium
JP6103526B2 (en) * 2013-03-15 2017-03-29 オリンパス株式会社 Imaging device, image display device, and display control method for image display device
JP2014207658A (en) * 2013-03-19 2014-10-30 パナソニック株式会社 Image processing device and imaging apparatus and image processing method
JP2014241505A (en) * 2013-06-11 2014-12-25 ソニー株式会社 Photographing apparatus, photographing method, template generation apparatus, template generation method, and program
JP6330283B2 (en) * 2013-09-30 2018-05-30 株式会社ニコン Subject tracking device, imaging device, and subject tracking program
JP6273766B2 (en) * 2013-10-24 2018-02-07 株式会社ニコン Image processing apparatus, imaging apparatus, and image processing program
JP6278688B2 (en) * 2013-12-20 2018-02-14 キヤノン株式会社 IMAGING DEVICE, IMAGING DEVICE CONTROL METHOD, AND PROGRAM
JP6442261B2 (en) * 2014-06-05 2018-12-19 キヤノン株式会社 Image pickup device, image pickup apparatus, and control method thereof
JP6085390B1 (en) * 2016-06-20 2017-02-22 ヤフー株式会社 Image processing apparatus, image processing method, and image processing program
JP6785164B2 (en) * 2017-01-27 2020-11-18 ヤフー株式会社 Image processing device, image processing method and image processing program
JP6607886B2 (en) * 2017-07-18 2019-11-20 シナノケンシ株式会社 Video analysis method
JP6630322B2 (en) * 2017-08-30 2020-01-15 シナノケンシ株式会社 Video analysis method
US10970425B2 (en) * 2017-12-26 2021-04-06 Seiko Epson Corporation Object detection and tracking
JP7146450B2 (en) * 2018-05-24 2022-10-04 日本信号株式会社 object detector
CN111050100A (en) * 2018-10-12 2020-04-21 黑快马股份有限公司 Automatic video recording and following shooting system and automatic video recording and following shooting method
CN112004019B (en) * 2019-05-27 2023-02-24 福建天晴数码有限公司 Tracking focusing method and computer readable storage medium
JP7366635B2 (en) 2019-08-07 2023-10-23 キヤノン株式会社 Imaging device, computer program and storage medium
CN111145311B (en) * 2019-12-30 2021-11-26 中国科学院自动化研究所 Multi-attribute editing method, system and device for high-resolution face image
CN112291480B (en) * 2020-12-03 2022-06-21 维沃移动通信有限公司 Tracking focusing method, tracking focusing device, electronic device and readable storage medium

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20010055063A1 (en) * 2000-05-26 2001-12-27 Honda Giken Kogyo Kabushiki Kaisha Position detection apparatus, position detection method and position detection program
US20070195165A1 (en) * 2005-05-20 2007-08-23 Olympus Medical Systems Corp. Image display apparatus
US20080316319A1 (en) * 2005-03-18 2008-12-25 Nikon Corporation Digital Camera
US20100040356A1 (en) * 2008-08-12 2010-02-18 Canon Kabushiki Kaisha Imaging apparatus and control method therefor
US7868917B2 (en) * 2005-11-18 2011-01-11 Fujifilm Corporation Imaging device with moving object prediction notification
US7945938B2 (en) * 2005-05-11 2011-05-17 Canon Kabushiki Kaisha Network camera system and control method therefore

Family Cites Families (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR940007163B1 (en) 1991-07-09 1994-08-06 삼성전자 주식회사 Auto-searching device of camcordor subject
FI97665C (en) * 1995-11-21 1997-01-27 Planmed Oy Procedures and apparatus for photographing an object
CN1178467C (en) * 1998-04-16 2004-12-01 三星电子株式会社 Method and device for automatically tracking a moving target
CN1254904A (en) * 1998-11-18 2000-05-31 株式会社新太吉 Method and equipment for picking-up/recognizing face
KR20010042310A (en) * 1999-02-01 2001-05-25 요트.게.아. 롤페즈 Descriptor for a video sequence and image retrieval system using said descriptor
US6754389B1 (en) * 1999-12-01 2004-06-22 Koninklijke Philips Electronics N.V. Program classification using object tracking
US20050128291A1 (en) * 2002-04-17 2005-06-16 Yoshishige Murakami Video surveillance system
US7466356B2 (en) * 2003-07-25 2008-12-16 Hewlett-Packard Development Company, L.P. Method and apparatus for setting a marker on an object and tracking the position of the object
US20060140445A1 (en) * 2004-03-22 2006-06-29 Cusack Francis J Jr Method and apparatus for capturing digital facial images optimally suited for manual and automated recognition
US7720257B2 (en) * 2005-06-16 2010-05-18 Honeywell International Inc. Object tracking system
JP4350725B2 (en) * 2005-08-05 2009-10-21 キヤノン株式会社 Image processing method, image processing apparatus, and program for causing computer to execute image processing method
JP4844073B2 (en) 2005-10-03 2011-12-21 株式会社ニコン Imaging device
JP4976160B2 (en) * 2007-02-22 2012-07-18 パナソニック株式会社 Imaging device
US20090009651A1 (en) * 2007-07-04 2009-01-08 Sanyo Electric Co., Ltd. Imaging Apparatus And Automatic Focus Control Method
CN101106700A (en) * 2007-08-01 2008-01-16 大连海事大学 Device and method for capturing intelligent target details in video surveillance system
JP4492697B2 (en) * 2007-12-28 2010-06-30 カシオ計算機株式会社 Imaging apparatus and program

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20010055063A1 (en) * 2000-05-26 2001-12-27 Honda Giken Kogyo Kabushiki Kaisha Position detection apparatus, position detection method and position detection program
US20080316319A1 (en) * 2005-03-18 2008-12-25 Nikon Corporation Digital Camera
US7945938B2 (en) * 2005-05-11 2011-05-17 Canon Kabushiki Kaisha Network camera system and control method therefore
US20070195165A1 (en) * 2005-05-20 2007-08-23 Olympus Medical Systems Corp. Image display apparatus
US7868917B2 (en) * 2005-11-18 2011-01-11 Fujifilm Corporation Imaging device with moving object prediction notification
US20100040356A1 (en) * 2008-08-12 2010-02-18 Canon Kabushiki Kaisha Imaging apparatus and control method therefor

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11889180B2 (en) 2019-01-03 2024-01-30 Huawei Technologies Co., Ltd. Photographing method and electronic device

Also Published As

Publication number Publication date
US20100053419A1 (en) 2010-03-04
CN101662587B (en) 2011-12-07
JP2010081587A (en) 2010-04-08
CN101662587A (en) 2010-03-03
US8284257B2 (en) 2012-10-09
JP5409189B2 (en) 2014-02-05

Similar Documents

Publication Publication Date Title
US8284257B2 (en) Image pick-up apparatus and tracking method therefor
US9317748B2 (en) Tracking apparatus
US8831282B2 (en) Imaging device including a face detector
CN103733607B (en) For detecting the apparatus and method of moving object
KR101519427B1 (en) Imaging device and its shutter drive mode selection method
EP2168005B1 (en) Focus control apparatus, image sensing apparatus, and control method therefor
US8929598B2 (en) Tracking apparatus, tracking method, and storage medium to store tracking program
US8411159B2 (en) Method of detecting specific object region and digital camera
US8830374B2 (en) Image capture device with first and second detecting sections for detecting features
JP5210843B2 (en) Imaging device
JP2009268086A (en) Imaging apparatus
US8878940B2 (en) Tracking apparatus for tracking target subject in input image
JP2015005799A (en) Subject detection device, control method of the same, imaging apparatus, control program of subject detection device, and storage medium
KR101728042B1 (en) Digital photographing apparatus and control method thereof
US9412008B2 (en) Tracking apparatus and tracking method
JP2013012941A (en) Tracking apparatus, tracking method, and tracking program
US20210256713A1 (en) Image processing apparatus and image processing method
JP4914045B2 (en) IMAGING DEVICE AND IMAGING DEVICE CONTROL METHOD
CN103581538B (en) Digital photographing apparatus and the method for controlling the digital photographing apparatus
JP2019201387A (en) Tracking controller
US9357124B2 (en) Focusing control device and controlling method of the same
JP5189913B2 (en) Image processing device
JP5118590B2 (en) Subject tracking method and imaging apparatus
US20130083963A1 (en) Electronic camera
JP4810440B2 (en) IMAGING DEVICE, ITS CONTROL METHOD, PROGRAM, AND STORAGE MEDIUM

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION