US20130300872A1 - Apparatus and method for displaying a blind spot - Google Patents
Apparatus and method for displaying a blind spot Download PDFInfo
- Publication number
- US20130300872A1 US20130300872A1 US13/996,346 US201113996346A US2013300872A1 US 20130300872 A1 US20130300872 A1 US 20130300872A1 US 201113996346 A US201113996346 A US 201113996346A US 2013300872 A1 US2013300872 A1 US 2013300872A1
- Authority
- US
- United States
- Prior art keywords
- vehicle
- image
- information
- display
- images
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 23
- 238000003384 imaging method Methods 0.000 claims abstract description 12
- 238000006073 displacement reaction Methods 0.000 claims description 35
- 238000010586 diagram Methods 0.000 description 14
- 241001300198 Caperonia palustris Species 0.000 description 3
- 235000000384 Veronica chamaedrys Nutrition 0.000 description 3
- 238000010276 construction Methods 0.000 description 3
- 238000006243 chemical reaction Methods 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 238000012545 processing Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 238000012805 post-processing Methods 0.000 description 1
- 238000007781 pre-processing Methods 0.000 description 1
- 230000011218 segmentation Effects 0.000 description 1
- 239000004575 stone Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G1/00—Traffic control systems for road vehicles
- G08G1/16—Anti-collision systems
- G08G1/167—Driving aids for lane monitoring, lane changing, e.g. blind spot detection
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R1/00—Optical viewing arrangements; Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
- B60R1/20—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
- B60R1/22—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle
- B60R1/23—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle with a predetermined field of view
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/70—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by an event-triggered choice to display a specific image among a selection of captured images
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/80—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement
- B60R2300/802—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement for monitoring and displaying vehicle exterior blind spot views
Definitions
- the present invention relates to an apparatus and method for displaying a blind spot in the surroundings of a vehicle and, more particularly, to an apparatus and method for providing a driver with information on blind spots in the front, rear, and sides of a vehicle that are hidden by a vehicle body and thus rarely seen within a field of view of the driver when the driver drives the vehicle by displaying one or more images of the surroundings of the vehicle provided by a camera, a GPS navigation system, and/or a street-view receiver.
- a vehicle runs on a road by using a motor, such as an engine installed in a vehicle body, as a power source.
- a vehicle is an apparatus for carrying persons or goods or performing various tasks. The driver of a vehicle safely drives the vehicle while watching a driving direction.
- a blind spot 20 is formed because a front portion near the vehicle 10 does not fall within a field of view of a driver due to the front portion in which an engine necessary to drive the vehicle 10 is installed. Furthermore, a blind spot 30 that partially limits a field of view of the rear is formed due to a trunk formed in the rear portion of the vehicle 10 , and blind spots 40 that partially limit fields of view of the sides are formed due to the doors of the vehicle.
- the blind spots 20 , 30 , and 40 refer to areas located at angles where things are not seed by an eye. If fields of view of the front and rear of a driver are limited by part of the vehicle body of the vehicle, there is no problem due to an obstacle in driving when the driver drives the vehicle on a road having extensive front views. However, the driver feels difficult to secure a field of view and to park the vehicle if obstacles are located in the blind spots because several obstacles are present in the surroundings of the vehicle 10 .
- a driver who is not accustomed to parking can hit an obstacle hidden in the front or rear of a vehicle when parking the vehicle. Even a driver who is accustomed to parking can park his vehicle while raising his head and watching the front and rear portions of a vehicle in order to avoid hidden obstacles, but there is a problem in that parking is very difficult because only a partial field of view is secured.
- An object of the present invention is to provide an apparatus and method for providing a driver with information on blind spots in the front, rear, and sides of a vehicle that are hidden by a vehicle body and thus rarely seen within a field of view of the driver when the driver drives the vehicle by displaying one or more images of the surroundings of the vehicle provided by a camera, a GPS navigation system, and/or a street-view receiver.
- the present invention provides an apparatus for displaying a blind spot, including an imaging unit or detecting surrounding images of a vehicle; a control unit for outputting at least one image captured by the imaging unit at a specific point of time or before the specific point of time as an image for display and displaying a vehicle icon indicative of a current position of the vehicle on the outputted image for display in a contour form or semi-transparently so that a blind spot covered by a vehicle body of the vehicle is displayed; and a screen display unit for displaying the image for display and the vehicle icon outputted from the control unit.
- the outputted image for display in accordance with the present invention is any one of an original image of the captured image and an image obtained by converting a point of time of the original image.
- the outputted image for display in accordance with the present invention is any one of an original image of the captured image, an image obtained by composing several images whose points of time have been converted, and a combination of some of several images.
- the blind spot covered by the vehicle body in accordance with the present invention includes a lower part of the vehicle.
- the imaging unit in accordance with the present invention is any one of a camera installed in the vehicle and configured to capture the surrounding images of the vehicle and provide the captured images, a GPS navigation system configured to detect information on a position of the vehicle and provide a surrounding image corresponding to the detected information, and an external information reception unit configured to receive the surrounding images of the vehicle from cameras installed in surrounding roads or buildings of the vehicle and provide the received surrounding images.
- control unit in accordance with the present invention calculates displacement information on a movement of the vehicle from a specific position using information on a position movement of the vehicle supplied outside the vehicle or driving control information detected from the vehicle and detects information on the current position of the vehicle based on the calculated displacement information.
- control unit in accordance with the present invention displays a position of the vehicle icon displayed on the image for display by incorporating information on a real-time position according to a movement of the vehicle into the position of the vehicle icon.
- the change of the position of the vehicle icon displayed on the image for display in accordance with the present invention includes changing only the position of the vehicle icon in a state in which the image for display is fixed, moving only the image for display in a state in which the position of the vehicle icon is fixed, or changing both the image for display and the position of the vehicle icon.
- the generated image for display in accordance with the present invention is obtained by merging image areas moved by a predetermined number of unit pixels using information on a position of the vehicle supplied from the vehicle or driving control information detected from the vehicle.
- the generated image for display in accordance with the present invention is obtained by comparing only images with each other and merging an image area including pixels having similarity and an image area having low similarity near the image area including pixels as a result of the comparison.
- information on the generated image in accordance with the present invention is obtained by extracting a reduced image search area according to information on a speed and steering angle of the vehicle based on information on a driving of the vehicle comprising the speed and movement direction of the vehicle supplied from the vehicle and merging an image area including pixels having similarity and an image area having low similarity near the image area including pixels using the extracted reduced image search area.
- the present invention provides a method for displaying a blind spot, including the steps of a) a control unit outputting at least one image, captured by an imaging unit at a specific point of time or prior to the specific point of time, as an image for display; b) the control unit calculating information on a displacement of a vehicle moved from a specific position based on information on a movement of a position of the vehicle supplied outside the vehicle or driving control information supplied from the vehicle and detecting a current position of the vehicle based on the calculated information on the movement displacement; and c) the control unit displaying the image for display and a vehicle icon indicative of the current position of the vehicle by displaying the vehicle icon based on the current position and the displacement information, detected at the step b), on the image for display outputted at the step a) in a contour form or semi-transparently so that a blind spot near the vehicle and a blind spot covered by a vehicle body are displayed.
- the image for display at the step a) in accordance with the present invention is any one of an original image of the captured image and an image obtained by converting a point of time of the original image.
- the image for display at the step a) in accordance with the present invention is any one of an original image of the captured image, an image obtained by composing several images whose points of time have been converted, and a combination of some of several images
- the image for display at the step a) in accordance with the present invention is an image captured by any one of a camera installed in the vehicle and configured to capture the surrounding images of the vehicle and provide the captured images, a GPS navigation system configured to detect information on a position of the vehicle and provide a surrounding image corresponding to the detected information, and an external information reception unit configured to receive the surrounding images of the vehicle from cameras installed in surrounding roads or buildings of the vehicle and provide the received surrounding images.
- the vehicle icon in accordance with the present invention is displayed so that a position of the vehicle icon displayed on the image for display is changed by incorporating information on a real-time position according to a movement of the vehicle into the position of the vehicle icon.
- the change of the position of the vehicle icon displayed on the image for display in accordance with the present invention includes changing only the position of the vehicle icon in a state in which the image for display is fixed, moving only the image for display in a state in which the position of the vehicle icon is fixed, or changing both the image for display and the position of the vehicle icon.
- the present invention is advantageous in that information on blind spots in the front, rear, and sides of a vehicle that are hidden by a vehicle body and thus rarely seen within a field of view of a driver when the driver drives the vehicle can be provided to the driver by displaying one or more images of the surroundings of the vehicle provided by the imaging unit.
- the present invention is advantageous in that a safe driving environment and a safe parking environment can be provided to a driver because blind spots rarely seen within a field of view of the driver can be checked.
- FIG. 1 is an exemplary diagram showing blind spots in the front and rear of a common vehicle.
- FIG. 2 is a block diagram showing an embodiment of an apparatus for displaying a blind spot in accordance with the present invention.
- FIG. 3 is an exemplary diagram showing an operating process of the apparatus for displaying a blind spot in accordance with the present invention.
- FIG. 4 is an exemplary diagram showing another operating process of the apparatus for displaying a blind spot in accordance with the present invention.
- FIG. 5 is an exemplary diagram in which a rear past image and a current image captured by the apparatus for displaying a blind spot in accordance with the present invention are merged.
- FIG. 6 is an exemplary diagram in which an embodiment of a vehicle icon is displayed on an image captured by the apparatus for displaying a blind spot in accordance with the present invention.
- FIG. 7 is an exemplary diagram in which another embodiment of a vehicle icon is displayed on an image captured by the apparatus for displaying a blind spot in accordance with the present invention.
- FIG. 8 is an exemplary diagram in which a vehicle icon is displayed on a road image captured by the apparatus for displaying a blind spot in accordance with the present invention.
- FIG. 9 is a flowchart illustrating a process of displaying a blind spot in accordance with the present invention.
- FIG. 10 is a block diagram showing another embodiment of an apparatus for displaying a blind spot in accordance with the present invention.
- 100 apparatus for displaying a blind spot 110: first camera 110′: first camera after a lapse of t seconds 120: second camera 120′: second camera after a lapse of t seconds 130: control unit 140: screen display unit 150: n th camera 200: vehicle 200′: vehicle after a lapse of t seconds 300: speed bump 400: vehicle icon 410: wheel 500: direction guide information 600: apparatus for displaying a blind spot 610: GPS navigation system 620: external information reception unit 630: control unit 640: screen display unit
- FIG. 2 is a block diagram showing an embodiment of an apparatus for displaying a blind spot in accordance with the present invention
- FIG. 3 is an exemplary diagram showing an operating process of the apparatus for displaying a blind spot in accordance with the present invention
- FIG. 4 is an exemplary diagram showing another operating process of the apparatus for displaying a blind spot in accordance with the present invention
- FIG. 5 is an exemplary diagram in which a rear past image and a current image captured by the apparatus for displaying a blind spot in accordance with the present invention are merged.
- the apparatus for displaying a blind spot 100 in accordance with the present invention includes a first camera 110 , a second camera 120 , and an n th camera 150 as imaging means for providing images of the surroundings of a vehicle, a control unit 130 , and a screen display unit 140 .
- the first camera 110 is an element installed in the vehicle 200 and configured to output a captured image of the rear of the vehicle while running.
- the first camera 110 is installed in a taillight installed in the room of the vehicle, a taillight, or a trunk door, etc.
- the second camera 120 is an element installed in the vehicle 200 and configured to output a captured image of the front of the vehicle while running.
- the second camera 120 is installed in a room mirror installed in the room of the vehicle, a head lamp, a radiator grill, etc.
- the n th camera 150 is an element installed in the vehicle 200 and configured to output captured images of the sides of the vehicle while running.
- the n th camera 150 is installed in the side-view mirrors of the vehicle, etc.
- CMOS type image sensor Any one of a CMOS type image sensor, a CCD type image sensor, and the third type image sensor can be used as the first, the second, and the n th cameras 110 , 120 , and 150 .
- a construction including one camera is illustrated, but a construction including a plurality of cameras for photographing the same driving direction is also possible.
- the control unit 130 displays one or more images captured by the first camera 110 , the second camera 120 , or the n th camera 150 at a specific point of time (hereinafter referred to as a ‘current point of time’) without change or displays images generated by performing any one of mergence, conversion, and combination using a previously stored image processing program through the screen display unit 140 .
- the control unit 130 detects at least one rear image A+B captured by the first camera 110 for photographing a specific area in the rear of the vehicle at a specific point of time and stores the detected rear images in storage means, such as a memory (not shown).
- the specific point of time may be a current point of time or may be any one point of time of a specific time interval between the past point of time and a current point of time.
- the at least one image A+B captured at the specific point of time may be one long image or may be one image A+B obtained by merging a first image A, captured at a specific past point of time, and a second image B captured by the first camera 110 ′ of the vehicle 200 ′ that has run after a lapse of a specific time (e.g., t seconds) at specific speed at the current point of time.
- a specific time e.g., t seconds
- control unit 130 detects information on a movement distance, movement speed, driving direction, and steering angle of the vehicle based on information on a movement distance calculated by comparing current images, captured by the ECU, TCU, speed sensor, steering angle sensor, wheel sensor, and camera of the vehicle, with images captured right before and detects information on the movement displacement of the vehicle based on driving control information including the detected information on the movement distance, movement speed, driving direction, and steering angle.
- control unit 130 calculates a movement position of the vehicle per hour from a specific position based on the detected information on the movement displacement of the vehicle, detects information on a current position of the vehicle based on the calculated movement position, and displays a vehicle icon 400 in a contour or semi-transparent form having a vehicle shape in order to indicate the current position of the vehicle 200 ′ at a specific position of the rear image A+B corresponding to the detected information on the current position of the vehicle.
- the control unit 130 displays the vehicle icon 400 in a contour or semi-transparent form having a vehicle shape on the detected rear image A+B instead of the real vehicle 200 , calculates a movement distance of the vehicle 200 ′ after a specific time from a specific position, for example, 0.5 second based on speed information detected by the speed sensor or TCU, etc. of the vehicle, and displays the contour or semi-transparent form having a vehicle shape at a position moved by a predetermined number of unit pixels or a specific distance, for example, in the area A on the rear image A+B based on the calculated movement distance of the vehicle.
- the specific time is illustrated as being 0.5 second, for convenience of description, but the present invention is not limited thereto.
- control unit 130 detects at least one front image C+D captured by the second camera 120 at a specific point of time and stores the detected front image C+D in storage means, such as the memory (not shown).
- the at least one front image C+D captured at the specific point of time may be one long image or may be an image obtained by merging a first front image C captured at a specific past point of time and a second front image D captured by the second camera 120 ′ of the vehicle 200 ′ that has moved at specific speed at a current point of time after a lapse of a specific time (e.g., t seconds).
- a specific time e.g., t seconds
- control unit 130 detects information on a movement displacement of the vehicle, including information on a speed, driving direction, and steering angle of the vehicle, from the vehicle, calculates a position of the vehicle that has moved per hour from a specific position based on the detected information on the movement displacement of the vehicle, detects a current position of the vehicle based on the calculated position, and displays the vehicle icon 400 in a contour or semi-transparent form having a vehicle shape in order to indicate the current position of the vehicle 200 ′ at a specific position of the front image C+D.
- the control unit 130 displays the contour or semi-transparent vehicle icon 400 on the detected front image C+D instead of the real vehicle 200 ′, calculates a movement distance of the vehicle 200 ′ after a specific time based on information on the speed of the vehicle, and displays the contour or semi-transparent vehicle icon at a position moved by a predetermined number of unit pixels, for example, in the area C on the front image C+D based on the calculated position of the vehicle.
- the information on the first rear image A that is, an image of a place through which the vehicle passed in the past, is separated into specific mergence areas A- 1 , A- 2 , and A- 3 on indication lines 1 , characteristic points for the separated mergence areas A- 1 , A- 2 , and A- 3 and information on the second rear image B, that is, a current image, are extracted by recognizing an object, pieces of information on the mergence areas A- 1 , A- 2 and A- 3 of the first rear image necessary for mergence are detected based on the extracted characteristic points, and the image A+B for display is generated by controlling the size of the images in the respective mergence areas A- 1 , A- 2 and A- 3 using a merger merge program or a stitching program.
- the image for display may be one image captured at a specific point of time t+1, may be an image obtained by merging an image captured at the specific point of time t+1 and an image captured at a previous point of time t, or may be an image obtained by merging the image captured at the specific point of time t+1, the image captured at the previous point of time t, and an image captured at a point of time t ⁇ 1 prior to the previous point of time t in order to display a blind spot not seen by a camera using the past image.
- image areas moved by a predetermined number of unit pixels based on information on the speed and the steering angle of a vehicle may be merged based on information on the driving of the vehicle including the speed and movement direction of the vehicle.
- image areas whose pixels are moved by a predetermined number of unit pixels (e.g., 20 pixels) in any one of up and down directions in the driving direction and then moved by a predetermined number of unit pixels (e.g., 10 pixels) left or right in the steering direction may be merged.
- the images can be merged more precisely by reducing the search area of an image according to information on the speed and steering angle of a vehicle and extracting image areas including pixels having similarity from the image having the reduced search area based on information on the driving of the vehicle including information on the speed and movement direction of the vehicle supplied from the vehicle.
- FIG. 6 shows that blind spots covered by the vehicle body of a vehicle are displayed along with a vehicle icon on an image detected by the apparatus for displaying a blind spot.
- FIG. 6 a shows a first rear image A captured by the camera prior to a specific point of time (past) when a vehicle runs backward
- FIG. 6 b shows a second rear image B captured by the camera at a specific point of time
- FIG. 6 c shows an image A+B for display obtained by merging the first rear image A and the second rear image B.
- the image A+B for display may be generated by merging the current image and the past image, but may be displayed using one long image.
- control unit 130 can display the vehicle icon 400 using a contour in such a way as to overlap with the second rear image A so that a driver can check a current position of the vehicle on the generated image A+B for display and check a precise position of the vehicle that runs backward.
- control unit 130 can also display the position of a wheel 410 along with the vehicle icon 400 so that a driver can check whether the wheel 410 is deviated from a parking guide line or the vehicle runs backward in a precise direction, for example.
- FIG. 7 shows that a blind spot covered by the vehicle body of a vehicle is displayed along with a vehicle icon on an image captured by the apparatus for displaying a blind spot.
- FIG. 7 a shows a second front image D of the current surroundings of the vehicle that runs forward captured by the second camera 120
- FIG. 7 b shows an image C+D for display obtained by merging a first front image C captured in the past prior to a specific time and the second front image D or an image for display generated by capturing one front image C+D including a speed bump 300 in a specific point of time using the second camera 120 .
- the driver can recognize that the vehicle passes through the speed bump 300 seen prior to the t seconds, but the real speed bump 300 is covered by the vehicle body and located in a blind spot that is rarely seen within a field of view of the driver.
- the control unit 130 calculates information on a movement displacement of the vehicle based on information on the movement distance, speed, and steering angle of the vehicle using a recognize distance and position calculation program, detects information on the movement distance of the vehicle from a specific point of time and on the position of the vehicle based on the calculated information, and displays a virtual vehicle icon 400 , indicating a current position of the vehicle, in a corresponding area on the front image C+D (e.g., the area C) based on the detected information on the distance and position in a contour or semi-transparent form having a vehicle shape. Accordingly, the driver can recognize a blind spot under the vehicle that is covered by the vehicle body.
- a recognize distance and position calculation program detects information on the movement distance of the vehicle from a specific point of time and on the position of the vehicle based on the calculated information, and displays a virtual vehicle icon 400 , indicating a current position of the vehicle, in a corresponding area on the front image C+D (e.g., the area C) based on the detected
- control unit 130 displays a virtual vehicle icon 400 in a place where the vehicle is now placed in a contour or semi-transparent form having a vehicle shape. Accordingly, the driver can recognize that the vehicle is now passing through a section in which the speed bump 300 is installed and can check a blind spot area C- 1 and a blind spot area C- 2 on the right and left sides of the vehicle icon 400 .
- control unit 130 displays information on the virtual vehicle icon 400 on the front image C+D in a contour or semi-transparent form having a vehicle shape so that the driver can recognize information on an obstacle and direction guide located in a road in the surroundings of the vehicle.
- FIG. 8 a shows information on a second front image D, that is, a current image
- FIG. 8 b shows information on a first front image C, that is, a past image prior to a specific time (e.g., 0.5 second)
- FIG. 8 c shows information on a front matching image C+D in which FIGS. 8 a and 8 b are matched.
- the control unit 130 displays a virtual vehicle icon 400 , displayed on the front image C+D only in a contour or semi-transparent form having a vehicle shape so that the driver can recognize the direction guide information 500 .
- control unit 130 can check whether or not a fixed object or a moving object is included in a current image and a past image prior to a specific point of time and display information on the distance up to the object based on a result of the check.
- the control unit 130 detects information on the pixels of objects having similar characteristics (e.g., the brightness of pixels and whether or not an edge exists) in a current image and a past image by processing an image captured at a current point of time and an image at a past point of time prior to a specific time and detects information (pixel information) on a specific object including the displacement.
- similar characteristics e.g., the brightness of pixels and whether or not an edge exists
- the control unit 130 can recognize an object by detecting similarity through a movement displacement and an angular displacement of pixels defined as vector by comparing two interested images through image processing, can recognize an object by using edges, such as the boundary line or contour of an object or lines or curves approximate to the contour of an object, can recognize an object by using edge points present within an image, unchanged points, or characteristic points that are not changed in relation to rotational conversion or a change in the size, or can recognize an object by extracting the characteristic points of the object included in an image.
- edges such as the boundary line or contour of an object or lines or curves approximate to the contour of an object
- control unit 130 detects a movement displacement (i.e., vector) between a current position of an object detected in a current image and a past position of the object detected in a past image, detects information on the driving of a vehicle including information on a steering angle, distance and speed of the vehicle supplied from the ECU, TCU, wheel sensor and/or steering angle sensor of the vehicle, and calculates a movement distance corresponding to the movement displacements of the two objects.
- a movement displacement i.e., vector
- control unit 130 displays a current position of an object and a past position of the object on a current image and sets a difference value between the two objects as a movement distance between the two objects using a distance that the vehicle has moved.
- control unit 130 detects a predetermined reference point on the current image and a displacement up to the current object and the past object, uses triangular information formed by the reference point and the current object and the past object, that is, a movement displacement between the current object and the past object, as a reference side, uses displacements formed by the reference point and the current object and by the reference point and the past object, respectively, as two other sides, calculates angles formed by the sides, and calculates the distance from the reference point to the current object based on the calculated angles.
- control unit 130 can estimate a movement displacement of the object, calculate an expected position and distance up to the object placed on a specific area of an image after a lapse of a specific time based on the estimated movement displacement, the speed of a vehicle, and the relative speed of the current object detected based on the estimated movement displacement, and provide the calculated position and distance.
- the control unit 130 can extract the distance up to an object through a stereo vision.
- the control unit 130 can estimate a movement displacement of the object, calculate an expected position and distance up to the object placed on a specific area of an image after a lapse of a specific time based on the estimated movement displacement, the speed of the vehicle, and the relative speed of the current object detected based on the estimated movement displacement, and provide the calculated position and distance.
- control unit 130 compares the calculated distance with predetermined reference distance information for determining whether or not there is a collision. If, as a result of the comparison, a collision is expected, the control unit 130 outputs a warning signal so that alarm or a warning screen is displayed.
- the stereo vision program When left and right images are received through stereo cameras installed in the head lamp of a vehicle, the taillight of the vehicle, or side-view mirrors on the left and right sides, the stereo vision program performs a pre-processing process on the stereo image input for matching brightness and epipolar lines for the left and right images, performs stereo matching for generating disparity between the pixels of areas overlapped with each other in the left and right images that are determined as the same object, removes noise components from the results of the disparity, obtains a stereo image by performing post-processing processes, such as projection, segmentation, and filtering, on information on the distance of the stereo image, and detects pieces of information, such as the distance, shape, and the direction of the object, through the stereo image.
- post-processing processes such as projection, segmentation, and filtering
- the stereo vision program is one embodiment, and a variety of known stereo vision programs may be used.
- control unit 130 can output the original image, captured by a camera, as an image for display without change, but may convert the captured original image into an image having a Bird Eye View point of time using a time point conversion program and output the converted image as an image for display.
- control unit 130 may output the original image, captured by the camera, as an image for display or may merge several images whose points of time have been converted or stitch the several images using an image processing program and output the merged or stitched image as an image for display.
- control unit 130 may combine the original image captured by the camera and some of the images whose points of time have been converted or the merged images using an image processing program and output the combined image as an image for display.
- the screen display unit 140 outputs an image outputted from the control unit 130 and the vehicle icon 400 in a contour or semi-transparent form having a vehicle shape together and includes output means, such as an LCD.
- the vehicle icon 400 has a contour or semi-transparent form having a vehicle shape and virtually displays a current position of a vehicle in an image for display.
- the vehicle icon 400 may display the position of the wheel (not shown) of a vehicle along with a contour or semi-transparent form having a vehicle shape. Accordingly, when a vehicle runs, an obstacle under the vehicle covered by the vehicle body of the vehicle and driving direction signs and characters for road guidance indicated in roads can be precisely delivered to a driver.
- only the position of the vehicle icon 400 indicated in an image for display may be changed in the state in which the image for display is fixed, only the image for display may be moved in the state in which the position of the vehicle icon 400 is fixed, or both the image for display and the position of the vehicle icon 400 may be changed.
- FIG. 9 is a flowchart illustrating a process of displaying a blind spot in accordance with the present invention. A method for displaying a blind spot in accordance with the present invention is described with reference to FIGS. 2 and 4 to 9 .
- the control unit 130 detects an image C or C+D for display captured by the second camera 120 at a specific point of time at step S 100 and temporarily stores the detected image.
- control unit 130 can store and output the image for display captured at step S 100 as the original image, but may convert the captured original image into an image having a Bird Eye View point of time using a time point conversion program and output the converted image as the image for display.
- the image for display captured at step S 100 may be the original image captured by the camera, an image obtained by merging or stitching several images whose points of time have been converted by the control unit 130 using an image processing program, or an image obtained by combining the original image and some of the images whose points of time have been converted and the merged images by the control unit 130 using an image processing program.
- the control unit 130 detects information on the movement displacement of the vehicle, including information on the movement distance, speed, driving direction, and steering angle of the vehicle, based on driving control information calculated by comparing a current image captured by the ECU, TCU, speed sensor, steering angle sensor, wheel sensor, and camera of the vehicle supplied from the vehicle with an image captured right before the current image and calculates a position according to the movement distance and steering direction of the vehicle per hour based on the detected movement displacement information at step S 110 .
- the control unit 130 matches the calculated movement distance and position with the image C+D for display detected at step S 100 according to a predetermined number of unit pixels and displays a virtual vehicle icon 400 indicative of a current position of the vehicle in a contour or semi-transparent form having a vehicle shape at step S 120 so that the virtual vehicle icon 400 is outputted through the screen display unit 140 .
- the virtual vehicle icon 400 outputted through the screen display unit 140 at step S 120 is displayed on the front image C+D in a contour or semi-transparent form having a vehicle shape so that obstacles (e.g., a speed bump, a boundary stone, and a median strip) within blind spots or direction guide information 500 , together with information on the current position of the vehicle, can be easily recognized by a driver.
- obstacles e.g., a speed bump, a boundary stone, and a median strip
- a safe driving environment and a safe parking environment can be provided to a driver because the driver can check blind spots covered by the vehicle body of a vehicle and thus rarely seen within a field of view of the driver and blind spots placed in the surroundings of the vehicle and not seen in a current image captured by a camera can be checked.
- the position of a virtual vehicle icon indicating a current position of a vehicle on a displayed image is changed and displayed on an image for display as the vehicle moves. It is evident that the position of the virtual vehicle icon may be changed and displayed on an image for display in which a plurality of images is continuously arranged as the vehicle moves.
- FIG. 10 is a block diagram showing another embodiment of an apparatus for displaying a blind spot in accordance with the present invention.
- the second embodiment relates to a construction for detecting images of the surroundings of a vehicle from devices other than cameras installed in the vehicle and displaying a blind spot based on the detected images.
- an apparatus for displaying a blind spot 600 in accordance with the second embodiment includes a GPS navigation system 610 and an external information reception unit 620 as imaging means for providing images of the surroundings of a vehicle, a control unit 630 , and a screen display unit 640 .
- the GPS navigation system 610 is installed in the vehicle.
- the GPS navigation system 610 receives information on the position of the vehicle from a GPS receiver, detects information on a current position of the vehicle based on the received information, reads a map image or real image of a corresponding position previously stored in a memory (not shown) based on the detected information on the position of the vehicle, and provides the read map image or real image.
- the external information reception unit 620 is connected to cameras (not shown) installed in roads and buildings in the surroundings of the vehicle over a radio network.
- the external information reception unit 620 receives images captured by the cameras and provides the images to the control unit 630 .
- a current intelligent road network system is equipped with fixed cameras for photographing some road sections or some areas of buildings, but the cameras are assigned unique IPs. Accordingly, a user can access the cameras using radio communication and easily receive and check images captured by the cameras.
- the external information reception unit 620 accesses, for example, IP cameras installed in roads or buildings in the surroundings of roads through radio communication, receives images provided by the IP cameras and information on positions where the IP cameras are installed, stores the received images and information in the memory (not shown), and provides the images received from the IP cameras to the control unit 630 . Furthermore, the external information reception unit 620 detects a current position of the vehicle from the GPS navigation system 610 or a GPS receiver (not shown) along with information on positions where the IP cameras are installed supplied from the IP cameras and sends the detected current position and the information on the positions to the control unit 630 .
- the GPS navigation system 610 or a GPS receiver not shown
- the control unit 630 detects a map image of a corresponding position or an image of a position where the vehicle is placed, which are supplied by the GPS navigation system 610 or the external information reception unit 620 at a specific point of time and generates an image for display using the detected one or more images without change or displays and outputs an image for display that has been reconfigured by merging the detected one or more images with an image captured at a specific point of time or images captured right before the specific point of time using an image processing program, converting points of time of the detected images using a time point conversion program, or combining some of the detected images.
- control unit 630 can output the original image captured by, for example, the IP camera as an image for display without change, but may convert the captured original image into an image having a Bird Eye View point of time using a time point conversion program and output the converted image as an image for display.
- control unit 630 may output the original image captured by the IP camera as an image for display or may merge or stitch past several images whose points of time have been converted using an image processing program and output the merged or stitched image as an image for display.
- the control unit 630 may combine the original image captured by the IP camera and some of the images whose points of time have been converted or the merged images using an image processing program and output the combined image as an image for display.
- control unit 630 compares driving control information, detected by the ECU, TCU, speed sensor, steering angle sensor, and wheel sensor of the vehicle, with driving control information received prior to a specific point of time and detects information on the movement displacement of the vehicle, such as the movement distance, speed, driving direction, and steering angle of the vehicle, based on a result of the comparison.
- the control unit 630 can compare a map image corresponding to the position information with information on the position of the vehicle and an image received prior to a specific point of time and detect information on a movement displacement of the vehicle based on a result of the comparison.
- control unit 630 can compare information on the position of a camera detected by the street-view receiver 620 and information on a movement distance using surrounding images and the GPS navigation system 610 and detect information on a movement displacement of the vehicle based on a result of the comparison.
- control unit 630 detects information on a current position of the vehicle that has moved from a specific position based on the detected information on the movement displacement of the vehicle and displays a contour or semi-transparent vehicle icon indicative of the detected current position of the vehicle on the generated image for display.
- the screen display unit 640 outputs the virtual vehicle icon having a contour or semi-transparent vehicle form along with the image outputted from the control unit 630 .
- the screen display unit 640 includes output means, such as an LCD.
- a safe driving environment and a safe parking environment can be provided to a driver.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Mechanical Engineering (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Closed-Circuit Television Systems (AREA)
- Traffic Control Systems (AREA)
- Image Processing (AREA)
- Fittings On The Vehicle Exterior For Carrying Loads, And Devices For Holding Or Mounting Articles (AREA)
- Position Fixing By Use Of Radio Waves (AREA)
Abstract
Description
- The present invention relates to an apparatus and method for displaying a blind spot in the surroundings of a vehicle and, more particularly, to an apparatus and method for providing a driver with information on blind spots in the front, rear, and sides of a vehicle that are hidden by a vehicle body and thus rarely seen within a field of view of the driver when the driver drives the vehicle by displaying one or more images of the surroundings of the vehicle provided by a camera, a GPS navigation system, and/or a street-view receiver.
- In general, a vehicle runs on a road by using a motor, such as an engine installed in a vehicle body, as a power source. A vehicle is an apparatus for carrying persons or goods or performing various tasks. The driver of a vehicle safely drives the vehicle while watching a driving direction.
- In a
common vehicle 10, however, as shown inFIG. 1 , ablind spot 20 is formed because a front portion near thevehicle 10 does not fall within a field of view of a driver due to the front portion in which an engine necessary to drive thevehicle 10 is installed. Furthermore, ablind spot 30 that partially limits a field of view of the rear is formed due to a trunk formed in the rear portion of thevehicle 10, andblind spots 40 that partially limit fields of view of the sides are formed due to the doors of the vehicle. - The
20, 30, and 40 refer to areas located at angles where things are not seed by an eye. If fields of view of the front and rear of a driver are limited by part of the vehicle body of the vehicle, there is no problem due to an obstacle in driving when the driver drives the vehicle on a road having extensive front views. However, the driver feels difficult to secure a field of view and to park the vehicle if obstacles are located in the blind spots because several obstacles are present in the surroundings of theblind spots vehicle 10. - Furthermore, a driver who is not accustomed to parking can hit an obstacle hidden in the front or rear of a vehicle when parking the vehicle. Even a driver who is accustomed to parking can park his vehicle while raising his head and watching the front and rear portions of a vehicle in order to avoid hidden obstacles, but there is a problem in that parking is very difficult because only a partial field of view is secured.
- Furthermore, although a driver already recognizes that an obstacle is present in the front or rear of a vehicle, there is a problem in that the driver can hit the obstacle because he does not temporarily remember the obstacle when the driver focuses on driving or parking.
- An object of the present invention is to provide an apparatus and method for providing a driver with information on blind spots in the front, rear, and sides of a vehicle that are hidden by a vehicle body and thus rarely seen within a field of view of the driver when the driver drives the vehicle by displaying one or more images of the surroundings of the vehicle provided by a camera, a GPS navigation system, and/or a street-view receiver.
- To achieve the above object, the present invention provides an apparatus for displaying a blind spot, including an imaging unit or detecting surrounding images of a vehicle; a control unit for outputting at least one image captured by the imaging unit at a specific point of time or before the specific point of time as an image for display and displaying a vehicle icon indicative of a current position of the vehicle on the outputted image for display in a contour form or semi-transparently so that a blind spot covered by a vehicle body of the vehicle is displayed; and a screen display unit for displaying the image for display and the vehicle icon outputted from the control unit.
- Furthermore, the outputted image for display in accordance with the present invention is any one of an original image of the captured image and an image obtained by converting a point of time of the original image.
- Furthermore, the outputted image for display in accordance with the present invention is any one of an original image of the captured image, an image obtained by composing several images whose points of time have been converted, and a combination of some of several images.
- Furthermore, the blind spot covered by the vehicle body in accordance with the present invention includes a lower part of the vehicle.
- Furthermore, the imaging unit in accordance with the present invention is any one of a camera installed in the vehicle and configured to capture the surrounding images of the vehicle and provide the captured images, a GPS navigation system configured to detect information on a position of the vehicle and provide a surrounding image corresponding to the detected information, and an external information reception unit configured to receive the surrounding images of the vehicle from cameras installed in surrounding roads or buildings of the vehicle and provide the received surrounding images.
- Furthermore, the control unit in accordance with the present invention calculates displacement information on a movement of the vehicle from a specific position using information on a position movement of the vehicle supplied outside the vehicle or driving control information detected from the vehicle and detects information on the current position of the vehicle based on the calculated displacement information.
- Furthermore, the control unit in accordance with the present invention displays a position of the vehicle icon displayed on the image for display by incorporating information on a real-time position according to a movement of the vehicle into the position of the vehicle icon.
- Furthermore, the change of the position of the vehicle icon displayed on the image for display in accordance with the present invention includes changing only the position of the vehicle icon in a state in which the image for display is fixed, moving only the image for display in a state in which the position of the vehicle icon is fixed, or changing both the image for display and the position of the vehicle icon.
- Furthermore, the generated image for display in accordance with the present invention is obtained by merging image areas moved by a predetermined number of unit pixels using information on a position of the vehicle supplied from the vehicle or driving control information detected from the vehicle.
- Furthermore, the generated image for display in accordance with the present invention is obtained by comparing only images with each other and merging an image area including pixels having similarity and an image area having low similarity near the image area including pixels as a result of the comparison.
- Furthermore, information on the generated image in accordance with the present invention is obtained by extracting a reduced image search area according to information on a speed and steering angle of the vehicle based on information on a driving of the vehicle comprising the speed and movement direction of the vehicle supplied from the vehicle and merging an image area including pixels having similarity and an image area having low similarity near the image area including pixels using the extracted reduced image search area.
- Furthermore, the present invention provides a method for displaying a blind spot, including the steps of a) a control unit outputting at least one image, captured by an imaging unit at a specific point of time or prior to the specific point of time, as an image for display; b) the control unit calculating information on a displacement of a vehicle moved from a specific position based on information on a movement of a position of the vehicle supplied outside the vehicle or driving control information supplied from the vehicle and detecting a current position of the vehicle based on the calculated information on the movement displacement; and c) the control unit displaying the image for display and a vehicle icon indicative of the current position of the vehicle by displaying the vehicle icon based on the current position and the displacement information, detected at the step b), on the image for display outputted at the step a) in a contour form or semi-transparently so that a blind spot near the vehicle and a blind spot covered by a vehicle body are displayed.
- Furthermore, the image for display at the step a) in accordance with the present invention is any one of an original image of the captured image and an image obtained by converting a point of time of the original image.
- Furthermore, the image for display at the step a) in accordance with the present invention is any one of an original image of the captured image, an image obtained by composing several images whose points of time have been converted, and a combination of some of several images
- Furthermore, the image for display at the step a) in accordance with the present invention is an image captured by any one of a camera installed in the vehicle and configured to capture the surrounding images of the vehicle and provide the captured images, a GPS navigation system configured to detect information on a position of the vehicle and provide a surrounding image corresponding to the detected information, and an external information reception unit configured to receive the surrounding images of the vehicle from cameras installed in surrounding roads or buildings of the vehicle and provide the received surrounding images.
- Furthermore, the vehicle icon in accordance with the present invention is displayed so that a position of the vehicle icon displayed on the image for display is changed by incorporating information on a real-time position according to a movement of the vehicle into the position of the vehicle icon.
- Furthermore, the change of the position of the vehicle icon displayed on the image for display in accordance with the present invention includes changing only the position of the vehicle icon in a state in which the image for display is fixed, moving only the image for display in a state in which the position of the vehicle icon is fixed, or changing both the image for display and the position of the vehicle icon.
- The present invention is advantageous in that information on blind spots in the front, rear, and sides of a vehicle that are hidden by a vehicle body and thus rarely seen within a field of view of a driver when the driver drives the vehicle can be provided to the driver by displaying one or more images of the surroundings of the vehicle provided by the imaging unit.
- Furthermore, the present invention is advantageous in that a safe driving environment and a safe parking environment can be provided to a driver because blind spots rarely seen within a field of view of the driver can be checked.
-
FIG. 1 is an exemplary diagram showing blind spots in the front and rear of a common vehicle. -
FIG. 2 is a block diagram showing an embodiment of an apparatus for displaying a blind spot in accordance with the present invention. -
FIG. 3 is an exemplary diagram showing an operating process of the apparatus for displaying a blind spot in accordance with the present invention. -
FIG. 4 is an exemplary diagram showing another operating process of the apparatus for displaying a blind spot in accordance with the present invention. -
FIG. 5 is an exemplary diagram in which a rear past image and a current image captured by the apparatus for displaying a blind spot in accordance with the present invention are merged. -
FIG. 6 is an exemplary diagram in which an embodiment of a vehicle icon is displayed on an image captured by the apparatus for displaying a blind spot in accordance with the present invention. -
FIG. 7 is an exemplary diagram in which another embodiment of a vehicle icon is displayed on an image captured by the apparatus for displaying a blind spot in accordance with the present invention. -
FIG. 8 is an exemplary diagram in which a vehicle icon is displayed on a road image captured by the apparatus for displaying a blind spot in accordance with the present invention. -
FIG. 9 is a flowchart illustrating a process of displaying a blind spot in accordance with the present invention. -
FIG. 10 is a block diagram showing another embodiment of an apparatus for displaying a blind spot in accordance with the present invention. -
-
100: apparatus for displaying a blind spot 110: first camera 110′: first camera after a lapse of t seconds 120: second camera 120′: second camera after a lapse of t seconds 130: control unit 140: screen display unit 150: nth camera 200: vehicle 200′: vehicle after a lapse of t seconds 300: speed bump 400: vehicle icon 410: wheel 500: direction guide information 600: apparatus for displaying a blind spot 610: GPS navigation system 620: external information reception unit 630: control unit 640: screen display unit - Preferred embodiments of an apparatus and method for displaying a blind spot in accordance with the present invention are described in detail with reference to the accompanying drawings.
-
FIG. 2 is a block diagram showing an embodiment of an apparatus for displaying a blind spot in accordance with the present invention,FIG. 3 is an exemplary diagram showing an operating process of the apparatus for displaying a blind spot in accordance with the present invention,FIG. 4 is an exemplary diagram showing another operating process of the apparatus for displaying a blind spot in accordance with the present invention, andFIG. 5 is an exemplary diagram in which a rear past image and a current image captured by the apparatus for displaying a blind spot in accordance with the present invention are merged. - As shown in
FIG. 2 toFIG. 5 , the apparatus for displaying ablind spot 100 in accordance with the present invention includes afirst camera 110, asecond camera 120, and an nth camera 150 as imaging means for providing images of the surroundings of a vehicle, acontrol unit 130, and ascreen display unit 140. - The
first camera 110 is an element installed in thevehicle 200 and configured to output a captured image of the rear of the vehicle while running. Thefirst camera 110 is installed in a taillight installed in the room of the vehicle, a taillight, or a trunk door, etc. - The
second camera 120 is an element installed in thevehicle 200 and configured to output a captured image of the front of the vehicle while running. Thesecond camera 120 is installed in a room mirror installed in the room of the vehicle, a head lamp, a radiator grill, etc. - The nth camera 150 is an element installed in the
vehicle 200 and configured to output captured images of the sides of the vehicle while running. The nth camera 150 is installed in the side-view mirrors of the vehicle, etc. - Any one of a CMOS type image sensor, a CCD type image sensor, and the third type image sensor can be used as the first, the second, and the nth cameras 110, 120, and 150.
- Furthermore, in the present embodiment, in order to capture a driving direction image of the front or rear of the vehicle, a construction including one camera is illustrated, but a construction including a plurality of cameras for photographing the same driving direction is also possible.
- The
control unit 130 displays one or more images captured by thefirst camera 110, thesecond camera 120, or the nth camera 150 at a specific point of time (hereinafter referred to as a ‘current point of time’) without change or displays images generated by performing any one of mergence, conversion, and combination using a previously stored image processing program through thescreen display unit 140. - When the
vehicle 200 runs backward from a specific departure position as shown inFIG. 3 , thecontrol unit 130 detects at least one rear image A+B captured by thefirst camera 110 for photographing a specific area in the rear of the vehicle at a specific point of time and stores the detected rear images in storage means, such as a memory (not shown). - Here, the specific point of time may be a current point of time or may be any one point of time of a specific time interval between the past point of time and a current point of time.
- Furthermore, the at least one image A+B captured at the specific point of time may be one long image or may be one image A+B obtained by merging a first image A, captured at a specific past point of time, and a second image B captured by the
first camera 110′ of thevehicle 200′ that has run after a lapse of a specific time (e.g., t seconds) at specific speed at the current point of time. - Furthermore, the
control unit 130 detects information on a movement distance, movement speed, driving direction, and steering angle of the vehicle based on information on a movement distance calculated by comparing current images, captured by the ECU, TCU, speed sensor, steering angle sensor, wheel sensor, and camera of the vehicle, with images captured right before and detects information on the movement displacement of the vehicle based on driving control information including the detected information on the movement distance, movement speed, driving direction, and steering angle. - Furthermore, the
control unit 130 calculates a movement position of the vehicle per hour from a specific position based on the detected information on the movement displacement of the vehicle, detects information on a current position of the vehicle based on the calculated movement position, and displays avehicle icon 400 in a contour or semi-transparent form having a vehicle shape in order to indicate the current position of thevehicle 200′ at a specific position of the rear image A+B corresponding to the detected information on the current position of the vehicle. - That is, after detecting the rear image A+B, the
control unit 130 displays thevehicle icon 400 in a contour or semi-transparent form having a vehicle shape on the detected rear image A+B instead of thereal vehicle 200, calculates a movement distance of thevehicle 200′ after a specific time from a specific position, for example, 0.5 second based on speed information detected by the speed sensor or TCU, etc. of the vehicle, and displays the contour or semi-transparent form having a vehicle shape at a position moved by a predetermined number of unit pixels or a specific distance, for example, in the area A on the rear image A+B based on the calculated movement distance of the vehicle. The specific time is illustrated as being 0.5 second, for convenience of description, but the present invention is not limited thereto. - Furthermore, when the vehicle runs forward as shown in
FIG. 4 , thecontrol unit 130 detects at least one front image C+D captured by thesecond camera 120 at a specific point of time and stores the detected front image C+D in storage means, such as the memory (not shown). - The at least one front image C+D captured at the specific point of time may be one long image or may be an image obtained by merging a first front image C captured at a specific past point of time and a second front image D captured by the
second camera 120′ of thevehicle 200′ that has moved at specific speed at a current point of time after a lapse of a specific time (e.g., t seconds). - Furthermore, the
control unit 130 detects information on a movement displacement of the vehicle, including information on a speed, driving direction, and steering angle of the vehicle, from the vehicle, calculates a position of the vehicle that has moved per hour from a specific position based on the detected information on the movement displacement of the vehicle, detects a current position of the vehicle based on the calculated position, and displays thevehicle icon 400 in a contour or semi-transparent form having a vehicle shape in order to indicate the current position of thevehicle 200′ at a specific position of the front image C+D. - That is, after detecting the front image C+D, the
control unit 130 displays the contour orsemi-transparent vehicle icon 400 on the detected front image C+D instead of thereal vehicle 200′, calculates a movement distance of thevehicle 200′ after a specific time based on information on the speed of the vehicle, and displays the contour or semi-transparent vehicle icon at a position moved by a predetermined number of unit pixels, for example, in the area C on the front image C+D based on the calculated position of the vehicle. - Meanwhile, in the process of merging a past image and a current image, as shown in
FIGS. 5 a and 5 b, for example, when information on the first rear image A and the second rear image B is detected, the information on the first rear image A, that is, information on a past image, is merged with information on the second rear image B, that is, information on a current image, thereby generating the image A+B for display, such as that shown inFIG. 5 c. - Here, the information on the first rear image A, that is, an image of a place through which the vehicle passed in the past, is separated into specific mergence areas A-1, A-2, and A-3 on
indication lines 1, characteristic points for the separated mergence areas A-1, A-2, and A-3 and information on the second rear image B, that is, a current image, are extracted by recognizing an object, pieces of information on the mergence areas A-1, A-2 and A-3 of the first rear image necessary for mergence are detected based on the extracted characteristic points, and the image A+B for display is generated by controlling the size of the images in the respective mergence areas A-1, A-2 and A-3 using a merger merge program or a stitching program. - That is, the image for display may be one image captured at a specific point of
time t+ 1, may be an image obtained by merging an image captured at the specific point of time t+1 and an image captured at a previous point of time t, or may be an image obtained by merging the image captured at the specific point oftime t+ 1, the image captured at the previous point of time t, and an image captured at a point of time t−1 prior to the previous point of time t in order to display a blind spot not seen by a camera using the past image. - In the mergence of images in the present embodiment, two images are compared with each other and image areas including pixels having similarity as a result of the comparison are merged. In contrast, image areas moved by a predetermined number of unit pixels based on information on the speed and the steering angle of a vehicle may be merged based on information on the driving of the vehicle including the speed and movement direction of the vehicle.
- That is, if speed of a vehicle is 10 Km and a steering angle of the vehicle is 10 degrees, image areas whose pixels are moved by a predetermined number of unit pixels (e.g., 20 pixels) in any one of up and down directions in the driving direction and then moved by a predetermined number of unit pixels (e.g., 10 pixels) left or right in the steering direction may be merged.
- Furthermore, in order to merge images more precisely, the images can be merged more precisely by reducing the search area of an image according to information on the speed and steering angle of a vehicle and extracting image areas including pixels having similarity from the image having the reduced search area based on information on the driving of the vehicle including information on the speed and movement direction of the vehicle supplied from the vehicle.
- Meanwhile, if the pixels of areas having similarity are extracted from a past image and the remaining pixel areas located in the surroundings of the pixels areas having similarity on the past image, that is, pixels areas having low similarity are merged and merged, an image for display including a blind spot not seen in a current image can be generated using the past image.
FIG. 6 shows that blind spots covered by the vehicle body of a vehicle are displayed along with a vehicle icon on an image detected by the apparatus for displaying a blind spot.FIG. 6 a shows a first rear image A captured by the camera prior to a specific point of time (past) when a vehicle runs backward,FIG. 6 b shows a second rear image B captured by the camera at a specific point of time, (current) andFIG. 6 c shows an image A+B for display obtained by merging the first rear image A and the second rear image B. - As shown in
FIG. 6 c, when the second rear image B captured at the specific point of time (current) and the first rear image A captured at the point of time prior to the specific point of time (past) are merged, one long image A+B for display is generated, so that a blind spot not seen in the second rear image B (i.e., a portion covered by the vehicle body) can be seen. - The image A+B for display may be generated by merging the current image and the past image, but may be displayed using one long image.
- Meanwhile, the
control unit 130 can display thevehicle icon 400 using a contour in such a way as to overlap with the second rear image A so that a driver can check a current position of the vehicle on the generated image A+B for display and check a precise position of the vehicle that runs backward. - Furthermore, the
control unit 130 can also display the position of awheel 410 along with thevehicle icon 400 so that a driver can check whether thewheel 410 is deviated from a parking guide line or the vehicle runs backward in a precise direction, for example. - Furthermore,
FIG. 7 shows that a blind spot covered by the vehicle body of a vehicle is displayed along with a vehicle icon on an image captured by the apparatus for displaying a blind spot. When the vehicle runs forward,FIG. 7 a shows a second front image D of the current surroundings of the vehicle that runs forward captured by thesecond camera 120, andFIG. 7 b shows an image C+D for display obtained by merging a first front image C captured in the past prior to a specific time and the second front image D or an image for display generated by capturing one front image C+D including aspeed bump 300 in a specific point of time using thesecond camera 120. - When the vehicle continues to run and then enter an area C on the front image C+D merged with a current position of the vehicle after t seconds, the driver can recognize that the vehicle passes through the
speed bump 300 seen prior to the t seconds, but thereal speed bump 300 is covered by the vehicle body and located in a blind spot that is rarely seen within a field of view of the driver. - In order for the driver to recognize information on the
speed bump 300 located in the blind spot, thecontrol unit 130 calculates information on a movement displacement of the vehicle based on information on the movement distance, speed, and steering angle of the vehicle using a recognize distance and position calculation program, detects information on the movement distance of the vehicle from a specific point of time and on the position of the vehicle based on the calculated information, and displays avirtual vehicle icon 400, indicating a current position of the vehicle, in a corresponding area on the front image C+D (e.g., the area C) based on the detected information on the distance and position in a contour or semi-transparent form having a vehicle shape. Accordingly, the driver can recognize a blind spot under the vehicle that is covered by the vehicle body. - Furthermore, the
control unit 130 displays avirtual vehicle icon 400 in a place where the vehicle is now placed in a contour or semi-transparent form having a vehicle shape. Accordingly, the driver can recognize that the vehicle is now passing through a section in which thespeed bump 300 is installed and can check a blind spot area C-1 and a blind spot area C-2 on the right and left sides of thevehicle icon 400. - Furthermore, the
control unit 130 displays information on thevirtual vehicle icon 400 on the front image C+D in a contour or semi-transparent form having a vehicle shape so that the driver can recognize information on an obstacle and direction guide located in a road in the surroundings of the vehicle. - Furthermore, an example in which obstacles covered by the vehicle body of a vehicle and located under the vehicle and in a road in the surroundings of the vehicle and direction guide information are displayed for a driver is described in more detail below with reference to
FIG. 8 . -
FIG. 8 a shows information on a second front image D, that is, a current image,FIG. 8 b shows information on a first front image C, that is, a past image prior to a specific time (e.g., 0.5 second), andFIG. 8 c shows information on a front matching image C+D in whichFIGS. 8 a and 8 b are matched. - If a driver has not recognized direction guide
information 500 on a road while driving, thecontrol unit 130 displays avirtual vehicle icon 400, displayed on the front image C+D only in a contour or semi-transparent form having a vehicle shape so that the driver can recognize the direction guideinformation 500. - Furthermore, the
control unit 130 can check whether or not a fixed object or a moving object is included in a current image and a past image prior to a specific point of time and display information on the distance up to the object based on a result of the check. - That is, if an image is captured by one camera, the
control unit 130 detects information on the pixels of objects having similar characteristics (e.g., the brightness of pixels and whether or not an edge exists) in a current image and a past image by processing an image captured at a current point of time and an image at a past point of time prior to a specific time and detects information (pixel information) on a specific object including the displacement. - The
control unit 130 can recognize an object by detecting similarity through a movement displacement and an angular displacement of pixels defined as vector by comparing two interested images through image processing, can recognize an object by using edges, such as the boundary line or contour of an object or lines or curves approximate to the contour of an object, can recognize an object by using edge points present within an image, unchanged points, or characteristic points that are not changed in relation to rotational conversion or a change in the size, or can recognize an object by extracting the characteristic points of the object included in an image. - Furthermore, the
control unit 130 detects a movement displacement (i.e., vector) between a current position of an object detected in a current image and a past position of the object detected in a past image, detects information on the driving of a vehicle including information on a steering angle, distance and speed of the vehicle supplied from the ECU, TCU, wheel sensor and/or steering angle sensor of the vehicle, and calculates a movement distance corresponding to the movement displacements of the two objects. - That is, the
control unit 130 displays a current position of an object and a past position of the object on a current image and sets a difference value between the two objects as a movement distance between the two objects using a distance that the vehicle has moved. - Furthermore, the
control unit 130 detects a predetermined reference point on the current image and a displacement up to the current object and the past object, uses triangular information formed by the reference point and the current object and the past object, that is, a movement displacement between the current object and the past object, as a reference side, uses displacements formed by the reference point and the current object and by the reference point and the past object, respectively, as two other sides, calculates angles formed by the sides, and calculates the distance from the reference point to the current object based on the calculated angles. - Furthermore, if a current object is placed on the area of a past image in the calculated area of an image, the
control unit 130 can estimate a movement displacement of the object, calculate an expected position and distance up to the object placed on a specific area of an image after a lapse of a specific time based on the estimated movement displacement, the speed of a vehicle, and the relative speed of the current object detected based on the estimated movement displacement, and provide the calculated position and distance. - Furthermore, if two or more stereo cameras are used, the
control unit 130 can extract the distance up to an object through a stereo vision. Here, thecontrol unit 130 can estimate a movement displacement of the object, calculate an expected position and distance up to the object placed on a specific area of an image after a lapse of a specific time based on the estimated movement displacement, the speed of the vehicle, and the relative speed of the current object detected based on the estimated movement displacement, and provide the calculated position and distance. - Furthermore, the
control unit 130 compares the calculated distance with predetermined reference distance information for determining whether or not there is a collision. If, as a result of the comparison, a collision is expected, thecontrol unit 130 outputs a warning signal so that alarm or a warning screen is displayed. - When left and right images are received through stereo cameras installed in the head lamp of a vehicle, the taillight of the vehicle, or side-view mirrors on the left and right sides, the stereo vision program performs a pre-processing process on the stereo image input for matching brightness and epipolar lines for the left and right images, performs stereo matching for generating disparity between the pixels of areas overlapped with each other in the left and right images that are determined as the same object, removes noise components from the results of the disparity, obtains a stereo image by performing post-processing processes, such as projection, segmentation, and filtering, on information on the distance of the stereo image, and detects pieces of information, such as the distance, shape, and the direction of the object, through the stereo image.
- The stereo vision program is one embodiment, and a variety of known stereo vision programs may be used.
- Furthermore, the
control unit 130 can output the original image, captured by a camera, as an image for display without change, but may convert the captured original image into an image having a Bird Eye View point of time using a time point conversion program and output the converted image as an image for display. - Furthermore, the
control unit 130 may output the original image, captured by the camera, as an image for display or may merge several images whose points of time have been converted or stitch the several images using an image processing program and output the merged or stitched image as an image for display. - Furthermore, the
control unit 130 may combine the original image captured by the camera and some of the images whose points of time have been converted or the merged images using an image processing program and output the combined image as an image for display. - The
screen display unit 140 outputs an image outputted from thecontrol unit 130 and thevehicle icon 400 in a contour or semi-transparent form having a vehicle shape together and includes output means, such as an LCD. - The
vehicle icon 400 has a contour or semi-transparent form having a vehicle shape and virtually displays a current position of a vehicle in an image for display. - Furthermore, the
vehicle icon 400 may display the position of the wheel (not shown) of a vehicle along with a contour or semi-transparent form having a vehicle shape. Accordingly, when a vehicle runs, an obstacle under the vehicle covered by the vehicle body of the vehicle and driving direction signs and characters for road guidance indicated in roads can be precisely delivered to a driver. - Furthermore, in a change in the position of the
vehicle icon 400, only the position of thevehicle icon 400 indicated in an image for display may be changed in the state in which the image for display is fixed, only the image for display may be moved in the state in which the position of thevehicle icon 400 is fixed, or both the image for display and the position of thevehicle icon 400 may be changed. -
FIG. 9 is a flowchart illustrating a process of displaying a blind spot in accordance with the present invention. A method for displaying a blind spot in accordance with the present invention is described with reference toFIGS. 2 and 4 to 9. - When the
vehicle 200 runs forward at step S100, thesecond camera 120 installed in the vehicle starts photographing as thevehicle 200 runs, and thecontrol unit 130 detects an image C or C+D for display captured by thesecond camera 120 at a specific point of time at step S100 and temporarily stores the detected image. - Furthermore, the
control unit 130 can store and output the image for display captured at step S100 as the original image, but may convert the captured original image into an image having a Bird Eye View point of time using a time point conversion program and output the converted image as the image for display. Furthermore, the image for display captured at step S100 may be the original image captured by the camera, an image obtained by merging or stitching several images whose points of time have been converted by thecontrol unit 130 using an image processing program, or an image obtained by combining the original image and some of the images whose points of time have been converted and the merged images by thecontrol unit 130 using an image processing program. - After the step S100, the
control unit 130 detects information on the movement displacement of the vehicle, including information on the movement distance, speed, driving direction, and steering angle of the vehicle, based on driving control information calculated by comparing a current image captured by the ECU, TCU, speed sensor, steering angle sensor, wheel sensor, and camera of the vehicle supplied from the vehicle with an image captured right before the current image and calculates a position according to the movement distance and steering direction of the vehicle per hour based on the detected movement displacement information at step S110. - The
control unit 130 matches the calculated movement distance and position with the image C+D for display detected at step S100 according to a predetermined number of unit pixels and displays avirtual vehicle icon 400 indicative of a current position of the vehicle in a contour or semi-transparent form having a vehicle shape at step S120 so that thevirtual vehicle icon 400 is outputted through thescreen display unit 140. - The
virtual vehicle icon 400 outputted through thescreen display unit 140 at step S120 is displayed on the front image C+D in a contour or semi-transparent form having a vehicle shape so that obstacles (e.g., a speed bump, a boundary stone, and a median strip) within blind spots or direction guideinformation 500, together with information on the current position of the vehicle, can be easily recognized by a driver. - Accordingly, a safe driving environment and a safe parking environment can be provided to a driver because the driver can check blind spots covered by the vehicle body of a vehicle and thus rarely seen within a field of view of the driver and blind spots placed in the surroundings of the vehicle and not seen in a current image captured by a camera can be checked.
- Meanwhile, in the present embodiment, an example in which the position of a virtual vehicle icon indicating a current position of a vehicle on a displayed image is changed and displayed on an image for display as the vehicle moves has been described. It is evident that the position of the virtual vehicle icon may be changed and displayed on an image for display in which a plurality of images is continuously arranged as the vehicle moves.
-
FIG. 10 is a block diagram showing another embodiment of an apparatus for displaying a blind spot in accordance with the present invention. The second embodiment relates to a construction for detecting images of the surroundings of a vehicle from devices other than cameras installed in the vehicle and displaying a blind spot based on the detected images. - As shown in
FIG. 10 , an apparatus for displaying ablind spot 600 in accordance with the second embodiment includes aGPS navigation system 610 and an externalinformation reception unit 620 as imaging means for providing images of the surroundings of a vehicle, acontrol unit 630, and ascreen display unit 640. - The
GPS navigation system 610 is installed in the vehicle. TheGPS navigation system 610 receives information on the position of the vehicle from a GPS receiver, detects information on a current position of the vehicle based on the received information, reads a map image or real image of a corresponding position previously stored in a memory (not shown) based on the detected information on the position of the vehicle, and provides the read map image or real image. - The external
information reception unit 620 is connected to cameras (not shown) installed in roads and buildings in the surroundings of the vehicle over a radio network. The externalinformation reception unit 620 receives images captured by the cameras and provides the images to thecontrol unit 630. - A current intelligent road network system is equipped with fixed cameras for photographing some road sections or some areas of buildings, but the cameras are assigned unique IPs. Accordingly, a user can access the cameras using radio communication and easily receive and check images captured by the cameras.
- That is, the external
information reception unit 620 accesses, for example, IP cameras installed in roads or buildings in the surroundings of roads through radio communication, receives images provided by the IP cameras and information on positions where the IP cameras are installed, stores the received images and information in the memory (not shown), and provides the images received from the IP cameras to thecontrol unit 630. Furthermore, the externalinformation reception unit 620 detects a current position of the vehicle from theGPS navigation system 610 or a GPS receiver (not shown) along with information on positions where the IP cameras are installed supplied from the IP cameras and sends the detected current position and the information on the positions to thecontrol unit 630. - The
control unit 630 detects a map image of a corresponding position or an image of a position where the vehicle is placed, which are supplied by theGPS navigation system 610 or the externalinformation reception unit 620 at a specific point of time and generates an image for display using the detected one or more images without change or displays and outputs an image for display that has been reconfigured by merging the detected one or more images with an image captured at a specific point of time or images captured right before the specific point of time using an image processing program, converting points of time of the detected images using a time point conversion program, or combining some of the detected images. - That is, the
control unit 630 can output the original image captured by, for example, the IP camera as an image for display without change, but may convert the captured original image into an image having a Bird Eye View point of time using a time point conversion program and output the converted image as an image for display. - Furthermore, the
control unit 630 may output the original image captured by the IP camera as an image for display or may merge or stitch past several images whose points of time have been converted using an image processing program and output the merged or stitched image as an image for display. Thecontrol unit 630 may combine the original image captured by the IP camera and some of the images whose points of time have been converted or the merged images using an image processing program and output the combined image as an image for display. - Furthermore, the
control unit 630 compares driving control information, detected by the ECU, TCU, speed sensor, steering angle sensor, and wheel sensor of the vehicle, with driving control information received prior to a specific point of time and detects information on the movement displacement of the vehicle, such as the movement distance, speed, driving direction, and steering angle of the vehicle, based on a result of the comparison. - Furthermore, when the
GPS navigation system 610 detects information on the position of the vehicle, supplied outside the vehicle, through a GPS receiver and outputs the detected information, thecontrol unit 630 can compare a map image corresponding to the position information with information on the position of the vehicle and an image received prior to a specific point of time and detect information on a movement displacement of the vehicle based on a result of the comparison. - Furthermore, the
control unit 630 can compare information on the position of a camera detected by the street-view receiver 620 and information on a movement distance using surrounding images and theGPS navigation system 610 and detect information on a movement displacement of the vehicle based on a result of the comparison. - Furthermore, the
control unit 630 detects information on a current position of the vehicle that has moved from a specific position based on the detected information on the movement displacement of the vehicle and displays a contour or semi-transparent vehicle icon indicative of the detected current position of the vehicle on the generated image for display. - The
screen display unit 640 outputs the virtual vehicle icon having a contour or semi-transparent vehicle form along with the image outputted from thecontrol unit 630. Thescreen display unit 640 includes output means, such as an LCD. - As described above, the preferred embodiments of the present invention have been described, but those skilled in the art will understand that the present invention can be modified and changed in various ways without departing from the spirit and scope of the present invention described in the claims.
- Accordingly, the disposition of obstacles placed in blind spots that are covered by the vehicle body of a vehicle and thus rarely seen within a field of view of a driver in a current position can be easily checked using images of the surroundings of the vehicle. Accordingly, a safe driving environment and a safe parking environment can be provided to a driver.
Claims (17)
Applications Claiming Priority (5)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| KR10-2010-0138493 | 2010-12-30 | ||
| KR1020100138493A KR101097063B1 (en) | 2010-12-30 | 2010-12-30 | Blind spot indicator and method |
| KR1020110132051A KR101295618B1 (en) | 2011-12-09 | 2011-12-09 | Apparatus and method for displaying blind spot |
| KR10-2011-0132051 | 2011-12-09 | ||
| PCT/KR2011/010280 WO2012091476A2 (en) | 2010-12-30 | 2011-12-29 | Apparatus and method for displaying a blind spot |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20130300872A1 true US20130300872A1 (en) | 2013-11-14 |
| US9418556B2 US9418556B2 (en) | 2016-08-16 |
Family
ID=46383739
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US13/996,346 Active 2033-05-07 US9418556B2 (en) | 2010-12-30 | 2011-12-29 | Apparatus and method for displaying a blind spot |
Country Status (5)
| Country | Link |
|---|---|
| US (1) | US9418556B2 (en) |
| EP (1) | EP2660104B1 (en) |
| JP (1) | JP2013541915A (en) |
| CN (1) | CN103237685B (en) |
| WO (1) | WO2012091476A2 (en) |
Cited By (26)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20140292805A1 (en) * | 2013-03-29 | 2014-10-02 | Fujitsu Ten Limited | Image processing apparatus |
| US9221396B1 (en) * | 2012-09-27 | 2015-12-29 | Google Inc. | Cross-validating sensors of an autonomous vehicle |
| DE102014225673A1 (en) * | 2014-12-12 | 2016-06-30 | Volkswagen Aktiengesellschaft | Method for displaying a positioning aid for a vehicle |
| US20160314364A1 (en) * | 2014-02-17 | 2016-10-27 | Hitachi Automotive Systems, Ltd. | Vehicle-Mounted Recognition Device |
| US20160311342A1 (en) * | 2014-09-15 | 2016-10-27 | Dti Group Limited | Arcing filtering using multiple image capture devices |
| US20160375828A1 (en) * | 2013-12-16 | 2016-12-29 | Lg Electronics Inc. | Stereo camera and vehicle comprising same |
| US20170132476A1 (en) * | 2015-11-08 | 2017-05-11 | Otobrite Electronics Inc. | Vehicle Imaging System |
| US9902322B2 (en) | 2015-10-30 | 2018-02-27 | Bendix Commercial Vehicle Systems Llc | Filling in surround view areas blocked by mirrors or other vehicle parts |
| JP2018063294A (en) * | 2016-10-11 | 2018-04-19 | アイシン精機株式会社 | Display control device |
| US10017112B2 (en) * | 2015-03-03 | 2018-07-10 | Hitachi Construction Machinery Co., Ltd. | Surroundings monitoring device of vehicle |
| US20180201191A1 (en) * | 2013-03-28 | 2018-07-19 | Aisin Seiki Kabushiki Kaisha | Surroundings-monitoring device and computer program product |
| WO2018156760A1 (en) * | 2017-02-22 | 2018-08-30 | Kevin Smith | Method, system, and device for forward vehicular vision |
| US10168824B2 (en) | 2014-11-26 | 2019-01-01 | Lg Electronics Inc. | Electronic device and control method for the electronic device |
| US10217006B2 (en) * | 2015-08-31 | 2019-02-26 | Continental Automotive Gmbh | Method and device for detecting objects in the dark using a vehicle camera and a vehicle lighting system |
| US20190266416A1 (en) * | 2015-11-08 | 2019-08-29 | Otobrite Electronics Inc. | Vehicle image system and method for positioning vehicle using vehicle image |
| CN111942288A (en) * | 2019-05-14 | 2020-11-17 | 欧特明电子股份有限公司 | Vehicle image system and vehicle positioning method using vehicle image |
| US10855954B2 (en) * | 2016-06-29 | 2020-12-01 | Aisin Seiki Kabushiki Kaisha | Periphery monitoring device |
| US20210080568A1 (en) * | 2018-04-25 | 2021-03-18 | Waymo Llc | Underbody Radar Units |
| US20210162923A1 (en) * | 2018-05-22 | 2021-06-03 | Continental Automotive Gmbh | Method and Device for Displaying Vehicle Surroundings |
| US11227409B1 (en) | 2018-08-20 | 2022-01-18 | Waymo Llc | Camera assessment techniques for autonomous vehicles |
| US20220324387A1 (en) * | 2021-04-08 | 2022-10-13 | Toyota Jidosha Kabushiki Kaisha | Display control system, display control method, and non-transitory storage medium |
| US11597318B2 (en) | 2014-04-09 | 2023-03-07 | Jaguar Land Rover Limited | Apparatus and method for displaying information |
| US20230192073A1 (en) * | 2021-11-04 | 2023-06-22 | STEER-Tech, LLC | Methods and systems for parking a vehicle |
| US11699207B2 (en) | 2018-08-20 | 2023-07-11 | Waymo Llc | Camera assessment techniques for autonomous vehicles |
| EP4471729A1 (en) * | 2023-05-30 | 2024-12-04 | Ampere SAS | System and method for displaying a view of a vehicle covered ground area |
| EP4546769A1 (en) * | 2023-10-25 | 2025-04-30 | Toyota Jidosha Kabushiki Kaisha | Video display device |
Families Citing this family (27)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| DE102012005277B3 (en) * | 2012-03-15 | 2013-07-18 | Mekra Lang Gmbh & Co. Kg | Rear area vision system |
| KR101449160B1 (en) * | 2012-12-12 | 2014-10-08 | 현대자동차주식회사 | Apparatus and method for providing information of blind spot |
| JP6375633B2 (en) * | 2014-02-12 | 2018-08-22 | 株式会社デンソー | Vehicle periphery image display device and vehicle periphery image display method |
| JP6326869B2 (en) * | 2014-03-05 | 2018-05-23 | 株式会社デンソー | Vehicle periphery image display device and vehicle periphery image display method |
| US9211889B1 (en) * | 2014-07-29 | 2015-12-15 | Robert Bosch Gmbh | Enhanced blind spot detection for vehicle with trailer |
| KR20160112545A (en) | 2015-03-19 | 2016-09-28 | 현대자동차주식회사 | Vehicle and method for controlling vehicle |
| US10185319B2 (en) * | 2015-11-16 | 2019-01-22 | Ford Global Technologies, Llc | Method and device for assisting a parking maneuver |
| DE102016208369A1 (en) * | 2016-05-17 | 2017-12-07 | Bayerische Motoren Werke Aktiengesellschaft | Method for determining data representing part of the environment below the vehicle |
| CN106004661B (en) * | 2016-05-31 | 2018-12-07 | 百度在线网络技术(北京)有限公司 | Exterior rearview mirror system and method applied to the exterior rearview mirror system |
| EP3263405B1 (en) * | 2016-06-27 | 2019-08-07 | Volvo Car Corporation | Around view monitoring system and method for vehicles |
| DE112017004968T5 (en) * | 2016-09-30 | 2019-06-13 | Aisin Seiki Kabushiki Kaisha | Environment monitoring device |
| TWI599989B (en) * | 2016-11-29 | 2017-09-21 | 財團法人工業技術研究院 | Image processing method and image system for transportation |
| CN106828319A (en) * | 2017-01-16 | 2017-06-13 | 惠州市德赛西威汽车电子股份有限公司 | A kind of panoramic looking-around display methods for showing body bottom image |
| CN108422932B (en) * | 2017-02-15 | 2021-06-01 | 奥迪股份公司 | Driving assistance systems, methods and vehicles |
| CN106985751A (en) * | 2017-04-07 | 2017-07-28 | 深圳市歌美迪电子技术发展有限公司 | Backsight display methods, device and equipment |
| US10497265B2 (en) * | 2017-05-18 | 2019-12-03 | Panasonic Intellectual Property Corporation Of America | Vehicle system, method of processing vehicle information, recording medium storing a program, traffic system, infrastructure system, and method of processing infrastructure information |
| CN107274342A (en) * | 2017-05-22 | 2017-10-20 | 纵目科技(上海)股份有限公司 | A kind of underbody blind area fill method and system, storage medium, terminal device |
| US9934440B1 (en) * | 2017-10-04 | 2018-04-03 | StradVision, Inc. | Method for monitoring blind spot of monitoring vehicle and blind spot monitor using the same |
| KR102051324B1 (en) * | 2017-10-31 | 2019-12-03 | 주식회사 켐트로닉스 | Surround view monitoring system |
| CN109717842A (en) * | 2017-10-31 | 2019-05-07 | 北京帕斯泰克科技有限公司 | A kind of video aiding sensors positioning device and method for pulse-taking instrument equipment |
| CN108016354A (en) * | 2017-12-04 | 2018-05-11 | 惠州市德赛西威汽车电子股份有限公司 | A kind of visible panoramic parking system in picture blind area and its method |
| CN107878330B (en) * | 2017-12-06 | 2020-05-05 | 湖北航天技术研究院特种车辆技术中心 | Vehicle chassis perspective method and vehicle chassis perspective device |
| JP7172309B2 (en) * | 2018-09-06 | 2022-11-16 | 株式会社アイシン | Perimeter monitoring device |
| CN113039781B (en) * | 2018-11-22 | 2024-11-08 | 索尼半导体解决方案公司 | Image processing device, camera system and image processing method |
| CN111998780B (en) * | 2019-05-27 | 2022-07-01 | 杭州海康威视数字技术股份有限公司 | Target ranging method, device and system |
| US11140364B2 (en) * | 2019-09-09 | 2021-10-05 | Texas Instruments Incorporated | Sensor fusion based perceptually enhanced surround view |
| DE102022209182A1 (en) | 2022-09-05 | 2024-03-07 | Robert Bosch Gesellschaft mit beschränkter Haftung | Display method for displaying a virtual environment model, computer program, computing device and vehicle |
Citations (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6483429B1 (en) * | 1999-10-21 | 2002-11-19 | Matsushita Electric Industrial Co., Ltd. | Parking assistance system |
| US20060192660A1 (en) * | 2005-02-24 | 2006-08-31 | Aisin Seiki Kabushiki Kaisha | Vehicle surrounding monitoring device |
| US20070088478A1 (en) * | 2005-10-19 | 2007-04-19 | Aisin Aw Co., Ltd. | Vehicle travel distance calculation method, vehicle travel distance calculation apparatus, vehicle current position detection method and vehicle current postition detection apparatus |
| US20070097209A1 (en) * | 2005-10-27 | 2007-05-03 | Aisin Aw Co., Ltd. | Parking assist method and a parking assist apparatus |
| US20090122140A1 (en) * | 2007-11-09 | 2009-05-14 | Kosuke Imamura | Method and apparatus for generating a bird's-eye view image |
| US20090297036A1 (en) * | 2005-01-31 | 2009-12-03 | Daimler Ag | Object detection on a pixel plane in a digital image sequence |
| US20100097457A1 (en) * | 2008-04-24 | 2010-04-22 | Gm Global Technology Opetations, Inc. | Clear path detection with patch smoothing approach |
Family Cites Families (17)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP2002337605A (en) * | 2001-05-18 | 2002-11-27 | Auto Network Gijutsu Kenkyusho:Kk | Perimeter recognition device for vehicles |
| JP4593070B2 (en) * | 2001-12-12 | 2010-12-08 | 株式会社エクォス・リサーチ | Image processing apparatus for vehicle |
| JP2005001570A (en) * | 2003-06-12 | 2005-01-06 | Equos Research Co Ltd | Parking assistance device |
| JP4323377B2 (en) * | 2004-05-24 | 2009-09-02 | オリンパス株式会社 | Image display device |
| JP4470041B2 (en) * | 2004-06-22 | 2010-06-02 | 株式会社エクォス・リサーチ | Obstacle detection support device |
| JP4877447B2 (en) * | 2004-08-31 | 2012-02-15 | 株式会社エクォス・リサーチ | Vehicle peripheral image display device |
| JP2006238131A (en) * | 2005-02-25 | 2006-09-07 | Aisin Seiki Co Ltd | Vehicle periphery monitoring device |
| JP2006306241A (en) * | 2005-04-27 | 2006-11-09 | Aisin Aw Co Ltd | Parking assisting method and parking assisting device |
| JP2008042235A (en) * | 2006-08-01 | 2008-02-21 | Honda Motor Co Ltd | Vehicle display device |
| JP4967604B2 (en) * | 2006-11-01 | 2012-07-04 | スズキ株式会社 | Parking assistance system |
| KR100833603B1 (en) | 2006-12-29 | 2008-05-30 | 주식회사 엠티오메가 | Car navigation device and method for providing bird view |
| JP2007302238A (en) * | 2007-05-29 | 2007-11-22 | Equos Research Co Ltd | Image processing apparatus for vehicle |
| CN101953163B (en) * | 2008-02-20 | 2013-04-17 | 歌乐牌株式会社 | Vehicle peripheral image display system |
| JP2009225322A (en) * | 2008-03-18 | 2009-10-01 | Hyundai Motor Co Ltd | Vehicle information display system |
| KR101093316B1 (en) | 2008-04-15 | 2011-12-14 | 주식회사 만도 | Image Matching Method and System for Driving a Vehicle |
| JP5118605B2 (en) * | 2008-10-30 | 2013-01-16 | クラリオン株式会社 | Vehicle periphery image display system |
| JP5300443B2 (en) * | 2008-12-01 | 2013-09-25 | 富士通テン株式会社 | Image processing device |
-
2011
- 2011-12-29 CN CN201180057909.5A patent/CN103237685B/en active Active
- 2011-12-29 JP JP2013534836A patent/JP2013541915A/en active Pending
- 2011-12-29 EP EP11853246.4A patent/EP2660104B1/en active Active
- 2011-12-29 WO PCT/KR2011/010280 patent/WO2012091476A2/en not_active Ceased
- 2011-12-29 US US13/996,346 patent/US9418556B2/en active Active
Patent Citations (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6483429B1 (en) * | 1999-10-21 | 2002-11-19 | Matsushita Electric Industrial Co., Ltd. | Parking assistance system |
| US20090297036A1 (en) * | 2005-01-31 | 2009-12-03 | Daimler Ag | Object detection on a pixel plane in a digital image sequence |
| US20060192660A1 (en) * | 2005-02-24 | 2006-08-31 | Aisin Seiki Kabushiki Kaisha | Vehicle surrounding monitoring device |
| US20070088478A1 (en) * | 2005-10-19 | 2007-04-19 | Aisin Aw Co., Ltd. | Vehicle travel distance calculation method, vehicle travel distance calculation apparatus, vehicle current position detection method and vehicle current postition detection apparatus |
| US20070097209A1 (en) * | 2005-10-27 | 2007-05-03 | Aisin Aw Co., Ltd. | Parking assist method and a parking assist apparatus |
| US20090122140A1 (en) * | 2007-11-09 | 2009-05-14 | Kosuke Imamura | Method and apparatus for generating a bird's-eye view image |
| US20100097457A1 (en) * | 2008-04-24 | 2010-04-22 | Gm Global Technology Opetations, Inc. | Clear path detection with patch smoothing approach |
Cited By (41)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US9868446B1 (en) | 2012-09-27 | 2018-01-16 | Waymo Llc | Cross-validating sensors of an autonomous vehicle |
| US11518395B1 (en) | 2012-09-27 | 2022-12-06 | Waymo Llc | Cross-validating sensors of an autonomous vehicle |
| US12397805B1 (en) * | 2012-09-27 | 2025-08-26 | Waymo Llc | Cross-validating sensors of an autonomous vehicle |
| US9555740B1 (en) | 2012-09-27 | 2017-01-31 | Google Inc. | Cross-validating sensors of an autonomous vehicle |
| US11872998B1 (en) | 2012-09-27 | 2024-01-16 | Waymo Llc | Cross-validating sensors of an autonomous vehicle |
| US9221396B1 (en) * | 2012-09-27 | 2015-12-29 | Google Inc. | Cross-validating sensors of an autonomous vehicle |
| US20180201191A1 (en) * | 2013-03-28 | 2018-07-19 | Aisin Seiki Kabushiki Kaisha | Surroundings-monitoring device and computer program product |
| US10710504B2 (en) * | 2013-03-28 | 2020-07-14 | Aisin Seiki Kabushiki Kaisha | Surroundings-monitoring device and computer program product |
| US9646572B2 (en) * | 2013-03-29 | 2017-05-09 | Fujitsu Ten Limited | Image processing apparatus |
| US20140292805A1 (en) * | 2013-03-29 | 2014-10-02 | Fujitsu Ten Limited | Image processing apparatus |
| US20160375828A1 (en) * | 2013-12-16 | 2016-12-29 | Lg Electronics Inc. | Stereo camera and vehicle comprising same |
| US9931983B2 (en) * | 2013-12-16 | 2018-04-03 | Lg Electronics Inc. | Stereo camera and vehicle comprising same |
| US20160314364A1 (en) * | 2014-02-17 | 2016-10-27 | Hitachi Automotive Systems, Ltd. | Vehicle-Mounted Recognition Device |
| US10019635B2 (en) * | 2014-02-17 | 2018-07-10 | Hitachi Automotive Systems, Ltd. | Vehicle-mounted recognition device |
| US11597318B2 (en) | 2014-04-09 | 2023-03-07 | Jaguar Land Rover Limited | Apparatus and method for displaying information |
| US10471848B2 (en) * | 2014-09-15 | 2019-11-12 | Dti Group Limited | Arcing filtering using multiple image capture devices |
| US20160311342A1 (en) * | 2014-09-15 | 2016-10-27 | Dti Group Limited | Arcing filtering using multiple image capture devices |
| US10168824B2 (en) | 2014-11-26 | 2019-01-01 | Lg Electronics Inc. | Electronic device and control method for the electronic device |
| DE102014225673B4 (en) * | 2014-12-12 | 2020-07-09 | Volkswagen Aktiengesellschaft | Method and system for displaying a positioning aid for a vehicle |
| DE102014225673A1 (en) * | 2014-12-12 | 2016-06-30 | Volkswagen Aktiengesellschaft | Method for displaying a positioning aid for a vehicle |
| US10017112B2 (en) * | 2015-03-03 | 2018-07-10 | Hitachi Construction Machinery Co., Ltd. | Surroundings monitoring device of vehicle |
| US10217006B2 (en) * | 2015-08-31 | 2019-02-26 | Continental Automotive Gmbh | Method and device for detecting objects in the dark using a vehicle camera and a vehicle lighting system |
| US9902322B2 (en) | 2015-10-30 | 2018-02-27 | Bendix Commercial Vehicle Systems Llc | Filling in surround view areas blocked by mirrors or other vehicle parts |
| US20190266416A1 (en) * | 2015-11-08 | 2019-08-29 | Otobrite Electronics Inc. | Vehicle image system and method for positioning vehicle using vehicle image |
| US20170132476A1 (en) * | 2015-11-08 | 2017-05-11 | Otobrite Electronics Inc. | Vehicle Imaging System |
| US10855954B2 (en) * | 2016-06-29 | 2020-12-01 | Aisin Seiki Kabushiki Kaisha | Periphery monitoring device |
| JP2018063294A (en) * | 2016-10-11 | 2018-04-19 | アイシン精機株式会社 | Display control device |
| WO2018156760A1 (en) * | 2017-02-22 | 2018-08-30 | Kevin Smith | Method, system, and device for forward vehicular vision |
| US20210080568A1 (en) * | 2018-04-25 | 2021-03-18 | Waymo Llc | Underbody Radar Units |
| US20210162923A1 (en) * | 2018-05-22 | 2021-06-03 | Continental Automotive Gmbh | Method and Device for Displaying Vehicle Surroundings |
| US12060011B2 (en) * | 2018-05-22 | 2024-08-13 | Continental Automotive Gmbh | Method and device for displaying vehicle surroundings |
| US11227409B1 (en) | 2018-08-20 | 2022-01-18 | Waymo Llc | Camera assessment techniques for autonomous vehicles |
| US12056898B1 (en) | 2018-08-20 | 2024-08-06 | Waymo Llc | Camera assessment techniques for autonomous vehicles |
| US11699207B2 (en) | 2018-08-20 | 2023-07-11 | Waymo Llc | Camera assessment techniques for autonomous vehicles |
| CN111942288A (en) * | 2019-05-14 | 2020-11-17 | 欧特明电子股份有限公司 | Vehicle image system and vehicle positioning method using vehicle image |
| US20220324387A1 (en) * | 2021-04-08 | 2022-10-13 | Toyota Jidosha Kabushiki Kaisha | Display control system, display control method, and non-transitory storage medium |
| US11912265B2 (en) * | 2021-11-04 | 2024-02-27 | STEER-Tech, LLC | Methods and systems for parking a vehicle |
| US20230192073A1 (en) * | 2021-11-04 | 2023-06-22 | STEER-Tech, LLC | Methods and systems for parking a vehicle |
| EP4471729A1 (en) * | 2023-05-30 | 2024-12-04 | Ampere SAS | System and method for displaying a view of a vehicle covered ground area |
| FR3149262A1 (en) * | 2023-05-30 | 2024-12-06 | Renault S.A.S | System and method for displaying a view of a ground area covered by a vehicle |
| EP4546769A1 (en) * | 2023-10-25 | 2025-04-30 | Toyota Jidosha Kabushiki Kaisha | Video display device |
Also Published As
| Publication number | Publication date |
|---|---|
| EP2660104B1 (en) | 2015-09-23 |
| WO2012091476A2 (en) | 2012-07-05 |
| EP2660104A4 (en) | 2014-06-11 |
| JP2013541915A (en) | 2013-11-14 |
| CN103237685B (en) | 2017-06-13 |
| US9418556B2 (en) | 2016-08-16 |
| WO2012091476A3 (en) | 2012-08-23 |
| EP2660104A2 (en) | 2013-11-06 |
| CN103237685A (en) | 2013-08-07 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US9418556B2 (en) | Apparatus and method for displaying a blind spot | |
| US8044781B2 (en) | System and method for displaying a 3D vehicle surrounding with adjustable point of view including a distance sensor | |
| US8089512B2 (en) | Driving support device, driving support method and computer program | |
| US8289391B2 (en) | Apparatus for vehicle surroundings monitoring | |
| KR101243108B1 (en) | Apparatus and method for displaying rear image of vehicle | |
| EP2544449B1 (en) | Vehicle perimeter monitoring device | |
| KR101097063B1 (en) | Blind spot indicator and method | |
| US20080198226A1 (en) | Image Processing Device | |
| CN111414796A (en) | Adaptive transparency of virtual vehicles in analog imaging systems | |
| US20110228980A1 (en) | Control apparatus and vehicle surrounding monitoring apparatus | |
| US20150042799A1 (en) | Object highlighting and sensing in vehicle image display systems | |
| US9965690B2 (en) | On-vehicle control device | |
| EP3530521B1 (en) | Driver assistance method and apparatus | |
| JP7426174B2 (en) | Vehicle surrounding image display system and vehicle surrounding image display method | |
| JP2004240480A (en) | Driving support device | |
| CN108973858B (en) | Device for ensuring safety of driving route | |
| CN110378836B (en) | Method, system and equipment for acquiring 3D information of object | |
| JP5888339B2 (en) | Display control device | |
| JP2015154125A (en) | Vehicle periphery image display device and vehicle periphery image display method | |
| JP2010028803A (en) | Image displaying method for parking aid | |
| JP5539250B2 (en) | Approaching object detection device and approaching object detection method | |
| JP4872245B2 (en) | Pedestrian recognition device | |
| KR101295618B1 (en) | Apparatus and method for displaying blind spot | |
| JP2000293693A (en) | Obstacle detection method and device | |
| JP2017056909A (en) | Vehicular image display device |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| AS | Assignment |
Owner name: WISE AUTOMOTIVE CORPORATION, KOREA, REPUBLIC OF Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:PARK, JAE-HONG;REEL/FRAME:030894/0513 Effective date: 20130725 |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
| MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2551); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY Year of fee payment: 4 |
|
| MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2552); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY Year of fee payment: 8 |