[go: up one dir, main page]

WO2024114119A1 - Sensor fusion method based on binocular camera guidance - Google Patents

Sensor fusion method based on binocular camera guidance Download PDF

Info

Publication number
WO2024114119A1
WO2024114119A1 PCT/CN2023/123988 CN2023123988W WO2024114119A1 WO 2024114119 A1 WO2024114119 A1 WO 2024114119A1 CN 2023123988 W CN2023123988 W CN 2023123988W WO 2024114119 A1 WO2024114119 A1 WO 2024114119A1
Authority
WO
WIPO (PCT)
Prior art keywords
target
information
point cloud
binocular camera
lidar
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/CN2023/123988
Other languages
French (fr)
Chinese (zh)
Inventor
陈志�
戴鹏程
王文军
张世泽
胡基贵
孙兆聪
范俊俊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
CRRC Nanjing Puzhen Co Ltd
Original Assignee
CRRC Nanjing Puzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by CRRC Nanjing Puzhen Co Ltd filed Critical CRRC Nanjing Puzhen Co Ltd
Publication of WO2024114119A1 publication Critical patent/WO2024114119A1/en
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/66Tracking systems using electromagnetic waves other than radio waves
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/86Combinations of lidar systems with systems other than lidar, radar or sonar, e.g. with direction finders
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/48Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
    • G01S7/497Means for monitoring or calibrating
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • G06V10/75Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/762Arrangements for image or video recognition or understanding using pattern recognition or machine learning using clustering, e.g. of similar faces in social networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Definitions

  • the present invention relates to the technical field of intelligent vehicle environment perception, and in particular to a sensor fusion method based on binocular camera guidance.
  • the method of using multiple sensors for fusion detection has gradually developed.
  • This detection method can also improve the robustness of the system in complex environments. Therefore, the detection method of multi-sensor information fusion is used more and more frequently in the research of target detection.
  • the data fusion of multiple sensors is divided into three levels: decision layer fusion, feature layer fusion and data layer fusion.
  • decision layer fusion decision layer fusion
  • feature layer fusion data layer fusion
  • data layer fusion In the paper "Outdoor Positioning by Fusion of Binocular Vision and 2D LiDAR”, Liu Zhengxuan used binocular vision to detect the relative posture of the vehicle, and combined the LiDAR data at multiple moments in a time window for fusion, and finally obtained a local sub-graph.
  • Patent "CN114463303A” proposes a road based on the fusion of binocular camera and laser radar Target detection method, this method combines LiDAR technology with visual sensors, and combines the YOLOv4 detection model to fuse the radar and binocular vision results, thereby providing a smarter and more accurate technical means for target detection;
  • Patent "CN111340797A” proposes a LiDAR and binocular camera data fusion detection method and system, this method uses the detection results of the binocular camera to fuse with the detection results of the LiDAR, and determines whether the fusion overlap rate is greater than a certain threshold. If it is greater than the set value, the fusion result is output, otherwise the target detection needs to be re-performed. This method improves the measurement accuracy.
  • the present invention proposes a sensor fusion method based on binocular camera guidance, which indirectly realizes the data fusion of binocular camera and lidar at the data layer. It can combine the advantages of binocular camera and lidar sensor, effectively improve the accuracy and robustness of target detection, and meet the needs of engineering applications.
  • the present invention provides a sensor fusion method based on binocular camera guidance, which is used to solve the problems existing in the background technology.
  • a sensor fusion method based on binocular camera guidance includes the following steps:
  • Step 1 Based on the visual image information collected by the binocular camera and the 3D point cloud information collected by the lidar, calibrate the left and right cameras that make up the binocular camera, and jointly calibrate the lidar and the left camera;
  • Step 2 Collect the forward environmental information of the electronically guided rubber-wheeled vehicle through binocular cameras and laser radar;
  • Step 3 Calculate the depth information of the front environment of the electronic guided rubber-wheeled vehicle through a stereo matching algorithm
  • Step 4 Obtain the visual information of the target in the image captured by the camera, including the target category and two-dimensional plane position information, through the target detection neural network based on YOLOv7;
  • Step 5 Combine the target visual information, environmental depth information and target time information output by the neural network to obtain the target detection result based on stereo vision;
  • Step 6 Based on the target time information in the target detection result obtained in step 5, establish the target's motion trajectory and predict the target's three-dimensional position information at the laser radar acquisition time;
  • Step 7 Based on the three-dimensional position information of the target predicted in step 6, segment and cluster the point cloud collected by the lidar, and output the target detection result;
  • Step 8 Track the target detected by the lidar based on the Kalman filter algorithm and the nearest neighbor algorithm, and correct and output the fused target based on the operation scenario.
  • step 1 The specific steps of performing camera calibration in step 1 are as follows:
  • the visual image information collected by the binocular camera is calibrated by Zhang Zhengyou method to obtain the left camera internal parameters ML , distortion coefficient DL , the right camera internal parameters MR , distortion coefficient DR and the external parameters MLR between the left and right cameras.
  • step 1 The specific steps of performing joint calibration in step 1 are as follows:
  • the calibration plate is used to perform joint calibration on the three-dimensional point cloud information collected by the laser radar and the two-dimensional position information in the visual image information collected by the left camera to obtain the external parameters M LC between the laser radar and the left camera.
  • step 3 The calculation method in step 3 is as follows:
  • the corresponding pixel points of the left and right views in the same scene are matched to obtain the disparity map.
  • the projection matrix Q from the two-dimensional position of the image to the three-dimensional position in space is obtained based on the Bouguet algorithm to calculate the target depth information.
  • the step four includes: collecting and creating a data set of target images in the operating environment, using the data set with existing target annotations as a training set for the neural network, obtaining the weight coefficient of the neural network through the training model, and then predicting the image captured by the left camera to obtain the target detection result containing the target category and the two-dimensional bounding box information describing the target plane position.
  • the step five includes: using the calculation method in step three to project the target plane information obtained in step four into three-dimensional space, obtaining the target detection result of the stereoscopic information described by the three-dimensional bounding box, and then combining the image timestamp to obtain the target detection result based on stereoscopic vision.
  • the step seven comprises:
  • the acquired stereo position information is clustered into point clouds to obtain target point cloud clusters, and the point cloud target detection results of the stereo information described by a three-dimensional bounding box are output.
  • the target information obtained by the binocular camera guides the laser radar to accurately measure the distance to the target;
  • the laser radar obtains the target category and historical trajectory information detected by the binocular camera, models and predicts the spatial position of the target, and assists in defining the detection range;
  • motion distortion compensation is used to eliminate the motion distortion inside the point cloud, as follows:
  • the IMU is used to compensate for the motion distortion of the lidar. While receiving the lidar scanning data, the data and the vehicle attitude angle data collected by the IMU are saved in the same circular queue.
  • the processor calculates the adoption time of each laser point based on the sampling time interval of the lidar and the data timestamp, and searches for two temporally adjacent frames of data in the IMU data queue. The laser points at the same measurement moment are paired with the vehicle motion attitude by using spherical linear interpolation.
  • the method for tracking the target based on the Kalman filter algorithm and the nearest neighbor algorithm in step eight is as follows:
  • the Kalman filter group is constructed and initialized, and the relevant parameters are configured to calculate the target.
  • the center of mass position of the point cloud is determined by the Kalman filter group.
  • the center of mass position of the target at the current timestamp is used to predict the center of mass position of the target at the next frame of the point cloud.
  • the nearest neighbor algorithm is used to search for a matching combination of real targets and predicted targets to achieve target association between frames, that is, to obtain the spatial position of the same target in two adjacent frames.
  • the target state of the Kalman filter group is updated using the spatial position and timestamp information of the target in the next frame of the point cloud, and then the target state is predicted. The above process is repeated to track the target.
  • the present invention has the following beneficial effects: the present invention adopts a binocular vision-guided fusion method, which can use visual target information to establish its motion trajectory, and then predict its position in three-dimensional space; adopts a data layer fusion method, uses visual prediction results to assist laser radar in defining the detection range, greatly reduces the amount of computational data, and has low requirements on computing platform performance;
  • the target detection results obtained based on binocular vision can accurately obtain the target depth, expand the target detection information and enhance the robustness of target detection.
  • FIG1 is a schematic diagram of a sensor fusion method of the present invention.
  • FIG2 is a flow chart of a sensor fusion method of the present invention.
  • Figure 3 is a flow chart of LiDAR point cloud clustering
  • Figure 4 is the effect diagram after point cloud clustering
  • Figure 5 is a schematic diagram of target tracking.
  • a sensor fusion method based on binocular camera guidance includes the following steps:
  • Step 1 Based on the visual image information collected by the binocular camera and the 3D point cloud information collected by the lidar, calibrate the left and right cameras that make up the binocular camera, and jointly calibrate the lidar and the left camera;
  • the fusion method proposed in the present invention first performs target detection and network tracking based on a binocular camera.
  • a binocular camera is generally composed of two horizontally placed left and right cameras. The two cameras image the target respectively. Since there is a baseline of a certain length between the cameras, the imaging position of the target will be different. The positional relationship between the target and the camera can be calculated using the difference in imaging position and the baseline length.
  • step 1 The specific steps for camera calibration in step 1 are as follows:
  • the visual image information collected by the binocular camera is calibrated by Zhang Zhengyou method to obtain the left camera internal parameters ML , distortion coefficient DL , the right camera internal parameters MR , distortion coefficient DR and the external parameters MLR between the left and right cameras;
  • step 1 The specific steps for joint calibration in step 1 are as follows:
  • the calibration plate is used to jointly calibrate the three-dimensional point cloud information collected by the laser radar and the two-dimensional position information in the visual image information collected by the left camera to obtain the external parameters M LC between the laser radar and the left camera, and then the measurement values in each coordinate system are converted to the train coordinate system;
  • Binocular correction dedistorts and aligns the views captured by the two cameras according to the relative position relationship obtained during camera calibration and the internal parameter data of the left and right cameras, so that the coordinates of the imaging origins of the left and right views are consistent, the optical axes of the two cameras are parallel, and the epipolar lines are aligned and coplanar with the imaging planes of the left and right cameras;
  • Step 2 Collect the forward environmental information of the electronically guided rubber-wheeled vehicle through binocular cameras and laser radar;
  • Step 2 is performed on the basis of keeping the positions of the binocular camera and the laser radar unchanged;
  • Step 3 Calculate the depth information of the front environment of the electronic guided rubber-wheeled vehicle through a stereo matching algorithm
  • step 3 The calculation method in step 3 is as follows: match the pixel points corresponding to the left and right views in the same scene to obtain the disparity map, obtain the projection matrix Q from the two-dimensional position of the image to the three-dimensional position of the space based on the Bouguet algorithm, and calculate the target depth information;
  • Step 4 Obtain the visual information of the target in the image captured by the camera, including the target category and two-dimensional plane position information, through the target detection neural network based on YOLOv7;
  • the above step 4 is as follows: collect and create a dataset of target images in the operating environment, use the dataset with existing target annotations as the training set of the YOLOv7 neural network, obtain the weight coefficient of the neural network through the training model, and then predict the images captured by the left camera to obtain the target category and description target.
  • the target detection result of the two-dimensional bounding box information of the plane position the present invention uses the target detection algorithm based on deep learning to detect the target in a single image, which can identify traffic participants such as pedestrians and vehicles, and classify them to obtain the target detection result based on stereo vision;
  • YOLOv7 adopted in the present invention improves the detection speed and accuracy by reforming multiple internal architectures based on YOLOv5;
  • YOLOv7 mainly uses MP structure and ELAN, which can perform deep learning more efficiently by controlling the shortest gradient path and deeper network structure;
  • the loss function of YOLOv7 is divided into three main parts: target confidence loss, coordinate loss and classification loss, which mainly performs data candidate matching in the matching strategy process, retains the data with the smallest loss, and realizes high-precision target detection;
  • Step 5 Combine the target visual information, environmental depth information and target time information output by the neural network to obtain the target detection result based on stereo vision;
  • the above step 5 uses the calculation method in step 3 to project the target plane information obtained in step 4 into three-dimensional space, obtain the target detection result of the three-dimensional information described by the three-dimensional bounding box, and then combine it with the image timestamp to obtain the target detection result based on stereoscopic vision.
  • Step 6 Based on the target time information in the target detection result obtained in step 5, establish the target's motion trajectory and predict the target's three-dimensional position information at the laser radar acquisition time;
  • Step 7 Based on the three-dimensional position information of the target predicted in step 6, segment and cluster the point cloud collected by the lidar, and output the target detection result;
  • step seven is specifically as follows: In combination with the operating environment characteristics of the electronic guided rubber-wheeled vehicle, the RANSAC algorithm is used to segment the ground, and the non-ground target point cloud is extracted.
  • the point cloud is clustered based on the stereo position information obtained in step six to obtain the target point cloud cluster, and the point cloud target detection result of the stereo information described by the three-dimensional bounding box is output.
  • the target information obtained by the binocular camera is used to guide the lidar to accurately measure the distance to the target; generally, the binocular camera is used to obtain the category and position of the target, and the sampling delay of the image and the processing time of the target detection algorithm need to be considered.
  • the lidar point cloud clustering mainly adopts the Euclidean clustering method for clustering.
  • This clustering method believes that in all the point cloud data collected by the lidar, the distance between any two points in the point cloud of the same object will be less than a certain value. Therefore, points with a distance less than a certain value can be clustered.
  • This clustering method has high accuracy, but its clustering speed is not advantageous. Although this method uses KD-Tree to speed up the data search speed, it is still insufficient in high-speed changing scenes.
  • the present invention innovatively uses binocular cameras to guide laser radars to cluster targets, which can not only improve the clustering speed, but also have better results on the accuracy and robustness of the final recognition results.
  • the steps of point cloud clustering are as follows:
  • the laser radar obtains the target category and historical trajectory information detected by the binocular camera, models and predicts the spatial position of the target, and assists in defining the detection range;
  • the IMU is used to compensate for the motion distortion of the laser radar. While receiving the laser radar scanning data, the data and the vehicle attitude angle data collected by the IMU are saved in the same circular queue.
  • the processor calculates the adoption time of each laser point according to the sampling time interval of the laser radar and the data timestamp, and searches for two frames of data adjacent in time in the IMU data queue.
  • the laser points at the same measurement moment are paired with the vehicle motion posture by using spherical linear interpolation. This allows the coordinates of each laser point of a frame of laser radar scanning data to be converted to the corresponding position coordinate system of the electronic guided rubber-wheeled vehicle at the same moment, so that all point clouds collected by the laser radar within a period of time are unified to one moment. Realize motion distortion compensation for lidar.
  • Step 8 Track the target detected by the lidar based on the Kalman filter algorithm and the nearest neighbor algorithm, and correct and output the fused target based on the operation scenario;
  • the method for tracking the target based on the Kalman filter algorithm and the nearest neighbor algorithm in the above step eight is as follows:
  • a Kalman filter group is constructed and initialized, and related parameters are configured.
  • the centroid position of the target point cloud is calculated.
  • the centroid position of the target at the next frame point cloud is predicted by the centroid position of the target at the current timestamp.
  • a matching combination of real targets and predicted targets is searched based on the nearest neighbor algorithm to achieve target association between frames, that is, the spatial position of the same target in two adjacent frames is obtained.
  • the target state of the Kalman filter group is updated using the spatial position and timestamp information of the target in the next frame point cloud, and then the target state is predicted. The above process is repeated to achieve target tracking.
  • the present invention proposes a sensor fusion framework guided by a binocular camera, and provides a fusion method of a sensor data layer guided by a binocular camera, which takes into account target detection performance and engineering applications, realizes the complementary advantages between sensors with different functions, and improves the intelligence of the target detection system.
  • the visual target detection system based on deep learning is mature and has lower performance requirements on the computing platform.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Multimedia (AREA)
  • Health & Medical Sciences (AREA)
  • Software Systems (AREA)
  • Artificial Intelligence (AREA)
  • Computing Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Remote Sensing (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Databases & Information Systems (AREA)
  • Electromagnetism (AREA)
  • Medical Informatics (AREA)
  • Computational Linguistics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biophysics (AREA)
  • Molecular Biology (AREA)
  • Mathematical Physics (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Biomedical Technology (AREA)
  • Image Processing (AREA)
  • Image Analysis (AREA)

Abstract

Disclosed in the present invention is a sensor fusion method based on binocular camera guidance. In the present invention, a fusion method based on binocular vision guidance is used, wherein a motion trajectory of a target can be established by using visual target information, and the position of the target in a three-dimensional space is then predicted; a data layer fusion method is used, wherein a visual prediction result is used to assist a laser radar in delimiting a detection range, such that the volume of operational data is greatly reduced, and the requirement for the performance of a computing platform is low; and on the basis of a target detection result acquired by means of binocular vision, a target depth can be accurately acquired, target detection information is expanded, and the robustness of target detection is enhanced.

Description

一种基于双目相机引导的传感器融合方法A sensor fusion method based on binocular camera guidance 技术领域Technical Field

本发明涉及智能车辆环境感知技术领域,特别涉及一种基于双目相机引导的传感器融合方法。The present invention relates to the technical field of intelligent vehicle environment perception, and in particular to a sensor fusion method based on binocular camera guidance.

背景技术Background technique

目前已有具体的研究工作与工程应用采用视觉或激光雷达等单一类型的传感器实现目标检测的功能。视觉传感器由于成本低、易于目标识别与采集信息数据丰富等特点被广泛地应用在目标检测中。激光雷达具有主动探测目标的能力,测距精度较高,近年来随着雷达工作原理的优化与升级,具备量产的可能。随着交通状况的日益复杂,基于单一传感器的检测方式受到算法复杂、检测信息量不足的制约,已经难以满足当前对于目标探测鲁棒性与检测速度的要求;At present, there are specific research works and engineering applications that use a single type of sensor such as vision or lidar to achieve the function of target detection. Visual sensors are widely used in target detection due to their low cost, easy target identification and rich information data collection. Lidar has the ability to actively detect targets and has high ranging accuracy. In recent years, with the optimization and upgrading of radar working principles, it has the possibility of mass production. With the increasing complexity of traffic conditions, the detection method based on a single sensor is restricted by the complexity of the algorithm and the insufficient amount of detection information, and it is difficult to meet the current requirements for target detection robustness and detection speed;

为了弥补单一传感器检测效果的不足,采用多种传感器进行融合检测的方法逐渐发展。这种检测方式在复杂的环境下还能提高系统的鲁棒性,因此多传感器信息融合的检测方式在目标检测的研究中使用得越来越频繁。目前将多传感器的数据融合分为决策层融合、特征层融合与数据层融合三个层次。刘诤轩在论文《融合双目视觉和2D激光雷达的室外定位》中采用双目视觉检测车辆的相对位姿,并结合一个时间窗口内多个时刻的激光雷达数据进行融合,最终得到一个局部子图,经过去噪声与匹配处理后实现车辆的高精度定位。巩朝光在论文《基于激光雷达与视觉融合的SLAM方法研究》中设计了激光雷达与双目相机的分类融合框架,通过对提取的视觉特征与激光点云进行分类与匹配,再根据匹配结果对特征线段进行合并,得到融合后的点云信息。这两篇论文都在双目相机与激光雷达的数据层层面进行数据融合,精度高但不利于工程实现;In order to make up for the shortcomings of the detection effect of a single sensor, the method of using multiple sensors for fusion detection has gradually developed. This detection method can also improve the robustness of the system in complex environments. Therefore, the detection method of multi-sensor information fusion is used more and more frequently in the research of target detection. At present, the data fusion of multiple sensors is divided into three levels: decision layer fusion, feature layer fusion and data layer fusion. In the paper "Outdoor Positioning by Fusion of Binocular Vision and 2D LiDAR", Liu Zhengxuan used binocular vision to detect the relative posture of the vehicle, and combined the LiDAR data at multiple moments in a time window for fusion, and finally obtained a local sub-graph. After denoising and matching, the high-precision positioning of the vehicle was achieved. In the paper "Research on SLAM Method Based on LiDAR and Vision Fusion", Gong Chaoguang designed a classification and fusion framework for LiDAR and binocular cameras. By classifying and matching the extracted visual features with the laser point cloud, the feature line segments are merged according to the matching results to obtain the fused point cloud information. Both papers perform data fusion at the data layer level of binocular cameras and LiDARs, which has high accuracy but is not conducive to engineering implementation;

专利“CN114463303A”提出了一种基于双目相机和激光雷达融合的道路 目标检测方法,该方法将激光雷达技术与视觉传感器相结合,结合YOLOv4检测模型对雷达与双目视觉结果进行融合,从而为目标检测提供更智准确的技术手段;专利“CN111340797A”提出了一种激光雷达与双目相机数据融合检测方法及系统,该方法利用双目相机的检测结果与激光雷达的检测结果进行融合,并判断融合重叠率是否大于一定阈值,若大于设定值则输出融合结果,否则需要重新进行目标检测,该方法提高了测量精度。这两篇专利在双目相机与激光雷达的决策层进行融合,有较好的工程效果但不利于性能的提升。现有研究大多利用传统方法进行相机与激光雷达数据层或决策层的融合以进行目标检测,这些方法较难实现工程应用性与检测性能的平衡;Patent "CN114463303A" proposes a road based on the fusion of binocular camera and laser radar Target detection method, this method combines LiDAR technology with visual sensors, and combines the YOLOv4 detection model to fuse the radar and binocular vision results, thereby providing a smarter and more accurate technical means for target detection; Patent "CN111340797A" proposes a LiDAR and binocular camera data fusion detection method and system, this method uses the detection results of the binocular camera to fuse with the detection results of the LiDAR, and determines whether the fusion overlap rate is greater than a certain threshold. If it is greater than the set value, the fusion result is output, otherwise the target detection needs to be re-performed. This method improves the measurement accuracy. These two patents fuse the decision layer of the binocular camera and the LiDAR, which has a good engineering effect but is not conducive to performance improvement. Most existing studies use traditional methods to fuse the camera and LiDAR data layer or decision layer for target detection. These methods are difficult to achieve a balance between engineering applicability and detection performance;

基于目前的研究现状,本发明提出了一种基于双目相机引导的传感器融合方法,它间接实现了双目相机与激光雷达在数据层进行数据融合,能够结合双目相机与激光雷达传感器的优势,有效地提升目标探测的准确性与鲁棒性,满足工程应用的需求。Based on the current research status, the present invention proposes a sensor fusion method based on binocular camera guidance, which indirectly realizes the data fusion of binocular camera and lidar at the data layer. It can combine the advantages of binocular camera and lidar sensor, effectively improve the accuracy and robustness of target detection, and meet the needs of engineering applications.

发明内容Summary of the invention

本发明提供一种基于双目相机引导的传感器融合方法,用以解决背景技术中所存在的问题。The present invention provides a sensor fusion method based on binocular camera guidance, which is used to solve the problems existing in the background technology.

一种基于双目相机引导的传感器融合方法,包括如下步骤:A sensor fusion method based on binocular camera guidance includes the following steps:

步骤一:基于双目相机采集的视觉图像信息以及激光雷达采集的三维点云信息,对组成双目相机的左相机和右相机进行相机标定,对激光雷达与左相机进行联合标定;Step 1: Based on the visual image information collected by the binocular camera and the 3D point cloud information collected by the lidar, calibrate the left and right cameras that make up the binocular camera, and jointly calibrate the lidar and the left camera;

步骤二:通过双目相机与激光雷达采集电子导向胶轮车的前向环境信息;Step 2: Collect the forward environmental information of the electronically guided rubber-wheeled vehicle through binocular cameras and laser radar;

步骤三:通过立体匹配算法计算电子导向胶轮车的前向环境的深度信息;Step 3: Calculate the depth information of the front environment of the electronic guided rubber-wheeled vehicle through a stereo matching algorithm;

步骤四:通过基于YOLOv7的目标检测神经网络,获取相机采集的图像中目标的视觉信息、包含目标类别及二维平面位置信息;Step 4: Obtain the visual information of the target in the image captured by the camera, including the target category and two-dimensional plane position information, through the target detection neural network based on YOLOv7;

步骤五:结合神经网络输出的目标视觉信息、环境深度信息与目标时刻信息,获取基于立体视觉的目标探测结果; Step 5: Combine the target visual information, environmental depth information and target time information output by the neural network to obtain the target detection result based on stereo vision;

步骤六:基于步骤五中获取的目标探测结果中的目标时刻信息,建立目标的运动轨迹,预测激光雷达采集时刻下目标的立体位置信息;Step 6: Based on the target time information in the target detection result obtained in step 5, establish the target's motion trajectory and predict the target's three-dimensional position information at the laser radar acquisition time;

步骤七:基于步骤六预测的目标的立体位置信息,对激光雷达采集的点云进行分割与聚类,输出目标探测结果;Step 7: Based on the three-dimensional position information of the target predicted in step 6, segment and cluster the point cloud collected by the lidar, and output the target detection result;

步骤八:基于卡尔曼滤波算法与近邻算法对激光雷达探测的目标进行跟踪,结合运营场景对融合后的目标进行修正输出。Step 8: Track the target detected by the lidar based on the Kalman filter algorithm and the nearest neighbor algorithm, and correct and output the fused target based on the operation scenario.

所述步骤一中进行相机标定的具体步骤如下:The specific steps of performing camera calibration in step 1 are as follows:

通过双目相机采集的视觉图像信息,采用张正友标定法,获取左相机内部参数ML、畸变系数DL、右相机内部参数MR、畸变系数DR以及左右相机间的外部参数MLRThe visual image information collected by the binocular camera is calibrated by Zhang Zhengyou method to obtain the left camera internal parameters ML , distortion coefficient DL , the right camera internal parameters MR , distortion coefficient DR and the external parameters MLR between the left and right cameras.

所述步骤一中进行联合标定的具体步骤如下:The specific steps of performing joint calibration in step 1 are as follows:

通过标定板在激光雷达采集的三维点云信息与左相机采集的视觉图像信息中的二维位置信息进行联合标定,获取激光雷达与左相机之间的外部参数MLCThe calibration plate is used to perform joint calibration on the three-dimensional point cloud information collected by the laser radar and the two-dimensional position information in the visual image information collected by the left camera to obtain the external parameters M LC between the laser radar and the left camera.

所述步骤三中的计算方法如下:The calculation method in step 3 is as follows:

将统一场景下左右视图对应的像素点进行匹配,得到视差图,基于Bouguet算法获取图像二维位置到空间三维位置的投影矩阵Q,计算目标深度信息。The corresponding pixel points of the left and right views in the same scene are matched to obtain the disparity map. The projection matrix Q from the two-dimensional position of the image to the three-dimensional position in space is obtained based on the Bouguet algorithm to calculate the target depth information.

所述步骤四包括:采集并制造运营环境中目标图像的数据集,将已有目标标注的数据集作为神经网络的训练集,通过训练模型获取神经网络的权重系数,再对左相机采集的图像进行预测,获取包含目标类别及描述目标平面位置的二维包围框信息的目标检测结果。The step four includes: collecting and creating a data set of target images in the operating environment, using the data set with existing target annotations as a training set for the neural network, obtaining the weight coefficient of the neural network through the training model, and then predicting the image captured by the left camera to obtain the target detection result containing the target category and the two-dimensional bounding box information describing the target plane position.

所述步骤五包括:使用步骤三中的计算方法将步骤四中获取的目标平面信息投影至三维空间,获取用三维包围框描述的立体信息的目标检测结果,再结合图像时间戳,获取基于立体视觉的目标探测结果。The step five includes: using the calculation method in step three to project the target plane information obtained in step four into three-dimensional space, obtaining the target detection result of the stereoscopic information described by the three-dimensional bounding box, and then combining the image timestamp to obtain the target detection result based on stereoscopic vision.

所述步骤七包括:The step seven comprises:

利用RANSAC算法对地面进行分割,提取非地面目标点云,基于步骤六 获取的立体位置信息进行点云聚类,获得目标点云簇,输出用三维包围框描述的立体信息的点云目标检测结果,通过双目相机获取的目标信息,引导激光雷达对目标进行精准测距;Use the RANSAC algorithm to segment the ground and extract the non-ground target point cloud. The acquired stereo position information is clustered into point clouds to obtain target point cloud clusters, and the point cloud target detection results of the stereo information described by a three-dimensional bounding box are output. The target information obtained by the binocular camera guides the laser radar to accurately measure the distance to the target;

所述点云聚类的步骤如下:The steps of point cloud clustering are as follows:

(1)激光雷达通过获取双目相机探测出的目标类别与历史轨迹信息,对目标的空间位置进行建模预测,辅助划定探测范围;(1) The laser radar obtains the target category and historical trajectory information detected by the binocular camera, models and predicts the spatial position of the target, and assists in defining the detection range;

(2)选择探测范围内点云集中的任意一个点M,并根据探测范围自适应设定聚类阈值,基于KDTree的数据结构搜索M近邻的k个点,筛选与点M的距离小于聚类阈值的点并存放于集合Q中,若Q中的点增加,则在Q中选取点M以外的任意点用以更新点M,持续搜索直至集合Q中元素不再增加;(2) Select any point M in the point cloud within the detection range, and adaptively set the clustering threshold according to the detection range. Search for k points of M's neighbors based on the KDTree data structure, select points whose distance to point M is less than the clustering threshold and store them in set Q. If the number of points in Q increases, select any point other than point M in Q to update point M. Continue searching until the number of elements in set Q stops increasing.

(3)对双目相机获取的图像中所有的检测物体重复步骤(1)和步骤(2),再利用聚类检测算法完成所有目标的检测,输出目标类别、点云、三维矩形框以及时间戳。(3) Repeat steps (1) and (2) for all the detection objects in the image acquired by the binocular camera, and then use the clustering detection algorithm to complete the detection of all targets, and output the target category, point cloud, three-dimensional rectangular box and timestamp.

在对激光雷达采集的点云进行聚类前,采用运动畸变补偿的方式消除点云内部的运动畸变,具体如下:Before clustering the point cloud collected by the LiDAR, motion distortion compensation is used to eliminate the motion distortion inside the point cloud, as follows:

采用IMU进行激光雷达的运动畸变补偿,在接收激光雷达扫描数据的同时,将该数据与IMU采集的车辆姿态角数据保存在同一循环队列中,处理器根据激光雷达的采样时间间隔与数据时间戳计算得到每个激光点的采用时间,在IMU的数据队列中寻找时间相邻的两帧数据,利用球面线性插值的方式,将同一测量时刻下的激光点与车辆运动姿态配对,从而能够将一帧激光雷达扫描数据的每一个激光点坐标转换到同一时刻下电子导向胶轮车的对应位置坐标系下,以使得将激光雷达在一段时间内收集的所有点云统一至一个时刻下,实现对激光雷达的运动畸变补偿。IMU is used to compensate for the motion distortion of the lidar. While receiving the lidar scanning data, the data and the vehicle attitude angle data collected by the IMU are saved in the same circular queue. The processor calculates the adoption time of each laser point based on the sampling time interval of the lidar and the data timestamp, and searches for two temporally adjacent frames of data in the IMU data queue. The laser points at the same measurement moment are paired with the vehicle motion attitude by using spherical linear interpolation. This allows the coordinates of each laser point of a frame of lidar scanning data to be converted to the corresponding position coordinate system of the electronically guided rubber-wheeled vehicle at the same moment, so that all point clouds collected by the lidar within a period of time are unified to one moment, thereby realizing the motion distortion compensation of the lidar.

所述步骤八中基于卡尔曼滤波算法与近邻算法实现对目标的跟踪的方法如下:The method for tracking the target based on the Kalman filter algorithm and the nearest neighbor algorithm in step eight is as follows:

基于目标运动模型构建并初始化卡尔曼滤波器组及配置相关参数,计算目 标点云的质心位置。基于卡尔曼滤波器组通过当前时间戳下目标的质心位置预测下一帧点云时刻目标的质心位置。在下一帧点云中基于近邻算法搜索互相匹配的真实目标与预测目标组合,实现帧间的目标关联,即获取到了相邻两帧同一目标的空间位置。利用目标在下一帧点云中的空间位置与时间戳信息对卡尔曼滤波器组的目标状态进行更新,再对目标状态进行预测。循环上述过程,可以实现对目标的跟踪。Based on the target motion model, the Kalman filter group is constructed and initialized, and the relevant parameters are configured to calculate the target. The center of mass position of the point cloud is determined by the Kalman filter group. The center of mass position of the target at the current timestamp is used to predict the center of mass position of the target at the next frame of the point cloud. In the next frame of the point cloud, the nearest neighbor algorithm is used to search for a matching combination of real targets and predicted targets to achieve target association between frames, that is, to obtain the spatial position of the same target in two adjacent frames. The target state of the Kalman filter group is updated using the spatial position and timestamp information of the target in the next frame of the point cloud, and then the target state is predicted. The above process is repeated to track the target.

与现有技术相比,本发明的有益效果是:本发明采用双目视觉引导的融合方法,可以利用视觉目标信息建立其运动轨迹,进而预测其在三维空间中的位置;采用数据层的融合方法,利用视觉的预测结果以辅助激光雷达划定探测范围,大幅度降低运算数据量,对计算平台性能要求较低;Compared with the prior art, the present invention has the following beneficial effects: the present invention adopts a binocular vision-guided fusion method, which can use visual target information to establish its motion trajectory, and then predict its position in three-dimensional space; adopts a data layer fusion method, uses visual prediction results to assist laser radar in defining the detection range, greatly reduces the amount of computational data, and has low requirements on computing platform performance;

基于双目视觉获取的目标探测结果,可以准确获取目标深度,扩展目标检测信息及增强目标检测的鲁棒性。The target detection results obtained based on binocular vision can accurately obtain the target depth, expand the target detection information and enhance the robustness of target detection.

附图说明BRIEF DESCRIPTION OF THE DRAWINGS

图1为本发明的传感器融合方法的原理图;FIG1 is a schematic diagram of a sensor fusion method of the present invention;

图2为本发明的传感器融合方法的流程图;FIG2 is a flow chart of a sensor fusion method of the present invention;

图3为激光雷达点云聚类的流程图;Figure 3 is a flow chart of LiDAR point cloud clustering;

图4为点云聚类后的效果图;Figure 4 is the effect diagram after point cloud clustering;

图5为目标跟踪示意图。Figure 5 is a schematic diagram of target tracking.

具体实施方式Detailed ways

下面结合附图,对本发明的一个具体实施方式进行详细描述,但应当理解本发明的保护范围并不受具体实施方式的限制。A specific implementation of the present invention is described in detail below in conjunction with the accompanying drawings, but it should be understood that the protection scope of the present invention is not limited by the specific implementation.

如图1至图5所示,本发明实施例提供的一种基于双目相机引导的传感器融合方法,包括如下步骤:As shown in FIG. 1 to FIG. 5 , a sensor fusion method based on binocular camera guidance provided by an embodiment of the present invention includes the following steps:

步骤一:基于双目相机采集的视觉图像信息以及激光雷达采集的三维点云信息,对组成双目相机的左相机和右相机进行相机标定,对激光雷达与左相机进行联合标定; Step 1: Based on the visual image information collected by the binocular camera and the 3D point cloud information collected by the lidar, calibrate the left and right cameras that make up the binocular camera, and jointly calibrate the lidar and the left camera;

本发明提出的融合方法,首先基于双目相机进行目标检测与网络跟踪,双目相机一般由左右两个水平放置的相机构成,两个相机分别对目标进行成像,由于相机间存在一定长度的基线,目标的成像位置会有所不同,可以利用成像位置的差异及基线长度计算出目标与相机间的位置关系;The fusion method proposed in the present invention first performs target detection and network tracking based on a binocular camera. A binocular camera is generally composed of two horizontally placed left and right cameras. The two cameras image the target respectively. Since there is a baseline of a certain length between the cameras, the imaging position of the target will be different. The positional relationship between the target and the camera can be calculated using the difference in imaging position and the baseline length.

步骤一中进行相机标定的具体步骤如下:The specific steps for camera calibration in step 1 are as follows:

通过双目相机采集的视觉图像信息,采用张正友标定法,获取左相机内部参数ML、畸变系数DL、右相机内部参数MR、畸变系数DR以及左右相机间的外部参数MLRThe visual image information collected by the binocular camera is calibrated by Zhang Zhengyou method to obtain the left camera internal parameters ML , distortion coefficient DL , the right camera internal parameters MR , distortion coefficient DR and the external parameters MLR between the left and right cameras;

步骤一中进行联合标定的具体步骤如下:The specific steps for joint calibration in step 1 are as follows:

通过标定板在激光雷达采集的三维点云信息与左相机采集的视觉图像信息中的二维位置信息进行联合标定,获取激光雷达与左相机之间的外部参数MLC,进而将各自坐标系下的测量值转换到列车坐标系中;The calibration plate is used to jointly calibrate the three-dimensional point cloud information collected by the laser radar and the two-dimensional position information in the visual image information collected by the left camera to obtain the external parameters M LC between the laser radar and the left camera, and then the measurement values in each coordinate system are converted to the train coordinate system;

双目校正根据相机标定时获取的相对位置关系与左右相机各自的内部参数据对两个相机拍摄的视图进行去畸变和行对准,使左右视图成像原点坐标一致、两相机的光轴平行、对极线对齐与左右相机的成像平面共面;Binocular correction dedistorts and aligns the views captured by the two cameras according to the relative position relationship obtained during camera calibration and the internal parameter data of the left and right cameras, so that the coordinates of the imaging origins of the left and right views are consistent, the optical axes of the two cameras are parallel, and the epipolar lines are aligned and coplanar with the imaging planes of the left and right cameras;

步骤二:通过双目相机与激光雷达采集电子导向胶轮车的前向环境信息;Step 2: Collect the forward environmental information of the electronically guided rubber-wheeled vehicle through binocular cameras and laser radar;

步骤二是在保持双目相机与激光雷达位置不变的的基础上进行的;Step 2 is performed on the basis of keeping the positions of the binocular camera and the laser radar unchanged;

步骤三:通过立体匹配算法计算电子导向胶轮车的前向环境的深度信息;Step 3: Calculate the depth information of the front environment of the electronic guided rubber-wheeled vehicle through a stereo matching algorithm;

步骤三中的计算方法如下:将统一场景下左右视图对应的像素点进行匹配,得到视差图,基于Bouguet算法获取图像二维位置到空间三维位置的投影矩阵Q,计算目标深度信息;The calculation method in step 3 is as follows: match the pixel points corresponding to the left and right views in the same scene to obtain the disparity map, obtain the projection matrix Q from the two-dimensional position of the image to the three-dimensional position of the space based on the Bouguet algorithm, and calculate the target depth information;

步骤四:通过基于YOLOv7的目标检测神经网络,获取相机采集的图像中目标的视觉信息、包含目标类别及二维平面位置信息;Step 4: Obtain the visual information of the target in the image captured by the camera, including the target category and two-dimensional plane position information, through the target detection neural network based on YOLOv7;

上述步骤四具体为:采集并制造运营环境中目标图像的数据集,将已有目标标注的数据集作为YOLOv7神经网络的训练集,通过训练模型获取神经网络的权重系数,再对左相机采集的图像进行预测,获取包含目标类别及描述目标 平面位置的二维包围框信息的目标检测结果,本发明利用基于深度学习的目标检测算法对单张图像进行目标检测,可以识别出行人、车辆等交通参与者,并对其进行分类,获取基于立体视觉的目标探测结果;The above step 4 is as follows: collect and create a dataset of target images in the operating environment, use the dataset with existing target annotations as the training set of the YOLOv7 neural network, obtain the weight coefficient of the neural network through the training model, and then predict the images captured by the left camera to obtain the target category and description target. The target detection result of the two-dimensional bounding box information of the plane position, the present invention uses the target detection algorithm based on deep learning to detect the target in a single image, which can identify traffic participants such as pedestrians and vehicles, and classify them to obtain the target detection result based on stereo vision;

另外,本发明所采用的YOLOv7是在YOLOv5的基础上通过对内部多项架构的改革提高了检测的速度与准确性;YOLOv7主要使用MP结构与ELAN,它可以通过控制最短的梯度路径和更深的网络结构更高效地进行深度学习;YOLOv7的损失函数分为目标置信度损失、坐标损失与分类损失三个主要部分,主要在匹配策略过程中进行数据候选,保留损失最小的数据,实现高精度的目标检测;In addition, YOLOv7 adopted in the present invention improves the detection speed and accuracy by reforming multiple internal architectures based on YOLOv5; YOLOv7 mainly uses MP structure and ELAN, which can perform deep learning more efficiently by controlling the shortest gradient path and deeper network structure; the loss function of YOLOv7 is divided into three main parts: target confidence loss, coordinate loss and classification loss, which mainly performs data candidate matching in the matching strategy process, retains the data with the smallest loss, and realizes high-precision target detection;

步骤五:结合神经网络输出的目标视觉信息、环境深度信息与目标时刻信息,获取基于立体视觉的目标探测结果;Step 5: Combine the target visual information, environmental depth information and target time information output by the neural network to obtain the target detection result based on stereo vision;

上述步骤五使用步骤三中的计算方法将步骤四中获取的目标平面信息投影至三维空间,获取用三维包围框描述的立体信息的目标检测结果,再结合图像时间戳,获取基于立体视觉的目标探测结果The above step 5 uses the calculation method in step 3 to project the target plane information obtained in step 4 into three-dimensional space, obtain the target detection result of the three-dimensional information described by the three-dimensional bounding box, and then combine it with the image timestamp to obtain the target detection result based on stereoscopic vision.

步骤六:基于步骤五中获取的目标探测结果中的目标时刻信息,建立目标的运动轨迹,预测激光雷达采集时刻下目标的立体位置信息;Step 6: Based on the target time information in the target detection result obtained in step 5, establish the target's motion trajectory and predict the target's three-dimensional position information at the laser radar acquisition time;

步骤七:基于步骤六预测的目标的立体位置信息,对激光雷达采集的点云进行分割与聚类,输出目标探测结果;Step 7: Based on the three-dimensional position information of the target predicted in step 6, segment and cluster the point cloud collected by the lidar, and output the target detection result;

上述步骤七具体为:结合电子导向胶轮车的运营环境特点,利用RANSAC算法对地面进行分割,提取非地面目标点云,基于步骤六获取的立体位置信息进行点云聚类,获得目标点云簇,输出用三维包围框描述的立体信息的点云目标检测结果,通过双目相机获取的目标信息,引导激光雷达对目标进行精准测距;一般应用双目相机获取目标的类别与位置,需要考虑图像存在的采样延时与目标检测算法的处理时间,目前激光雷达点云聚类主要采用欧氏聚类方法进行聚类,这种聚类方法认为在激光雷达采集的所有的点云数据中,相同的物体的点云中任意两点的距离会小于一定数值,因此可以将距离小于一定数值的点 合成一类,最终完成目标聚类。这种聚类方式精度较高,但是它的聚类速度并不占优势,虽然这种方法采用KD-Tree加快了数据的查找速度,但是在高速变化的场景中仍然显得不足;本发明开创性地采用了双目相机引导激光雷达进行目标聚类的方式,不仅能够提高聚类的速度,对最终识别结果的精准性与鲁棒性也有较好的结果,点云聚类的步骤如下:The above step seven is specifically as follows: In combination with the operating environment characteristics of the electronic guided rubber-wheeled vehicle, the RANSAC algorithm is used to segment the ground, and the non-ground target point cloud is extracted. The point cloud is clustered based on the stereo position information obtained in step six to obtain the target point cloud cluster, and the point cloud target detection result of the stereo information described by the three-dimensional bounding box is output. The target information obtained by the binocular camera is used to guide the lidar to accurately measure the distance to the target; generally, the binocular camera is used to obtain the category and position of the target, and the sampling delay of the image and the processing time of the target detection algorithm need to be considered. At present, the lidar point cloud clustering mainly adopts the Euclidean clustering method for clustering. This clustering method believes that in all the point cloud data collected by the lidar, the distance between any two points in the point cloud of the same object will be less than a certain value. Therefore, points with a distance less than a certain value can be clustered. This clustering method has high accuracy, but its clustering speed is not advantageous. Although this method uses KD-Tree to speed up the data search speed, it is still insufficient in high-speed changing scenes. The present invention innovatively uses binocular cameras to guide laser radars to cluster targets, which can not only improve the clustering speed, but also have better results on the accuracy and robustness of the final recognition results. The steps of point cloud clustering are as follows:

(1)激光雷达通过获取双目相机探测出的目标类别与历史轨迹信息,对目标的空间位置进行建模预测,辅助划定探测范围;(1) The laser radar obtains the target category and historical trajectory information detected by the binocular camera, models and predicts the spatial position of the target, and assists in defining the detection range;

(2)选择探测范围内点云集中的任意一个点M,并根据探测范围自适应设定聚类阈值,基于KDTree的数据结构搜索M近邻的k个点,筛选与点M的距离小于聚类阈值的点并存放于集合Q中,若Q中的点增加,则在Q中选取点M以外的任意点用以更新点M,持续搜索直至集合Q中元素不再增加;(2) Select any point M in the point cloud within the detection range, and adaptively set the clustering threshold according to the detection range. Search for k points of M's neighbors based on the KDTree data structure, select points whose distance to point M is less than the clustering threshold and store them in set Q. If the number of points in Q increases, select any point other than point M in Q to update point M. Continue searching until the number of elements in set Q stops increasing.

(3)对双目相机获取的图像中所有的检测物体重复步骤(1)和步骤(2),再利用聚类检测算法完成所有目标的检测,输出目标类别、点云、三维矩形框以及时间戳。(3) Repeat steps (1) and (2) for all the detection objects in the image acquired by the binocular camera, and then use the clustering detection algorithm to complete the detection of all targets, and output the target category, point cloud, three-dimensional rectangular box and timestamp.

另外在激光雷达跟随胶轮车辆运动的过程中,由于电子导向胶轮车在高速运动,不同的时刻胶轮车辆的基准位姿不相同,导致激光点采集的激光点云会出现运动畸变,同时,如果激光雷达扫描的频率较低,激光点的数据并非瞬间获得,也会导致激光点云出现运动畸变,因此在对激光雷达采集的点云进行聚类前,采用运动畸变补偿的方式消除点云内部的运动畸变,具体如下:In addition, when the laser radar follows the movement of the rubber-wheeled vehicle, the electronically guided rubber-wheeled vehicle moves at high speed, and the reference posture of the rubber-wheeled vehicle is different at different times, resulting in motion distortion of the laser point cloud collected by the laser point. At the same time, if the laser radar scanning frequency is low, the data of the laser point is not obtained instantaneously, which will also cause motion distortion of the laser point cloud. Therefore, before clustering the point cloud collected by the laser radar, motion distortion compensation is used to eliminate the motion distortion inside the point cloud, as follows:

采用IMU进行激光雷达的运动畸变补偿,在接收激光雷达扫描数据的同时,将该数据与IMU采集的车辆姿态角数据保存在同一循环队列中,处理器根据激光雷达的采样时间间隔与数据时间戳计算得到每个激光点的采用时间,在IMU的数据队列中寻找时间相邻的两帧数据,利用球面线性插值的方式,将同一测量时刻下的激光点与车辆运动姿态配对,从而能够将一帧激光雷达扫描数据的每一个激光点坐标转换到同一时刻下电子导向胶轮车的对应位置坐标系下,以使得将激光雷达在一段时间内收集的所有点云统一至一个时刻下, 实现对激光雷达的运动畸变补偿。IMU is used to compensate for the motion distortion of the laser radar. While receiving the laser radar scanning data, the data and the vehicle attitude angle data collected by the IMU are saved in the same circular queue. The processor calculates the adoption time of each laser point according to the sampling time interval of the laser radar and the data timestamp, and searches for two frames of data adjacent in time in the IMU data queue. The laser points at the same measurement moment are paired with the vehicle motion posture by using spherical linear interpolation. This allows the coordinates of each laser point of a frame of laser radar scanning data to be converted to the corresponding position coordinate system of the electronic guided rubber-wheeled vehicle at the same moment, so that all point clouds collected by the laser radar within a period of time are unified to one moment. Realize motion distortion compensation for lidar.

步骤八:基于卡尔曼滤波算法与近邻算法对激光雷达探测的目标进行跟踪,结合运营场景对融合后的目标进行修正输出;Step 8: Track the target detected by the lidar based on the Kalman filter algorithm and the nearest neighbor algorithm, and correct and output the fused target based on the operation scenario;

上述步骤八中基于卡尔曼滤波算法与近邻算法实现对目标的跟踪的方法如下:The method for tracking the target based on the Kalman filter algorithm and the nearest neighbor algorithm in the above step eight is as follows:

基于目标运动模型构建并初始化卡尔曼滤波器组及配置相关参数,计算目标点云的质心位置,基于卡尔曼滤波器组通过当前时间戳下目标的质心位置预测下一帧点云时刻目标的质心位置,在下一帧点云中基于近邻算法搜索互相匹配的真实目标与预测目标组合,实现帧间的目标关联,即获取到了相邻两帧同一目标的空间位置,利用目标在下一帧点云中的空间位置与时间戳信息对卡尔曼滤波器组的目标状态进行更新,再对目标状态进行预测,循环上述过程,可以实现对目标的跟踪。本发明提出一种由双目相机引导的传感器融合框架,提供了一种由双目相机引导的传感器数据层的融合方法,兼顾了目标探测性能与工程应用,实现了不同功能的传感器间优势互补,提升了目标探测系统的智能化程度,基于深度学习的视觉目标检测系统,相比3D目标检测系统,发展成熟,对计算平台的性能要求更低。Based on the target motion model, a Kalman filter group is constructed and initialized, and related parameters are configured. The centroid position of the target point cloud is calculated. Based on the Kalman filter group, the centroid position of the target at the next frame point cloud is predicted by the centroid position of the target at the current timestamp. In the next frame point cloud, a matching combination of real targets and predicted targets is searched based on the nearest neighbor algorithm to achieve target association between frames, that is, the spatial position of the same target in two adjacent frames is obtained. The target state of the Kalman filter group is updated using the spatial position and timestamp information of the target in the next frame point cloud, and then the target state is predicted. The above process is repeated to achieve target tracking. The present invention proposes a sensor fusion framework guided by a binocular camera, and provides a fusion method of a sensor data layer guided by a binocular camera, which takes into account target detection performance and engineering applications, realizes the complementary advantages between sensors with different functions, and improves the intelligence of the target detection system. Compared with the 3D target detection system, the visual target detection system based on deep learning is mature and has lower performance requirements on the computing platform.

对于本领域技术人员而言,显然本发明不限于上述示范性实施例的细节,而且在不背离本发明的精神和基本特征的情况下,能够以其他的具体形式实现本发明。因此,无论从哪一点来看,均应将实施例看作是示范性的,而且是非限制性的,本发明的范围由所附权利要求而不是上述说明限定,因此旨在将落在权利要求的等同要件的含义和范围内的所有变化囊括在本发明内。不应将权利要求中的任何附图标记视为限制所涉及的权利要求。It will be apparent to those skilled in the art that the invention is not limited to the details of the exemplary embodiments described above and that the invention can be implemented in other specific forms without departing from the spirit and essential features of the invention. Therefore, the embodiments should be considered exemplary and non-limiting in all respects, and the scope of the invention is defined by the appended claims rather than the foregoing description, and it is intended that all variations falling within the meaning and scope of the equivalent elements of the claims be included in the invention. Any reference numeral in a claim should not be considered as limiting the claim to which it relates.

此外,应当理解,虽然本说明书按照实施方式加以描述,但并非每个实施方式仅包含一个独立的技术方案,说明书的这种叙述方式仅仅为清楚起见,本领域技术人员应当将说明书作为一个整体,各实施例中的技术方案也可以经适当组合,形成本领域技术人员可以理解的其他实施方式。 In addition, it should be understood that although the present specification is described according to implementation modes, not every implementation mode contains only one independent technical solution. This description of the specification is only for the sake of clarity. Those skilled in the art should regard the specification as a whole. The technical solutions in each embodiment may also be appropriately combined to form other implementation modes that can be understood by those skilled in the art.

Claims (9)

一种基于双目相机引导的传感器融合方法,其特征在于,包括如下步骤:A sensor fusion method based on binocular camera guidance, characterized in that it includes the following steps: 步骤一:基于双目相机采集的视觉图像信息以及激光雷达采集的三维点云信息,对组成双目相机的左相机和右相机进行相机标定,对激光雷达与左相机进行联合标定;Step 1: Based on the visual image information collected by the binocular camera and the 3D point cloud information collected by the lidar, calibrate the left and right cameras that make up the binocular camera, and jointly calibrate the lidar and the left camera; 步骤二:通过双目相机与激光雷达采集电子导向胶轮车的前向环境信息;Step 2: Collect the forward environmental information of the electronically guided rubber-wheeled vehicle through binocular cameras and laser radar; 步骤三:通过立体匹配算法计算电子导向胶轮车的前向环境的深度信息;Step 3: Calculate the depth information of the front environment of the electronic guided rubber-wheeled vehicle through a stereo matching algorithm; 步骤四:通过基于YOLOv7的目标检测神经网络,获取相机采集的图像中目标的视觉信息、包含目标类别及二维平面位置信息;Step 4: Obtain the visual information of the target in the image captured by the camera, including the target category and two-dimensional plane position information, through the target detection neural network based on YOLOv7; 步骤五:结合神经网络输出的目标视觉信息、环境深度信息与目标时刻信息,获取基于立体视觉的目标探测结果;Step 5: Combine the target visual information, environmental depth information and target time information output by the neural network to obtain the target detection result based on stereo vision; 步骤六:基于步骤五中获取的目标探测结果中的目标时刻信息,建立目标的运动轨迹,预测激光雷达采集时刻下目标的立体位置信息;Step 6: Based on the target time information in the target detection result obtained in step 5, establish the target's motion trajectory and predict the target's three-dimensional position information at the laser radar acquisition time; 步骤七:基于步骤六预测的目标的立体位置信息,对激光雷达采集的点云进行分割与聚类,输出目标探测结果;Step 7: Based on the three-dimensional position information of the target predicted in step 6, segment and cluster the point cloud collected by the lidar, and output the target detection result; 步骤八:基于卡尔曼滤波算法与近邻算法对激光雷达探测的目标进行跟踪,结合运营场景对融合后的目标进行修正输出。Step 8: Track the target detected by the lidar based on the Kalman filter algorithm and the nearest neighbor algorithm, and correct and output the fused target based on the operation scenario. 如权利要求1所述的一种基于双目相机引导的传感器融合方法,其特征在于,所述步骤一中进行相机标定的具体步骤如下:The sensor fusion method based on binocular camera guidance according to claim 1 is characterized in that the specific steps of performing camera calibration in step 1 are as follows: 通过双目相机采集的视觉图像信息,采用张正友标定法,获取左相机内部参数ML、畸变系数DL、右相机内部参数MR、畸变系数DR以及左右相机间的外部参数MLRThe visual image information collected by the binocular camera is calibrated by Zhang Zhengyou method to obtain the left camera internal parameters ML , distortion coefficient DL , the right camera internal parameters MR , distortion coefficient DR and the external parameters MLR between the left and right cameras. 如权利要求1或2所述的一种基于双目相机引导的传感器融合方法,其特征在于,所述步骤一中进行联合标定的具体步骤如下:The sensor fusion method based on binocular camera guidance according to claim 1 or 2 is characterized in that the specific steps of performing joint calibration in step 1 are as follows: 通过标定板在激光雷达采集的三维点云信息与左相机采集的视觉图像信 息中的二维位置信息进行联合标定,获取激光雷达与左相机之间的外部参数MLCThe 3D point cloud information collected by the laser radar and the visual image information collected by the left camera are compared by the calibration plate. The two-dimensional position information in the information is jointly calibrated to obtain the external parameters M LC between the laser radar and the left camera. 如权利要求1所述的一种基于双目相机引导的传感器融合方法,其特征在于,所述步骤三中的计算方法如下:The sensor fusion method based on binocular camera guidance according to claim 1, characterized in that the calculation method in step three is as follows: 将统一场景下左右视图对应的像素点进行匹配,得到视差图,基于Bouguet算法获取图像二维位置到空间三维位置的投影矩阵Q,计算目标深度信息。The corresponding pixel points of the left and right views in the same scene are matched to obtain the disparity map. The projection matrix Q from the two-dimensional position of the image to the three-dimensional position in space is obtained based on the Bouguet algorithm to calculate the target depth information. 如权利要求1所述的一种基于双目相机引导的传感器融合方法,其特征在于,所述步骤四包括:采集并制造运营环境中目标图像的数据集,将已有目标标注的数据集作为神经网络的训练集,通过训练模型获取神经网络的权重系数,再对左相机采集的图像进行预测,获取包含目标类别及描述目标平面位置的二维包围框信息的目标检测结果。A sensor fusion method based on binocular camera guidance as described in claim 1, characterized in that step four includes: collecting and creating a data set of target images in the operating environment, using the data set with existing target annotations as a training set for the neural network, obtaining the weight coefficient of the neural network through the training model, and then predicting the image captured by the left camera to obtain a target detection result containing the target category and the two-dimensional bounding box information describing the target plane position. 如权利要求1所述的一种基于双目相机引导的传感器融合方法,其特征在于,所述步骤五包括:使用步骤三中的计算方法将步骤四中获取的目标平面信息投影至三维空间,获取用三维包围框描述的立体信息的目标检测结果,再结合图像时间戳,获取基于立体视觉的目标探测结果。A sensor fusion method based on binocular camera guidance as described in claim 1, characterized in that the step five includes: using the calculation method in step three to project the target plane information obtained in step four into three-dimensional space, obtaining the target detection result of the stereo information described by the three-dimensional bounding box, and then combining the image timestamp to obtain the target detection result based on stereo vision. 如权利要求1所述的一种基于双目相机引导的传感器融合方法,其特征在于,所述步骤七包括:The sensor fusion method based on binocular camera guidance according to claim 1, characterized in that the step seven comprises: 利用RANSAC算法对地面进行分割,提取非地面目标点云,基于步骤六获取的立体位置信息进行点云聚类,获得目标点云簇,输出用三维包围框描述的立体信息的点云目标检测结果,通过双目相机获取的目标信息,引导激光雷达对目标进行精准测距;The RANSAC algorithm is used to segment the ground and extract the non-ground target point cloud. The point cloud is clustered based on the stereo position information obtained in step 6 to obtain the target point cloud cluster. The point cloud target detection result of the stereo information described by the three-dimensional bounding box is output. The target information obtained by the binocular camera is used to guide the laser radar to accurately measure the distance to the target. 所述点云聚类的步骤如下:The steps of point cloud clustering are as follows: (1)激光雷达通过获取双目相机探测出的目标类别与历史轨迹信息,对目标的空间位置进行建模预测,辅助划定探测范围;(1) The laser radar obtains the target category and historical trajectory information detected by the binocular camera, models and predicts the spatial position of the target, and assists in defining the detection range; (2)选择探测范围内点云集中的任意一个点M,并根据探测范围自适应设定聚类阈值,基于KDTree的数据结构搜索M近邻的k个点,筛选与点M 的距离小于聚类阈值的点并存放与集合Q中,若Q中的点增加,则在Q中选取点M以外的任意点用以更新点M,持续搜索直至集合Q中元素不再增加;(2) Select any point M in the point cloud within the detection range, and adaptively set the clustering threshold according to the detection range. Search for k points of M’s neighbors based on the KDTree data structure, and filter out the points that are close to point M. The points whose distance is less than the clustering threshold are stored in the set Q. If the number of points in Q increases, any point other than point M is selected in Q to update point M. The search continues until the number of elements in the set Q stops increasing. (3)对双目相机获取的图像中所有的检测物体重复步骤(1)和步骤(2),再利用聚类检测算法完成所有目标的检测,输出目标类别、点云、三维矩形框以及时间戳。(3) Repeat steps (1) and (2) for all the detection objects in the image acquired by the binocular camera, and then use the clustering detection algorithm to complete the detection of all targets, and output the target category, point cloud, three-dimensional rectangular box and timestamp. 如权利要求7所述的一种基于双目相机引导的传感器融合方法,其特征在于,在对激光雷达采集的点云进行聚类前,采用运动畸变补偿的方式消除点云内部的运动畸变,具体如下:The sensor fusion method based on binocular camera guidance as claimed in claim 7 is characterized in that before clustering the point cloud collected by the laser radar, the motion distortion inside the point cloud is eliminated by using motion distortion compensation, which is specifically as follows: 采用IMU进行激光雷达的运动畸变补偿,在接收激光雷达扫描数据的同时,将该数据与IMU采集的车辆姿态角数据保存在同一循环队列中,处理器根据激光雷达的采样时间间隔与数据时间戳计算得到每个激光点的采用时间,在IMU的数据队列中寻找时间相邻的两帧数据,利用球面线性插值的方式,将同一测量时刻下的激光点与车辆运动姿态配对,从而能够将一帧激光雷达扫描数据的每一个激光点坐标转换到同一时刻下电子导向胶轮车的对应位置坐标系下,以使得将激光雷达在一段时间内收集的所有点云统一至一个时刻下,实现对激光雷达的运动畸变补偿。IMU is used to compensate for the motion distortion of the lidar. While receiving the lidar scanning data, the data and the vehicle attitude angle data collected by the IMU are saved in the same circular queue. The processor calculates the adoption time of each laser point based on the sampling time interval of the lidar and the data timestamp, and searches for two temporally adjacent frames of data in the IMU data queue. The laser points at the same measurement moment are paired with the vehicle motion attitude by using spherical linear interpolation. This allows the coordinates of each laser point of a frame of lidar scanning data to be converted to the corresponding position coordinate system of the electronically guided rubber-wheeled vehicle at the same moment, so that all point clouds collected by the lidar within a period of time are unified to one moment, thereby realizing the motion distortion compensation of the lidar. 如权利要求1所述的一种基于双目相机引导的传感器融合方法,其特征在于,所述步骤八中基于卡尔曼滤波算法与近邻算法实现对目标的跟踪的方法如下:The sensor fusion method based on binocular camera guidance according to claim 1 is characterized in that the method for tracking the target based on the Kalman filter algorithm and the nearest neighbor algorithm in step eight is as follows: 基于目标运动模型构建并初始化卡尔曼滤波器组及配置相关参数,计算目标点云的质心位置,基于卡尔曼滤波器组通过当前时间戳下目标的质心位置预测下一帧点云时刻目标的质心位置,在下一帧点云中基于近邻算法搜索互相匹配的真实目标与预测目标组合,实现帧间的目标关联,即获取到了相邻两帧同一目标的空间位置,利用目标在下一帧点云中的空间位置与时间戳信息对卡尔曼滤波器组的目标状态进行更新,再对目标状态进行预测,循环上述过程,可以实现对目标的跟踪。 Based on the target motion model, a Kalman filter group is constructed and initialized, and related parameters are configured. The center of mass position of the target point cloud is calculated. The center of mass position of the target at the next frame point cloud is predicted based on the Kalman filter group through the center of mass position of the target at the current timestamp. In the next frame point cloud, a matching combination of real target and predicted target is searched based on the nearest neighbor algorithm to achieve target association between frames, that is, the spatial position of the same target in two adjacent frames is obtained. The target state of the Kalman filter group is updated using the spatial position and timestamp information of the target in the next frame point cloud, and then the target state is predicted. The above process is repeated to achieve target tracking.
PCT/CN2023/123988 2022-11-29 2023-10-11 Sensor fusion method based on binocular camera guidance Ceased WO2024114119A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202211507517.0 2022-11-29
CN202211507517 2022-11-29

Publications (1)

Publication Number Publication Date
WO2024114119A1 true WO2024114119A1 (en) 2024-06-06

Family

ID=86697442

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2023/123988 Ceased WO2024114119A1 (en) 2022-11-29 2023-10-11 Sensor fusion method based on binocular camera guidance

Country Status (2)

Country Link
CN (1) CN115937810A (en)
WO (1) WO2024114119A1 (en)

Cited By (35)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN118334016A (en) * 2024-06-12 2024-07-12 青岛交通工程监理咨询有限公司 Highway engineering construction quality detection method
CN118408553A (en) * 2024-07-01 2024-07-30 西南科技大学 A UAV navigation method based on three-dimensional reconstruction and recognition of the environment
CN118584523A (en) * 2024-08-01 2024-09-03 浙江易视通联信息科技有限公司 Rescue positioning data fusion analysis method and system based on time series update
CN118603052A (en) * 2024-08-08 2024-09-06 陕西炬烽建筑劳务有限公司 A terrain survey device for civil engineering
CN118730129A (en) * 2024-07-11 2024-10-01 奇瑞汽车股份有限公司 A multi-sensor multi-modal tightly coupled fusion mapping and positioning method and system
CN118778064A (en) * 2024-09-04 2024-10-15 中科蓝卓(北京)信息科技有限公司 Method and system for monitoring dangerous trees along railways based on optical SAR fusion and binocular depth estimation technology
CN118816935A (en) * 2024-09-18 2024-10-22 青岛慧拓智能机器有限公司 A multi-sensor fusion odometer generation method and system based on surface element fitting
CN118823275A (en) * 2024-09-18 2024-10-22 湖南理工职业技术学院 A method for modeling and positioning an augmented reality environment for a mobile robot
CN118859937A (en) * 2024-07-01 2024-10-29 吉首大学 A real-time position control method and system for multiple intelligent vehicles
CN118869899A (en) * 2024-07-15 2024-10-29 浙江大学 A first-person perspective data acquisition method and system based on binocular event camera and depth camera
CN119024268A (en) * 2024-10-25 2024-11-26 西北工业大学 High-precision signal time difference direction finding positioning method and system combined with dual-path vision pod spatial positioning
CN119063732A (en) * 2024-08-26 2024-12-03 上海卫星工程研究所 Information processing method, system and medium for integrating on-board real-time target pre-detection and intelligent precision detection
CN119090839A (en) * 2024-08-28 2024-12-06 北京建筑大学 An intelligent detection method for abnormal rail fasteners based on multi-source information fusion
CN119274030A (en) * 2024-12-09 2025-01-07 山东高速集团有限公司创新研究院 A highway intelligent inspection method and equipment based on multi-dimensional visual fusion
CN119360330A (en) * 2024-09-29 2025-01-24 武汉科技大学 An unpaved road identification and classification system
CN119395718A (en) * 2024-10-14 2025-02-07 东北大学 A multi-sensor fusion simultaneous estimation method for the position and pose of a moving object and multiple targets in a metaverse environment
CN119477980A (en) * 2024-11-04 2025-02-18 中国人民解放军国防科技大学 An interactive multi-model infrared time-sensitive target tracking method based on image feature fusion
CN119495099A (en) * 2024-11-15 2025-02-21 哈尔滨工业大学 A 3D data annotation method for embodied intelligent robots
CN119540536A (en) * 2024-11-22 2025-02-28 北京航空航天大学 A vehicle target detection method based on the fusion of binocular camera and lidar
CN119580162A (en) * 2024-11-28 2025-03-07 中国矿业大学 A real-time detection method for deviation of mining conveyor belts based on ELAN-CP
CN119600105A (en) * 2024-11-20 2025-03-11 上海联适导航技术股份有限公司 A fast and accurate real-time moving target position estimation method and its application
CN119741384A (en) * 2024-12-26 2025-04-01 西安交通大学 A method and system for calibrating ice sensor based on binocular vision method
CN119779391A (en) * 2024-12-20 2025-04-08 同济大学 Visual inspection device for road surface flatness and compactness mounted on the roller platform
CN119832176A (en) * 2024-12-19 2025-04-15 中国科学院上海微系统与信息技术研究所 Unmanned system autonomous exploration method based on multi-vision scene understanding
CN120014180A (en) * 2025-04-18 2025-05-16 宁波傲视智绘光电科技有限公司 Target three-dimensional reconstruction system and method for constructing three-dimensional contour of moving target
CN120031958A (en) * 2025-04-22 2025-05-23 南京四维向量科技有限公司 A method for measuring recognition and positioning accuracy based on laser tracker and vision
CN120141510A (en) * 2025-05-16 2025-06-13 西安达升科技股份有限公司 Laser vision fusion navigation positioning method, device, equipment and storage medium
CN120213927A (en) * 2025-05-28 2025-06-27 中国矿业大学 Open pit mine slope three-dimensional texture imaging device and slope three-dimensional texture acquisition method
CN120353205A (en) * 2025-06-23 2025-07-22 精兵特种装备(福建)有限公司 Equipment target control data acquisition system and acquisition method thereof
CN120510409A (en) * 2025-07-22 2025-08-19 陕西优鹏安盈科技有限责任公司 High-precision speed measuring method for vehicle by combining laser radar and vision technology
CN120543951A (en) * 2025-07-25 2025-08-26 山东高速建设管理集团有限公司 A visual target detection method and system based on hypergraph computing
CN120720997A (en) * 2025-09-04 2025-09-30 江铃汽车股份有限公司 A system and method for synchronously completing part error prevention detection and gap measurement
CN120740619A (en) * 2025-08-26 2025-10-03 理工雷科智途(北京)科技有限公司 Self-adaptive map building and positioning method, system and medium based on deep learning
CN120993441A (en) * 2025-10-22 2025-11-21 自然资源部北海预报减灾中心(自然资源部青岛海洋中心) A multi-sensor fusion intelligent collision avoidance method and system
CN121121081A (en) * 2025-11-11 2025-12-12 湖南工程学院 A machine vision-based target detection method and system

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115937810A (en) * 2022-11-29 2023-04-07 中车南京浦镇车辆有限公司 A sensor fusion method based on binocular camera guidance
CN116563186B (en) * 2023-05-12 2024-07-12 中山大学 Real-time panoramic sensing system and method based on special AI sensing chip
CN117911525A (en) * 2023-12-20 2024-04-19 清华大学 Multi-mode multi-path complementary visual data calibration method and device
CN119758363B (en) * 2025-03-08 2025-07-11 萃同电子科技(上海)有限公司 Positioning method and system for complementary positioning fusion of laser radar and stereoscopic vision

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111951305A (en) * 2020-08-20 2020-11-17 重庆邮电大学 A target detection and motion state estimation method based on vision and lidar
CN114463303A (en) * 2022-01-29 2022-05-10 合肥工业大学 Road target detection method based on fusion of binocular camera and laser radar
CN114926808A (en) * 2022-03-30 2022-08-19 吉林大学 Target detection and tracking method based on sensor fusion
US20220309835A1 (en) * 2021-03-26 2022-09-29 Harbin Institute Of Technology, Weihai Multi-target detection and tracking method, system, storage medium and application
CN115937810A (en) * 2022-11-29 2023-04-07 中车南京浦镇车辆有限公司 A sensor fusion method based on binocular camera guidance

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111340797B (en) * 2020-03-10 2023-04-28 山东大学 Laser radar and binocular camera data fusion detection method and system
CN114677531B (en) * 2022-03-23 2024-07-09 东南大学 Multi-mode information fusion method for detecting and positioning targets of unmanned surface vehicle

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111951305A (en) * 2020-08-20 2020-11-17 重庆邮电大学 A target detection and motion state estimation method based on vision and lidar
US20220309835A1 (en) * 2021-03-26 2022-09-29 Harbin Institute Of Technology, Weihai Multi-target detection and tracking method, system, storage medium and application
CN114463303A (en) * 2022-01-29 2022-05-10 合肥工业大学 Road target detection method based on fusion of binocular camera and laser radar
CN114926808A (en) * 2022-03-30 2022-08-19 吉林大学 Target detection and tracking method based on sensor fusion
CN115937810A (en) * 2022-11-29 2023-04-07 中车南京浦镇车辆有限公司 A sensor fusion method based on binocular camera guidance

Cited By (38)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN118334016A (en) * 2024-06-12 2024-07-12 青岛交通工程监理咨询有限公司 Highway engineering construction quality detection method
CN118859937A (en) * 2024-07-01 2024-10-29 吉首大学 A real-time position control method and system for multiple intelligent vehicles
CN118408553A (en) * 2024-07-01 2024-07-30 西南科技大学 A UAV navigation method based on three-dimensional reconstruction and recognition of the environment
CN118730129A (en) * 2024-07-11 2024-10-01 奇瑞汽车股份有限公司 A multi-sensor multi-modal tightly coupled fusion mapping and positioning method and system
CN118869899A (en) * 2024-07-15 2024-10-29 浙江大学 A first-person perspective data acquisition method and system based on binocular event camera and depth camera
CN118584523A (en) * 2024-08-01 2024-09-03 浙江易视通联信息科技有限公司 Rescue positioning data fusion analysis method and system based on time series update
CN118584523B (en) * 2024-08-01 2024-11-01 浙江易视通联信息科技有限公司 Rescue positioning data fusion analysis method and system based on time series update
CN118603052A (en) * 2024-08-08 2024-09-06 陕西炬烽建筑劳务有限公司 A terrain survey device for civil engineering
CN119063732A (en) * 2024-08-26 2024-12-03 上海卫星工程研究所 Information processing method, system and medium for integrating on-board real-time target pre-detection and intelligent precision detection
CN119090839A (en) * 2024-08-28 2024-12-06 北京建筑大学 An intelligent detection method for abnormal rail fasteners based on multi-source information fusion
CN118778064A (en) * 2024-09-04 2024-10-15 中科蓝卓(北京)信息科技有限公司 Method and system for monitoring dangerous trees along railways based on optical SAR fusion and binocular depth estimation technology
CN118823275A (en) * 2024-09-18 2024-10-22 湖南理工职业技术学院 A method for modeling and positioning an augmented reality environment for a mobile robot
CN118816935A (en) * 2024-09-18 2024-10-22 青岛慧拓智能机器有限公司 A multi-sensor fusion odometer generation method and system based on surface element fitting
CN119360330A (en) * 2024-09-29 2025-01-24 武汉科技大学 An unpaved road identification and classification system
CN119395718A (en) * 2024-10-14 2025-02-07 东北大学 A multi-sensor fusion simultaneous estimation method for the position and pose of a moving object and multiple targets in a metaverse environment
CN119024268A (en) * 2024-10-25 2024-11-26 西北工业大学 High-precision signal time difference direction finding positioning method and system combined with dual-path vision pod spatial positioning
CN119477980A (en) * 2024-11-04 2025-02-18 中国人民解放军国防科技大学 An interactive multi-model infrared time-sensitive target tracking method based on image feature fusion
CN119495099A (en) * 2024-11-15 2025-02-21 哈尔滨工业大学 A 3D data annotation method for embodied intelligent robots
CN119600105A (en) * 2024-11-20 2025-03-11 上海联适导航技术股份有限公司 A fast and accurate real-time moving target position estimation method and its application
CN119540536A (en) * 2024-11-22 2025-02-28 北京航空航天大学 A vehicle target detection method based on the fusion of binocular camera and lidar
CN119580162B (en) * 2024-11-28 2025-10-14 中国矿业大学 A real-time detection method for mining conveyor belt deviation based on ELAN-CP
CN119580162A (en) * 2024-11-28 2025-03-07 中国矿业大学 A real-time detection method for deviation of mining conveyor belts based on ELAN-CP
CN119274030A (en) * 2024-12-09 2025-01-07 山东高速集团有限公司创新研究院 A highway intelligent inspection method and equipment based on multi-dimensional visual fusion
CN119832176A (en) * 2024-12-19 2025-04-15 中国科学院上海微系统与信息技术研究所 Unmanned system autonomous exploration method based on multi-vision scene understanding
CN119779391A (en) * 2024-12-20 2025-04-08 同济大学 Visual inspection device for road surface flatness and compactness mounted on the roller platform
CN119741384A (en) * 2024-12-26 2025-04-01 西安交通大学 A method and system for calibrating ice sensor based on binocular vision method
CN120014180A (en) * 2025-04-18 2025-05-16 宁波傲视智绘光电科技有限公司 Target three-dimensional reconstruction system and method for constructing three-dimensional contour of moving target
CN120031958A (en) * 2025-04-22 2025-05-23 南京四维向量科技有限公司 A method for measuring recognition and positioning accuracy based on laser tracker and vision
CN120141510A (en) * 2025-05-16 2025-06-13 西安达升科技股份有限公司 Laser vision fusion navigation positioning method, device, equipment and storage medium
CN120213927A (en) * 2025-05-28 2025-06-27 中国矿业大学 Open pit mine slope three-dimensional texture imaging device and slope three-dimensional texture acquisition method
CN120353205A (en) * 2025-06-23 2025-07-22 精兵特种装备(福建)有限公司 Equipment target control data acquisition system and acquisition method thereof
CN120510409A (en) * 2025-07-22 2025-08-19 陕西优鹏安盈科技有限责任公司 High-precision speed measuring method for vehicle by combining laser radar and vision technology
CN120543951A (en) * 2025-07-25 2025-08-26 山东高速建设管理集团有限公司 A visual target detection method and system based on hypergraph computing
CN120740619A (en) * 2025-08-26 2025-10-03 理工雷科智途(北京)科技有限公司 Self-adaptive map building and positioning method, system and medium based on deep learning
CN120720997A (en) * 2025-09-04 2025-09-30 江铃汽车股份有限公司 A system and method for synchronously completing part error prevention detection and gap measurement
CN120720997B (en) * 2025-09-04 2025-11-28 江铃汽车股份有限公司 A system and method for simultaneously performing error-proofing inspection and gap measurement of parts.
CN120993441A (en) * 2025-10-22 2025-11-21 自然资源部北海预报减灾中心(自然资源部青岛海洋中心) A multi-sensor fusion intelligent collision avoidance method and system
CN121121081A (en) * 2025-11-11 2025-12-12 湖南工程学院 A machine vision-based target detection method and system

Also Published As

Publication number Publication date
CN115937810A (en) 2023-04-07

Similar Documents

Publication Publication Date Title
WO2024114119A1 (en) Sensor fusion method based on binocular camera guidance
CN115731268B (en) Unmanned aerial vehicle multi-target tracking method based on vision/millimeter wave radar information fusion
CN113985445B (en) 3D target detection algorithm based on camera and laser radar data fusion
CN111693972B (en) A Vehicle Position and Velocity Estimation Method Based on Binocular Sequence Images
CN110223348B (en) Adaptive pose estimation method for robot scene based on RGB-D camera
CN115032651A (en) A target detection method based on the fusion of lidar and machine vision
CN106681353B (en) Obstacle avoidance method and system for UAV based on binocular vision and optical flow fusion
CN112258600A (en) Simultaneous positioning and map construction method based on vision and laser radar
CN116147618B (en) Real-time state sensing method and system suitable for dynamic environment
CN113223045B (en) Vision and IMU sensor fusion positioning system based on dynamic object semantic segmentation
CN113468950A (en) Multi-target tracking method based on deep learning in unmanned driving scene
CN110246159A (en) The 3D target motion analysis method of view-based access control model and radar information fusion
CN115273034A (en) Traffic target detection and tracking method based on vehicle-mounted multi-sensor fusion
CN114545435B (en) A dynamic target perception system and method integrating camera and laser radar
CN114998276B (en) A real-time detection method for robot dynamic obstacles based on 3D point cloud
CN105160649A (en) Multi-target tracking method and system based on kernel function unsupervised clustering
CN116878501A (en) High-precision positioning and mapping system and method based on multi-sensor fusion
CN116468786B (en) A semantic SLAM method based on point-line joint for dynamic environment
Li et al. BA-LIOM: Tightly coupled laser-inertial odometry and mapping with bundle adjustment
CN117421384A (en) Sliding window optimization method for visual-inertial SLAM system based on common-view projection matching
CN112945233A (en) Global drift-free autonomous robot simultaneous positioning and map building method
Xu et al. Detection-first tightly-coupled LiDAR-Visual-Inertial SLAM in dynamic environments
CN114529603B (en) An odometry method based on the fusion of laser SLAM and monocular vision SLAM
CN113902859B (en) A semi-dense mapping method, storage medium and device for monocular camera imaging
CN115457497A (en) Method for detecting vehicle speed based on 3D target detection and multi-target tracking

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23896275

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 23896275

Country of ref document: EP

Kind code of ref document: A1