GB2641767A - Method for verifying ground-truth data in a scene and system - Google Patents
Method for verifying ground-truth data in a scene and systemInfo
- Publication number
- GB2641767A GB2641767A GB2408372.7A GB202408372A GB2641767A GB 2641767 A GB2641767 A GB 2641767A GB 202408372 A GB202408372 A GB 202408372A GB 2641767 A GB2641767 A GB 2641767A
- Authority
- GB
- United Kingdom
- Prior art keywords
- data
- scene
- ground
- truth data
- labels
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/774—Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Evolutionary Computation (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Software Systems (AREA)
- Computing Systems (AREA)
- Medical Informatics (AREA)
- Databases & Information Systems (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Data Mining & Analysis (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Molecular Biology (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Image Analysis (AREA)
- Traffic Control Systems (AREA)
- Image Processing (AREA)
- Radar Systems Or Details Thereof (AREA)
Abstract
Method for verifying ground truth data in a scene 13 and training an auto-labelling pipeline, comprising: capturing images of objects in a scene using a capture device 11 such as cameras, LIDAR, RADAR, ultrasonic sensor; generating ground truth data by applying an auto-labelling pipeline 12 to the images; verifying the positional alignment between the manually obtained labels and the automatic generated labels for each object of the ground-truth data; identifying data points in the ground truth data that do not align with the manual labels; and utilizing the deviating data points either as examples for improving the autolabelling pipeline or disregarding them. A tool with a simple user interface may be used for documenting the scene setup information without additional burden. The automatically generated labels may be obtained from a machine learning and/or image processing process. The method may be applied to a vehicle 14 advanced driver assistance system (ADAS).
Description
[0001] Mercedes-Benz Group AG
[0002] METHOD FOR VERIFYING GROUND-TRUTH DATA IN A SCENE AND SYSTEM
[0003] FIELD OF THE INVENTION
[0004] The invention relates to the field of automobiles. More specifically, the present invention relates to a method for verifying Ground-Truth Data in a scene according to claim 1. Furthermore, the present invention relates to a system, a corresponding computer program product, and a corresponding non-transitory computer-readable storage medium.
[0005] BACKGROUND INFORMATION
[0006] Learning-based models require Ground-Truth (GT) data, specifically labeled with the desired output for a given input. The model learns the underlying data distribution by minimizing the loss, i.e., the difference between prediction and GT. A problem can arise when the GT data lacks the desired/needed quality.
[0007] One specific application is as follows: A perception system in ADAS/AD aims to generate a 3D geometric environment representation from cost-effective sensor data such as cameras (which only provide a 2D projection), radar, and ultrasonic (both with coarse detections). To generate precise GT labels (the 3D geometric environment), data collection vehicles are equipped with LiDAR as an expert sensor. Automatic labeling is used to generate the labels. However, even the expert sensor can generate erroneous GT data, for example, False Negative (FN) issues due to filtering steps or False Positive (FP) issues due to reflections.
[0008] Some low-quality labels in the GT could significantly impair the entire training process. Since data generation is inexpensive, it is already sufficient to identify and remove these poor-quality labels from the training set.
[0009] The objective of the invention is to develop a method for capturing Groud-TruthData with high quality standards to achieve better performances, using this Data.
[0010] This objective is solved by means of an inventive method with the features of claim 1, as well as by means of an inventive system. Advantageous embodiments of the invention and further developments can be found in the dependent claims and in the description.
[0011] SUMMARY OF THE INVENTION
[0012] One aspect of the invention relates to a method for verifying Ground-Truth data in a scene using a learning-based system.
[0013] A scene for verifying Ground-Truth data using a learning-based system could encompass various contexts depending on the application. For instance, it may be a traffic monitoring scenario at a bustling intersection where vehicles traverse, pedestrians cross the street, and traffic lights regulate flow. The scene may be carefully observed to capture relevant features and information essential for the specific application of the learning-based system.
[0014] Ground-Truth refers to the true or correct data used as a reference for evaluating or validating algorithms or systems. These data serve as a benchmark for assessing the accuracy and performance of models or systems.
[0015] The procedure for verifying Ground-Truth data in a scene using a learning-based system begins with the capture of manual data points using a sensing. For data capture, various sensing devices such as sensors and camera systems could be employed. Therefore an extended data collection may be implemented, where for each data point additionally the scene setup is documented. This includes number of vehicles and other relevant objects in the scene (for parking use cases, for example number of cones, number of pedestrians etc.). This can be tracked in a tool with a very simple UI and no overhead for the data collection team.
[0016] These manual data points serve as a reference for the GT data, representing the actual features in the scene.
[0017] To ensure the accuracy of the GT data, an Auto-Labeling Pipeline is employed to generate additional documentation of the scene for each manual data point. This involves automatically assigning labels to the captured objects or features.
[0018] Subsequently, the correspondence between the manual and automatic labels for each data point is verified. Discrepancies between these labels are identified using comparison and evaluation algorithms. Additionally, an expanded data collection takes place, documenting not only the capture of each data point but also the scene setup. This allows for a more comprehensive analysis of the scene and potential improvement in the quality of GT data.
[0019] The identified deviating data points can either be used as training data to enhance the Auto-Labeling Pipeline or be disregarded. This decision is made through algorithms for data selection and prioritization, aimed at optimizing the quality of GT data.
[0020] Following this, an automated process is implemented, utilizing the manual scene annotations and the generated GT data to verify the correspondence between manual and automatic labels. Through the application of machine learning algorithms and pattern recognition techniques, the accuracy of the GT data is efficiently ensured, leading to enhanced performance of the learning-based system.
[0021] Further advantages, features, and details of the invention derive from the following description of preferred embodiments as well as from the drawing. The features and feature combinations previously mentioned in the description as well as the features and feature combinations mentioned in the following description of the figure and/or shown in the figure alone can be employed not only in the respectively indicated combination but also in any other combination or taken alone without leaving the scope of the invention.
[0022] BRIEF DESCRIPTION OF THE DRAWING
[0023] The novel features and characteristic of the disclosure are set forth in the appended claims. The accompanying drawings, which are incorporated in and constitute a part of this disclosure, illustrate exemplary embodiments and together with the description, serve to explain the disclosed principles. In the figures, the left-most digit(s) of a reference number identifies the figure in which the reference number first appears. The same numbers are used throughout the figures to reference like features and components. Some embodiments of system and/or methods in accordance with embodiments of the present subject matter are now described below, by way of example only, and with reference to the accompanying figures.
[0024] The drawing shows in: [0019] Fig. 1
[0025] DETAILED DESCRIPTION
[0026] In the present document, the word "exemplary" is used herein to mean "serving as an example, instance, or illustration". Any embodiment or implementation of the present subject matter described herein as "exemplary" is not necessarily to be construed as preferred or advantageous over other embodiments.
[0027] While the disclosure is susceptible to various modifications and alternative forms, specific embodiments thereof have been shown by way of example in the drawing and will be described in detail below. It should be understood, however, that it is not intended to limit the disclosure to the particular forms disclosed, but on the contrary, the disclosure is to cover all modifications, equivalents, and alternatives falling within the scope of the disclosure.
[0028] The terms "comprises", "comprising", or any other variations thereof, are intended to cover a non-exclusive inclusion so that a setup, device or method that comprises a list of components or steps does not include only those components or steps but may include other components or steps not expressly listed or inherent to such setup or device or method. In other words, one or more elements in a system or apparatus preceded by "comprises" or "comprise" does not or do not, without more constraints, preclude the existence of other elements or additional elements in the system or method.
[0029] In the following detailed description of the embodiment of the disclosure, reference is made to the accompanying drawing that forms part hereof, and in which is shown by way of illustration a specific embodiment in which the disclosure may be practiced. This embodiment is described in sufficient detail to enable those skilled in the art to practice the disclosure, and it is to be understood that other embodiments may be utilized and that changes may be made without departing from the scope of the present disclosure. The following description is, therefore, not to be taken in a limiting sense.
[0030] Fig. 1 illustrates the process of verifying ground-truth data in a scene 13 using a machine-learning based system 10, for example in a vehicle 14. The diagram marks several components that clarify the process and interaction of the elements. The system's 10 data capture device 11 (optical capture device or sensor device as example) is a central unit for manual data collection and serves as the primary data source. The auto-labeling pipeline 12 uses these manual inputs to generate automatic labels that then serve as ground-truth data. The scene 13, where this data collection occurs, is documented by both the data capture device 11 and the auto-labeling pipeline 12. The manual data points are collected directly in this scene 13 by the capture device 11 and form the basis for the automatic labels produced by the pipeline 12.
[0031] These labels are combined with the manual data points to form the ground-truth data, with each additional piece of information contributing to the documentation of the scene 13. A critical step in the process is verifying the match between manual and automatic labels to ensure data accuracy. This also involves identifying deviant data points, which indicate where the auto-labeling pipeline 12 could be improved.
[0032] Finally, it is decided whether these deviant data points will be used to enhance the pipeline 12 or ignored, directly affecting the system's further optimization. This structured and detailed depiction facilitates understanding of the complex procedures for data verification and optimization within the system 10.
[0033] In summary, the invention proposes a system 10 and a method to annotate and verify ground truth generated by auto-labeling.
[0034] List of reference signs system 11 capture device 12 auto-labeling pipeline 13 scene 14 vehicle
Claims (6)
1. Mercedes-Benz Group AGCLAIMS1. Method for verifying Ground-Truth data in a scene (13) using a learning-based system (10), comprising the following steps: - Capture of manual data points using a system's capture device (11); - Generation of Ground-Truth data by additional documentation of the scene for each manual data point using an auto-labeling pipeline (12); - Verification of the alignment between the manual and automatic labels for each individual manual data point and the corresponding data points of the Ground-Truth data to ensure the expected accuracy of the Ground-Truth data; - Identification of data points in the Ground-Truth data that do not align with the manual labels; - Utilization of these deviating data points either as examples for improving the auto-labeling pipeline (12) or disregarding them.
2. Method according to claim 1, characterized in that the utilization of a tool with a simple user interface for documenting the required scene setup information without additional burden.
3. Method according to any one of claims 1 or 2, characterized in that the automatic generation of the automatic labels is performed using techniques of machine learning and image processing.
4. System to execute a Method or verifying Ground-Truth data in a scene (13) according to any of the preceding claims.
5. Computer program product comprising program code means for performing a method according to claim 1 to 3.
6. A non-transitory computer-readable storage medium comprising at least the computer program product according to claim 5.
Priority Applications (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| GB2408372.7A GB2641767A (en) | 2024-06-12 | 2024-06-12 | Method for verifying ground-truth data in a scene and system |
| DE102024130910.5A DE102024130910A1 (en) | 2024-06-12 | 2024-10-23 | METHOD FOR VERIFIING GROUND TRUTH DATA IN A SCENE AND SYSTEM |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| GB2408372.7A GB2641767A (en) | 2024-06-12 | 2024-06-12 | Method for verifying ground-truth data in a scene and system |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| GB202408372D0 GB202408372D0 (en) | 2024-07-24 |
| GB2641767A true GB2641767A (en) | 2025-12-17 |
Family
ID=91948675
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| GB2408372.7A Pending GB2641767A (en) | 2024-06-12 | 2024-06-12 | Method for verifying ground-truth data in a scene and system |
Country Status (2)
| Country | Link |
|---|---|
| DE (1) | DE102024130910A1 (en) |
| GB (1) | GB2641767A (en) |
Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20180373980A1 (en) * | 2017-06-27 | 2018-12-27 | drive.ai Inc. | Method for training and refining an artificial intelligence |
| US20190102656A1 (en) * | 2017-09-29 | 2019-04-04 | Here Global B.V. | Method, apparatus, and system for providing quality assurance for training a feature prediction model |
| WO2019232641A1 (en) * | 2018-06-07 | 2019-12-12 | Element Ai Inc. | Automated labeling of data with user validation |
| US10699192B1 (en) * | 2019-01-31 | 2020-06-30 | StradVision, Inc. | Method for optimizing hyperparameters of auto-labeling device which auto-labels training images for use in deep learning network to analyze images with high precision, and optimizing device using the same |
| CN111820890A (en) * | 2020-07-24 | 2020-10-27 | 武汉中旗生物医疗电子有限公司 | Electrocardiosignal quality labeling method and device |
-
2024
- 2024-06-12 GB GB2408372.7A patent/GB2641767A/en active Pending
- 2024-10-23 DE DE102024130910.5A patent/DE102024130910A1/en active Pending
Patent Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20180373980A1 (en) * | 2017-06-27 | 2018-12-27 | drive.ai Inc. | Method for training and refining an artificial intelligence |
| US20190102656A1 (en) * | 2017-09-29 | 2019-04-04 | Here Global B.V. | Method, apparatus, and system for providing quality assurance for training a feature prediction model |
| WO2019232641A1 (en) * | 2018-06-07 | 2019-12-12 | Element Ai Inc. | Automated labeling of data with user validation |
| US10699192B1 (en) * | 2019-01-31 | 2020-06-30 | StradVision, Inc. | Method for optimizing hyperparameters of auto-labeling device which auto-labels training images for use in deep learning network to analyze images with high precision, and optimizing device using the same |
| CN111820890A (en) * | 2020-07-24 | 2020-10-27 | 武汉中旗生物医疗电子有限公司 | Electrocardiosignal quality labeling method and device |
Also Published As
| Publication number | Publication date |
|---|---|
| GB202408372D0 (en) | 2024-07-24 |
| DE102024130910A1 (en) | 2025-12-18 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN107945198B (en) | Method and apparatus for labeling point cloud data | |
| CN112966631A (en) | License plate detection and identification system and method under unlimited security scene | |
| CN112990200B (en) | A data labeling method, device, computer equipment and storage medium | |
| CN110533950A (en) | Detection method, device, electronic device and storage medium for parking space usage status | |
| WO2016157499A1 (en) | Image processing apparatus, object detection apparatus, and image processing method | |
| CN110569697A (en) | Vehicle component detection method, device and equipment | |
| CN111967368B (en) | Traffic light identification method and device | |
| CN111008576B (en) | Pedestrian detection and its model training, updating method, device and readable storage medium | |
| CN111435446A (en) | License plate identification method and device based on L eNet | |
| CN115100628B (en) | Obstacle recognition method and device for automatic driving vehicle | |
| CN111444911B (en) | Training method and device of license plate recognition model and license plate recognition method and device | |
| CN110751012A (en) | Target detection evaluation method and device, electronic equipment and storage medium | |
| CN110837760B (en) | Target detection method, training method and apparatus for target detection | |
| CN112733864B (en) | Model training method, target detection method, device, equipment and storage medium | |
| CN108573244B (en) | Vehicle detection method, device and system | |
| CN118247359A (en) | Automatic calibration method and device for fish-eye camera, computer equipment and storage medium | |
| CN112215222A (en) | License plate recognition method, device, equipment and storage medium | |
| CN112700653A (en) | Method, device and equipment for judging illegal lane change of vehicle and storage medium | |
| CN112287905A (en) | Vehicle damage identification method, device, equipment and storage medium | |
| CN111639640A (en) | License plate recognition method, device and equipment based on artificial intelligence | |
| GB2641767A (en) | Method for verifying ground-truth data in a scene and system | |
| CN111160183A (en) | Method and device for detecting vehicle running a red light | |
| CN116740712A (en) | Infrared image target labeling method, device, electronic equipment, and storage medium | |
| CN115240095A (en) | Method, device, equipment and storage medium for identification of vehicle painting method | |
| CN115147626A (en) | Defect detection method and related apparatus, equipment and storage medium |