SE1550006A1 - Method and system for categorization of a scene - Google Patents
Method and system for categorization of a scene Download PDFInfo
- Publication number
- SE1550006A1 SE1550006A1 SE1550006A SE1550006A SE1550006A1 SE 1550006 A1 SE1550006 A1 SE 1550006A1 SE 1550006 A SE1550006 A SE 1550006A SE 1550006 A SE1550006 A SE 1550006A SE 1550006 A1 SE1550006 A1 SE 1550006A1
- Authority
- SE
- Sweden
- Prior art keywords
- images
- scene
- probability
- determining
- foreground object
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/52—Surveillance or monitoring of activities, e.g. for recognising suspicious objects
- G06V20/54—Surveillance or monitoring of activities, e.g. for recognising suspicious objects of traffic, e.g. cars on the road, trains or boats
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/254—Fusion techniques of classification results, e.g. of results related to same input data
- G06F18/256—Fusion techniques of classification results, e.g. of results related to same input data of results relating to different input data, e.g. multimodal recognition
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/215—Motion-based segmentation
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/246—Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
- G06T7/248—Analysis of motion using feature-based methods, e.g. the tracking of corners or segments involving reference images or patches
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/254—Analysis of motion involving subtraction of images
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/285—Analysis of motion using a sequence of stereo image pairs
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/30—Determination of transform parameters for the alignment of images, i.e. image registration
- G06T7/33—Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
- G06T7/344—Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods involving models
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/74—Image or video pattern matching; Proximity measures in feature spaces
- G06V10/75—Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
- G06V10/751—Comparing pixel values or logical combinations thereof, or feature values having positional relevance, e.g. template matching
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/80—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
- G06V10/809—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of classification results, e.g. where the classifiers operate on the same input data
- G06V10/811—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of classification results, e.g. where the classifiers operate on the same input data the classifiers operating on different input data, e.g. multi-modal recognition
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/52—Surveillance or monitoring of activities, e.g. for recognising suspicious objects
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
- G06V20/58—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
- G06V20/584—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads of vehicle lights or traffic lights
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/222—Studio circuitry; Studio devices; Studio equipment
- H04N5/2224—Studio circuitry; Studio devices; Studio equipment related to virtual studio applications
- H04N5/2226—Determination of depth image, e.g. for foreground/background separation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/08—Detecting or categorising vehicles
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Software Systems (AREA)
- Medical Informatics (AREA)
- General Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Health & Medical Sciences (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- General Engineering & Computer Science (AREA)
- Signal Processing (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Description
scene, preferably also allowing for the possibility of extracting further information relating to the detected objects while still keeping the privacy at an in comparison high level.
SUMMARY OF THE INVENTION According to an aspect of the invention, the above is at least partly met by a computer-implemented method for classifying a moving object within a scene, the moving object being one of a plurality of different types of predetermined objects, the method comprising receiving an image stream comprising a plurality of images, each of the plurality of images being a captured representation of the scene and comprising depth information as to content within the scene, wherein the content of the scene comprises at least one physical foreground object and a physical background, processing the plurality of images to produce an equal plurality of depth maps of the scene, forrning height maps for each of the plurality of images based on a predeterrnined model of the background and the corresponding depth map, extracting the at least one physical foreground object from each of the plurality of images based on the corresponding height map, determining, for each of the plurality of images, a relative position of the at least one physical foreground object within the scene based on the height maps, deterrnining a first probability for the type of object within each of the images by matching the extracted physical foreground object with a predeterrnined set of different object types, Wherein each of the predetermined set of different object types are defined to have at least a length, a height and a width being within a predetermined range, deterrnining a second probability for the type of the object by determining a difference in the relative position for the object in at least two images of the plurality of images, converting the difference in relative position to a relative speed for the object, and matching the relative speed with speed profiles for the predeterrnined set of different object types, and de?ning the type of the foreground object based on the first and the second probability.
The general concept of the present invention is based on the fact that it may be possible to improve the classification of objects within a scene by determination of different information about the extracted object, and combine these pieces of information (e.g. a ?rst and a second piece of information about the object as indicated above) based on a probability level for each of the pieces of information. This will advantageously reduce any short term "noise" aspects that may result in error e. g. one of the pieces of information. However, when seen over time and when taking into account the reliability for each of the pieces of information a robust object Classification may be achieved. 10 15 20 25 30 This is speci?cally use?al when performing object Classification in a "mixed object" scene, e. g. where both a pedestrian and a vehicle are present. This will of course be further advantageous in scenes comprises a plurality of pedestrians and vehicles of different forms and shapes, e. g. including cars, trucks and bikes/bicycles.
It may also be possible to further improve the deterrnination of the type of object by including at least a third probability for the type of the obj ect, wherein the third probability is deterinined by matching the extracted foreground object with a set of prede?ned images for the plurality of different types of objects, and deterrnining a probability level, being a representation for the third probability, based on a resulting matching level.
It is advantageous, and within the scope of the invention, to store, for each of the plurality of images, data representing the type of the extracted object(s) and its relative position. Once the data is stored, e.g. within a database, it may for example be possible to track the objects within subsequent images of the image stream. It may also be possible to determine movement pattem for the different types of objects within the scene.
In the most general implementation of the invention, image data is received and processed. However, it may be possible to further control at least two cameras for acquiring the image stream, wherein each of the at least to cameras are arranged to capturing a predeterrnined spatially separated corresponding image of the scene. Each image pair will then need to be processed for deterrnining the depth information based on the corresponding images captured by the at least two cameras, for example using a known stereo/ 3D processing algorithm.
In some implementations of the invention it may be advantageous to determine the absolute location of the object(s) within the scene. This may for example be achieve by means of an absolute location deterrnined within the scene, where the relative locations for the objects based may be transforrned to absolute positions based on the absolute position within the scene. In a possible embodiment, the absolute position within the scene may be acquired from a GPS receiver arranged within the scene.
According to another aspect of the invention there is provided an image processing system for classifying a moving object within a scene, the moving object being one of a plurality of different types of predetermined objects, the image processing system comprising a control unit con?gured to receiving an image stream comprising a plurality of images, each of the plurality of images being a captured representation of the scene and comprising depth information as to content within the scene, wherein the content of the scene comprises at least one physical foreground object and a physical background, processing the 10 15 20 25 30 plurality of images to produce an equal plurality of depth maps of the scene, forming height maps for each of the plurality of images based on a predeterrnined model of the background and the corresponding depth map, extracting the at least one physical foreground object from each of the plurality of images based on the corresponding height map, deterrnining, for each of the plurality of images, a relative position of the at least one physical foreground object within the scene based on the height maps, deterrnining a first probability for the type of object within each of the images by matching the extracted physical foreground object with a predetermined set of different object types, wherein each of the predeterrnined set of different object types are defined to have at least a length, a height and a width being within a predeterrnined range, deterrnining a second probability for the type of the object by deterrnining a difference in the relative position for the object in at least two images of the plurality of images, converting the difference in relative position to a relative speed for the object, and matching the relative speed With speed profiles for the predeterrnined set of different object types, and defining the type of the foreground object based on the first and the second probability. This aspect of the invention provides similar advantages as discussed above in relation to the previous aspects of the invention.
In a preferred embodiment, the system further comprises at least two cameras wherein each of the at least to cameras are arranged to capturing a predeterrnined spatially separated corresponding image of the scene. In such an embodiment, the control unit thus further configured to control at least two cameras for acquiring the image stream, and to determine the depth information based on the corresponding images captured by the at least two cameras.
Within the context of the invention it may alternatively be possible to include a time-of-?ight (ToF) camera for acquiring the depth information as to the objects within the scene. In a possible implementation of the invention, a combination of "normal" 2D cameras are used together with at least one ToF camera According to a still further aspect of the invention there is provided a computer program product comprising a computer readable medium having stored thereon computer program means for operating an image processing system comprising a control unit, the image processing system configured for classifying a moving object within a scene, the moving object being one of a plurality of different types of predetermined objects, the computer program product comprising code for receiving an image stream comprising a plurality of images, each of the plurality of images being a captured representation of the scene and comprising depth information as to content within the scene, wherein the content 10 15 20 25 30 of the scene comprises at least one physical foreground object and a physical background, code for processing the plurality of images to produce an equal plurality of depth maps of the scene, code for forrning height maps for each of the plurality of images based on a predeterrnined model of the background and the corresponding depth map, code for extracting the at least one physical foreground object from each of the plurality of images based on the corresponding height map, code for deterrnining, for each of the plurality of images, a relative position of the at least one physical foreground object within the scene based on the height maps, code for deterrnining a first probability for the type of object within each of the images by matching the extracted physical foreground object with a predeterrnined set of different object types, wherein each of the predeterrnined set of different object types are defined to have at least a length, a height and a width being within a predeterrnined range, code for determining a second probability for the type of the object by deterrnining a difference in the relative position for the object in at least two images of the plurality of images, converting the difference in relative position to a relative speed for the object, and matching the relative speed With speed profiles for the predeterrnined set of different object types, and code for defining the type of the foreground object based on the first and the second probability. Also this aspect of the invention provides similar advantages as discussed above in relation to the previous aspects of the invention.
The control unit preferably including a micro processor or any other type of computing device. Similarly, a so?ware executed by the control unit for operating the inventive wireless communication device may be stored on a computer readable medium, being any type of memory device, including one of a removable nonvolatile random access memory, a hard disk drive, a ?oppy disk, a CD-ROM, a DVD-ROM, a USB memory, an SD memory card, or a similar computer readable medium known in the art. Accordingly, operation of the wireless communication device may be at least partly automated, implemented as e.g. software, hardware and a combination thereof.
Further features of, and advantages with, the present invention will become apparent when studying the appended claims and the following description. The skilled addressee realize that different features of the present invention may be combined to create embodiments other than those described in the following, without departing from the scope of the present invention. 10 15 20 25 30 BRIEF DESCRIPTION OF THE DRAWINGS The various aspects of the invention, including its particular features and advantages, will be readily understood from the following detailed description and the accompanying drawings, in which: Fig. 1 illustrates an image processing system according to the invention arranged for collecting and processing image data of a scene; Fig. 2 provides a conceptual illustration of the image processing system illustrated in Fig. 1; Fig. 3 is a ?ow chart illustrating the exemplary steps for operating the system according to the invention, and Fig. 4 depicts details of object classification according to the invention.
DETAILED DESCRIPTION OF PREFERRED EMBODIMENTS The present invention will now be described more fully hereinafter with reference to the accompanying drawings, in which currently preferred embodiments of the invention are shown. This invention may, however, be embodied in many different forms and should not be construed as limited to the embodiments set forth herein; rather, these embodiments are provided for thoroughness and completeness, and fully convey the scope of the invention to the skilled addressee. Like reference characters refer to like elements throughout.
Referring now to the drawings and Fig. 1 in particular, there is depicted an image processing system 100 arranged in an elevated position as to a ground/street level, in the illustrated embodiment af?xed to an elongated light post 102. The elevated position of the image processing system 100 provides for the possibility of monitoring a scene comprising a plurality of different physical objects. Specifically, in the disclosed embodiment the scene comprises a plurality of different types of physical foreground objects including pedestrians 104, different types of vehicles, such as a truck 106 and a car 108, as well as a plurality of bicycles 110. The scene also comprises a background, for example including a street 112, trees 114, etc. It should be noted that the image processing system 100 may be affixed to or integrated into any form of elevated structure. As such, it could be possible to arrange the image processing system 100 onto e. g. a bridge, an overpass, etc. The image processing system 100 could possibly be integrated with a sign post or any other form of arrangement. The image processing system 100 could be stationary of mobile, connected to the main grid or battery powered. 10 15 20 25 30 The truck 106 and the car 108 travel along the street 112, in opposite directions in different lanes, Whereas the pedestrians 104 and the bicycles 110 shares the pavement space at the outsides of the street traveling at mixed directions at each of the pavements.
In the illustrated embodiment, the image processing system 100 collects image data for a preselected portion of the scene. The size of the preselected portion of the scene may be dependent on the capacity of the image processing system 100, the resolution of the cameras, etc. The collection of image data and possible image processing performed by the image processing system 100 may as exemplified in Fig. 1 be communicated, e. g. Wirelessly or Wired, to a remote server (not shown), for example over the Intemet. Any type of Wired or wireless communication protocol could be possible in accordance to the invention.
Tuming now to Fig. 2, Which conceptually illustrates the image processing system 100 as shown in Fig. 1. The image processing system 100 comprises a first (left) 202 and a second (right) camera 204 having optical lenses connected thereto, respectively. The image processing system 100 comprises a casing 206, preferably configured for handling outdoor use, the casing holding the cameras 202, 204 as Well as processing and communication ?anctionality for the image processing system 100. Typically, the image processing system 100 comprises a control unit 208 arranged in communication With the cameras 202, 204 as Well as a communication module 210 (for allowing communication With e.g. the above mentioned remote server) and a memory element 212 (e. g. for interrnediate storage of image data). It should be understood that the cameras 202, 204 may not necessarily be provided as separate elements, but may be integrated as one single unit provided With a frst and a second set of optical lenses arranged to a first and a second image sensor (i.e. provided for the cameras 202, 204 respectively).
As understood from the above, the image processing system 100 is configured for stereo collection of image data using the first and the second camera 202, 204. To improve the possibility of creating useful three dimensional image data it is desirable to separate the first and the second camera 202, 204 With a predeterrnined distance, the distance being dependent on the desired implementation. Furthermore, within the scope of the invention, it should be understood that it could be possible to use other forms of cameras for image collection, such as using a time-o f-flight (ToF) camera. Also, the general concept of the invention may be implemented based on image data captured using "external" cameras (i.e. not necessarily including the cameras themselves). That is, the invention may for example be implemented as a server solution Where the server receives image data from e.g. 10 15 20 25 30 remotely arranged Cameras. As such, at least some of the processing performed could be implemented in a distributed manner, e. g. partly as a "cloud" solution.
During operation of the image processing system 100, With further reference to Fig. 3, the process starts by the control unit 208 receiving, S1, an image stream from the first and the second camera 202, 204. The control unit 208 subsequently process, S2, the image stream for forrning a three dimensional representation of the scene monitored by the cameras 202, 204. The three dimensional representation is typically formed based on depth information extracted based on the predeterrnined separation of the ?rst 202 and the second 204 camera, thereby forming depth maps for each image of the image stream (i.e. pair of images from the cameras 202, 204).
Based on the depth map and a predetermined model of the background it is possible to forrn, S3, height maps for each of the image pair. The background model could for example be an "empty" scene where no foreground objects are present. The background model could also be deterrnined in other manners as Would be readily understood by the skilled addressee.
The height maps are subsequently used for extracting a physical foreground object from the scene, such as a pedestrian 104, a car 106 or a bicycle 110. The extracted foreground objects are typically defined to have a "boundary box" closely surrounding the object. The boundary box may typically be rectangular but may also be defined in a more freeforrn manner.
For each of the image pair a relative position for the foreground object is deterrnined, S5. The relative position is defined as being relative to e.g. one or a plurality of predetermined "points" Within the scene. However, it may of course be possible to make the position "absolute" by correlating one or a plurality of points Within the scene to e.g. be a real GPS position, or similar.
For each extracted object Within the scene a first probability as to the type of object is determined, S6. In a preferred embodiment of the invention a set of object types are prede?ned. Each of the set of the predefined object types are de?ned to have at least a length, a height and a width being within a predeterrnined range. The foreground object being extracted from the scene is accordingly matched to the set of predefined object types, and a probability is deterinined for the extracted foreground object as compared to the predefined object types. As understood, the set of prede?ned object types typically includes a set of height, width and length parameters defining the pedestrian, the bicycle, the car and the truck.
The size (e.g. the boundary box as mentioned above) of the extracted foreground object is 10 15 20 25 30 thus compared to the predefmed parameters for each of the type of objects for deterrnining the first probability.
A second probability is deterrnined, S7, where the extracted foreground object essentially is "tracked" between subsequent image pairs of the image stream. In accordance to the invention a difference in the relative position is determined for the extracted foreground object, i.e. from one image pair to the next. Accordingly, it may be possible to, for each of the predefined object types, form a set of "speed" parameters/profile defining how fast the speci?c object type likely will move between subsequent image pairs. The tracking will generally also include determination of the direction of movement of the extracted object.
Finally, the type of the foreground object is deterrnined, S8, by combining the first and the second probability, possibly with further information for increasing the probability of correctly determining the object type for the extracted foreground object.
This is further exempli?ed in Fig. 4 showing details of the object classification method according to the invention. Specifically, the ?owchart illustrates the above discussed geometric matching classification, e.g. including matching of size and "moving object positions" for subsequent image pairs with predefmed object type parameters (i.e. 404 and 406). In addition, a further, e. g. possibly defined as a third probability may be deterrnined by performing an image matching 408 between the extracted foreground object and representative image data stored in a database 410. The image matching may preferably take into account the direction of movement of the extracted object, for example for use in improving the selection of the "correct" image to match with, i.e. as the stored images to match with may be annotated to include information as to an aspect angle for the stored image. The third probability for the object type may typically be combined with the first and the second probability.
It should be understood that the combination of the first, second and third probability may be weighted, e.g. more weight could for example be given to the height, width and length parameter matching as compared to the remaining two probabilities. Any type of weighting may be possible and is within the scope of the invention. Also, it should be understood that the combined probability may be accumulated over time, i.e. by allowing further geometric matching classifications and/or image based classifications to be made for the same object. That is, it will likely be preferred to continue to track the object as long as the object is visible within the scene. Thereby, the likelihood of determining the correct object type will be increased. 10 15 20 25 30 10 Once the object type has been deterrnined, i.e. being one of the prede?ned types of object, this knowledge may be used for a numerous different applications. For example, the image processing system 100 could be used in a traf?c counting scenario comprising tracking mixed object types for categorizing the behavior for each of the object types. Similarly, the image processing system 100 according to the invention could be used for controlling traffic patterns, i.e. to control traf?c lights in a city, for example by adapting time given for passing a street. For example, at peak hours with a high density of bicycles, the bicycles could be given more passing time as compared to cars, possibly improving the overall traffic environment within the city.
In addition, the control functionality of the present disclosure may be implemented using existing computer processors, or by a special purpose computer processor for an appropriate system, incorporated for this or another purpose, or by a hardwired system.
Embodiments within the scope of the present disclosure include program products comprising machine-readable media for carrying or having machine-executable instructions or data structures stored thereon. Such machine-readable media can be any available media that can be accessed by a general purpose or special purpose computer or other machine with a processor. By way of example, such machine-readable media can comprise RAM, ROM, EPROM, EEPROM, CD-ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to carry or store desired program code in the form of machine-executable instructions or data structures and which can be accessed by a general purpose or special purpose computer or other machine with a processor. When information is transferred or provided over a network or another communications connection (either hardwired, Wireless, or a combination of hardwired or wireless) to a machine, the machine properly views the connection as a machine-readable medium. Thus, any such connection is properly terrned a machine-readable medium.
Combinations of the above are also included within the scope of machine-readable media.
Machine-executable instructions include, for example, instructions and data which cause a general purpose computer, special purpose computer, or special purpose processing machines to perform a certain function or group of functions.
Although the figures may show a sequence the order of the steps may differ from what is depicted. Also two or more steps may be performed concurrently or with partial concurrence. Such variation will depend on the software and hardware systems chosen and on designer choice. All such Variations are within the scope of the disclosure. Likewise, software implementations could be accomplished with standard programming techniques 10 ll with rule based logic and other logic to accomplish the various connection steps, processing steps, comparison steps and decision steps. Additionally, even though the invention has been described With reference to specific exemplifying embodiments thereof, many different alterations, modiñcations and the like Will become apparent for those skilled in the art.
Further, a single unit may perforrn the functions of several means recited in the claims. In the claims, any reference signs placed between parentheses shall not be construed as limiting to the claim. Furthermore, in the claims, the Word "comprising" does not exclude other elements or steps, and the inde?nite article "a" or "an" does not exclude a plurality.
Variations to the disclo sed embodiments can be understood and effected by the skilled addressee in practicing the claimed inventíon, from a study of the drawings, the disclosure, and the appended claims. The person skilled in the art realizes that the present invention is not limited to the preferred embodiments.
Claims (7)
1. kod for att bearbeta namnda flertalet bilder for att skapa en lika stor mangd 30 av djupkartor for scenen;
2. kod for att skapa hojdkartor for vart och ett av namnda flertal bilder baserat pa en forutbestamd modell av bakgrunden och den motsvarande djupkartan;
3. kod for att extrahera atminstone ett fysiskt forgrundsobjekt fran var och en av namnda flertalet av bilder baserat pa den motsvarande hojdkartan;
4. kod for att bestamma, for var och en av namnda flertal bilder, en relativ position hos namnda atminstone ett fysiskt forgrundsobjekt i scenen baserat pa hOjdkartorna;
5. kod for att bestamma en forsta sannolikhetsniva for typen av objekt inom var och en av bildema genom att matcha det extraherade fysiska forgrundsobjektet med en forutbestamd uppsattning av olika objekttyper, varvid var och en av den forutbestamda uppsattningen av olika objekttyper är definierade att ha kminstone en langd, en hojd och en bredd som är inom ett forutbestamt omfang;
6. kod for att bestamma en andra sannolikhetsniva for typen av objekt genom att bestamma en skillnad i den relativa positionen for objektet i atminstone tva bilder av namnda flertal bilder, omvandla skillnaden i relativ position till en relativ hastighet for objektet och matcha den relativa hastigheten med hastighetsprofiler for den forutbestamda uppsattningen av olika objekttyper; och
7. kod for att definiera typcn av forgrundsobjektet bascrat pa en kombination av den forsta och den andra sannolikhetsnivan. zot 214 102 100 /
Priority Applications (5)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| SE1550006A SE1550006A1 (sv) | 2015-01-07 | 2015-01-07 | Method and system for categorization of a scene |
| PCT/SE2015/051382 WO2016111640A1 (en) | 2015-01-07 | 2015-12-21 | Method and system for categorization of a scene |
| US15/538,498 US10127458B2 (en) | 2015-01-07 | 2015-12-21 | Method and system for categorization of a scene |
| EP15877226.9A EP3243165B1 (en) | 2015-01-07 | 2015-12-21 | Method and system for categorization of a scene |
| CA2970911A CA2970911C (en) | 2015-01-07 | 2015-12-21 | Method and system for categorization of a scene |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| SE1550006A SE1550006A1 (sv) | 2015-01-07 | 2015-01-07 | Method and system for categorization of a scene |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| SE538405C2 SE538405C2 (sv) | 2016-06-14 |
| SE1550006A1 true SE1550006A1 (sv) | 2016-06-14 |
Family
ID=56108394
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| SE1550006A SE1550006A1 (sv) | 2015-01-07 | 2015-01-07 | Method and system for categorization of a scene |
Country Status (5)
| Country | Link |
|---|---|
| US (1) | US10127458B2 (sv) |
| EP (1) | EP3243165B1 (sv) |
| CA (1) | CA2970911C (sv) |
| SE (1) | SE1550006A1 (sv) |
| WO (1) | WO2016111640A1 (sv) |
Families Citing this family (10)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US10453187B2 (en) * | 2017-07-21 | 2019-10-22 | The Boeing Company | Suppression of background clutter in video imagery |
| KR102370770B1 (ko) * | 2017-09-15 | 2022-03-04 | 킴벌리-클라크 월드와이드, 인크. | 세면실 디바이스 증강 현실 설치 시스템 |
| US10650530B2 (en) * | 2018-03-29 | 2020-05-12 | Uveye Ltd. | Method of vehicle image comparison and system thereof |
| US10643332B2 (en) * | 2018-03-29 | 2020-05-05 | Uveye Ltd. | Method of vehicle image comparison and system thereof |
| US10861165B2 (en) * | 2019-01-11 | 2020-12-08 | Microsoft Technology Licensing, Llc | Subject tracking with aliased time-of-flight data |
| WO2021204344A1 (en) * | 2020-04-06 | 2021-10-14 | HELLA GmbH & Co. KGaA | Method and system for detecting a vehicle having at least one wheel |
| US11461993B2 (en) * | 2021-01-05 | 2022-10-04 | Applied Research Associates, Inc. | System and method for determining the geographic location in an image |
| EP4123598B1 (en) * | 2021-07-19 | 2025-10-15 | Axis AB | Masking of objects in a video stream |
| EP4343709B1 (en) * | 2022-09-23 | 2024-10-30 | Axis AB | Method and image-processing device for determining a probability value indicating that an object captured in a stream of image frames belongs to an object type |
| CN118397202B (zh) * | 2024-06-28 | 2024-09-06 | 杭州群核信息技术有限公司 | 图像处理方法、装置、设备以及存储介质 |
Family Cites Families (27)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP3573512B2 (ja) * | 1994-05-17 | 2004-10-06 | オリンパス株式会社 | 画像処理方法及び画像処理装置 |
| JP3745117B2 (ja) * | 1998-05-08 | 2006-02-15 | キヤノン株式会社 | 画像処理装置及び画像処理方法 |
| US6570608B1 (en) * | 1998-09-30 | 2003-05-27 | Texas Instruments Incorporated | System and method for detecting interactions of people and vehicles |
| US7127083B2 (en) * | 2003-11-17 | 2006-10-24 | Vidient Systems, Inc. | Video surveillance system with object detection and probability scoring based on object class |
| US7486803B2 (en) * | 2003-12-15 | 2009-02-03 | Sarnoff Corporation | Method and apparatus for object tracking prior to imminent collision detection |
| JP2005215909A (ja) * | 2004-01-29 | 2005-08-11 | Hitachi Ltd | 動画像処理技術を使用した市街の交通状況調査システム |
| KR100519782B1 (ko) * | 2004-03-04 | 2005-10-07 | 삼성전자주식회사 | 스테레오 카메라를 이용한 사람 검출 방법 및 장치 |
| US7466860B2 (en) * | 2004-08-27 | 2008-12-16 | Sarnoff Corporation | Method and apparatus for classifying an object |
| GB0502369D0 (en) * | 2005-02-04 | 2005-03-16 | British Telecomm | Classifying an object in a video frame |
| US7602944B2 (en) * | 2005-04-06 | 2009-10-13 | March Networks Corporation | Method and system for counting moving objects in a digital video stream |
| US7567704B2 (en) | 2005-11-30 | 2009-07-28 | Honeywell International Inc. | Method and apparatus for identifying physical features in video |
| US7965866B2 (en) | 2007-07-03 | 2011-06-21 | Shoppertrak Rct Corporation | System and process for detecting, tracking and counting human objects of interest |
| DE102007058959A1 (de) * | 2007-12-07 | 2009-06-10 | Robert Bosch Gmbh | Konfigurationsmodul für ein Überwachungssystem, Überwachungssystem, Verfahren zur Konfiguration des Überwachungssystems sowie Computerprogramm |
| JP5264396B2 (ja) * | 2008-10-03 | 2013-08-14 | キヤノン株式会社 | 画像処理装置及び画像種別特定方法 |
| CN101388145B (zh) * | 2008-11-06 | 2010-09-15 | 北京汇大基业科技有限公司 | 道路交通安全自动警示方法及装置 |
| AU2008365306B2 (en) | 2008-12-19 | 2014-02-13 | Saab Ab | System and method for mixing a scene with a virtual scenario |
| WO2010124062A1 (en) * | 2009-04-22 | 2010-10-28 | Cernium Corporation | System and method for motion detection in a surveillance video |
| US20130163879A1 (en) * | 2010-08-30 | 2013-06-27 | Bk-Imaging Ltd. | Method and system for extracting three-dimensional information |
| US8903119B2 (en) * | 2010-10-11 | 2014-12-02 | Texas Instruments Incorporated | Use of three-dimensional top-down views for business analytics |
| US8704887B2 (en) * | 2010-12-02 | 2014-04-22 | GM Global Technology Operations LLC | Multi-object appearance-enhanced fusion of camera and range sensor data |
| CN103959308B (zh) | 2011-08-31 | 2017-09-19 | Metaio有限公司 | 以参考特征匹配图像特征的方法 |
| US9338409B2 (en) | 2012-01-17 | 2016-05-10 | Avigilon Fortress Corporation | System and method for home health care monitoring |
| WO2014186642A2 (en) | 2013-05-17 | 2014-11-20 | International Electronic Machines Corporation | Operations monitoring in an area |
| CN103324955A (zh) * | 2013-06-14 | 2013-09-25 | 浙江智尔信息技术有限公司 | 一种基于视频处理的行人检测方法 |
| US9656667B2 (en) * | 2014-01-29 | 2017-05-23 | Continental Automotive Systems, Inc. | Method for minimizing automatic braking intrusion based on collision confidence |
| EP3026653A1 (en) * | 2014-11-27 | 2016-06-01 | Kapsch TrafficCom AB | Method of controlling a traffic surveillance system |
| US9754490B2 (en) * | 2015-11-04 | 2017-09-05 | Zoox, Inc. | Software application to request and control an autonomous vehicle service |
-
2015
- 2015-01-07 SE SE1550006A patent/SE1550006A1/sv not_active IP Right Cessation
- 2015-12-21 CA CA2970911A patent/CA2970911C/en active Active
- 2015-12-21 WO PCT/SE2015/051382 patent/WO2016111640A1/en not_active Ceased
- 2015-12-21 EP EP15877226.9A patent/EP3243165B1/en active Active
- 2015-12-21 US US15/538,498 patent/US10127458B2/en active Active
Also Published As
| Publication number | Publication date |
|---|---|
| US10127458B2 (en) | 2018-11-13 |
| CA2970911A1 (en) | 2016-07-14 |
| EP3243165B1 (en) | 2023-08-02 |
| SE538405C2 (sv) | 2016-06-14 |
| EP3243165A4 (en) | 2018-08-15 |
| EP3243165A1 (en) | 2017-11-15 |
| WO2016111640A1 (en) | 2016-07-14 |
| CA2970911C (en) | 2021-03-02 |
| US20170344835A1 (en) | 2017-11-30 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| SE1550006A1 (sv) | Method and system for categorization of a scene | |
| CN112417953B (zh) | 道路状况检测和地图数据更新方法、装置、系统及设备 | |
| CN108345822B (zh) | 一种点云数据处理方法及装置 | |
| US20150339924A1 (en) | Parking space occupancy | |
| EP3516582A1 (en) | Autonomous route determination | |
| KR102657921B1 (ko) | 융합된 이미지를 사용한 엔드 투 엔드 시스템 트레이닝 | |
| CN115240094A (zh) | 一种垃圾检测方法和装置 | |
| CN112836657A (zh) | 一种基于轻量化YOLOv3的行人检测方法及系统 | |
| Józsa et al. | Towards 4D virtual city reconstruction from Lidar point cloud sequences | |
| KR20250089500A (ko) | 바운딩 박스를 사용하는 레이더 및 비전 기반의 내비게이션 | |
| US12272120B2 (en) | Enriching feature maps using multiple pluralities of windows to generate bounding boxes | |
| KR20250046254A (ko) | 에이전트 우선순위화를 위한 방법 및 시스템 | |
| KR20250085758A (ko) | 장면 종속적 대상체 쿼리를 사용한 바운딩 박스의 생성 | |
| CN119380309B (zh) | 路网生成方法、智能设备及计算可读存储介质 | |
| CN117593685B (zh) | 真值数据的构建方法和装置、存储介质 | |
| KR20250157517A (ko) | 엔드 투 엔드 변환기 기반 경계 박스 추적 | |
| KR20250157411A (ko) | 경계 박스를 사용한 대상체의 다중 모달 센서 기반 검출 및 추적 | |
| KR20250057828A (ko) | 이전-시점 특징 맵을 사용하는 이후-시점 특징 맵의 보강 | |
| KR20260005265A (ko) | 3d 대상체 검출을 위한 고정밀 맵 융합 | |
| Koetsier et al. | Trajectory extraction for analysis of unsafe driving behaviour | |
| Ke | Real-time video analytics empowered by machine learning and edge computing for smart transportation applications | |
| Nadella | Extracting road traffic data through video analysis using automatic camera calibration and deep neural networks | |
| Menard | Tracking, Recognizing, and Analyzing Flow For Intersection Control: TRAFFIC: Video-based vehicle recognition and counting at intersections. | |
| KR20250167088A (ko) | 경계 박스를 사용한 다중 모달 센서 기반 운행 | |
| Brunnegård et al. | Visual SLAM using sparse maps based on feature points |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| NUG | Patent has lapsed |