CN108009473A - Based on goal behavior attribute video structural processing method, system and storage device - Google Patents
Based on goal behavior attribute video structural processing method, system and storage device Download PDFInfo
- Publication number
- CN108009473A CN108009473A CN201711055281.0A CN201711055281A CN108009473A CN 108009473 A CN108009473 A CN 108009473A CN 201711055281 A CN201711055281 A CN 201711055281A CN 108009473 A CN108009473 A CN 108009473A
- Authority
- CN
- China
- Prior art keywords
- target
- frame
- detection
- video
- targets
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
- G06V20/41—Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/71—Indexing; Data structures therefor; Storage structures
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/78—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/783—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
- G06F16/7837—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using objects detected or recognised in the video content
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/52—Surveillance or monitoring of activities, e.g. for recognising suspicious objects
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Multimedia (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Software Systems (AREA)
- Data Mining & Analysis (AREA)
- Library & Information Science (AREA)
- Databases & Information Systems (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Molecular Biology (AREA)
- Evolutionary Computation (AREA)
- Mathematical Physics (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Health & Medical Sciences (AREA)
- Image Analysis (AREA)
Abstract
本发明公开了一种基于目标行为属性的视频结构化处理的方法,该方法包括:利用YOLO目标检测算法获取目标基本属性;利用多目标跟踪算法获取上述已检测目标的轨迹信息;利用基于运动光流特征的异常行为分析算法提取异常视频帧;根据自定义构建的元数据结构,利用上述方法获取相对应的目标类别属性以及目标轨迹等特征信息;采用加权判定方法对所提取的元数据中存在的误检数据进行修正;将所获取的数据上传到后端的服务器进一步处理。通过上述方式,本发明可以将非结构化的视频数据转为具有实用价值的结构化数据,提升了视频监控系统的网络传输效率以及降低后端服务器负载率。本发明还提供了一种基于目标行为属性的实时处理系统以及装置。
The invention discloses a method for structured video processing based on target behavior attributes. The method includes: using the YOLO target detection algorithm to obtain the basic attributes of the target; using a multi-target tracking algorithm to obtain the track information of the detected target; The abnormal behavior analysis algorithm of stream features extracts abnormal video frames; according to the custom-built metadata structure, use the above method to obtain the corresponding target category attributes and target trajectory and other characteristic information; Correct the misdetection data; upload the obtained data to the back-end server for further processing. Through the above method, the present invention can convert unstructured video data into structured data with practical value, which improves the network transmission efficiency of the video monitoring system and reduces the load rate of the back-end server. The invention also provides a real-time processing system and device based on the target behavior attribute.
Description
技术领域technical field
本发明涉及计算机视觉领域,特别是涉及一种基于目标行为属性视频结构化处理方法、系统及存储装置。The invention relates to the field of computer vision, in particular to a video structure processing method, system and storage device based on target behavior attributes.
背景技术Background technique
随着智能监控技术的发展,对于视频的处理尤为重要。现有技术中,对于视频的处理多采用图像特征检测的方法,但是由于视频中维数会很高,且会有大量的冗余特征和无关特征,这样就造成了视频处理的压力,无法实现快速的处理视频,且会降低获取目标特征的准确率。所以为了满足智能化监控技术发展的需求,需要一种高实时性和高准确率的视频处理的方法、视频处理系统。With the development of intelligent monitoring technology, the processing of video is particularly important. In the prior art, the method of image feature detection is often used for video processing, but since the dimensionality of the video will be very high, and there will be a large number of redundant features and irrelevant features, this will cause the pressure of video processing and cannot be realized. The video is processed quickly, and the accuracy of obtaining target features will be reduced. Therefore, in order to meet the needs of the development of intelligent monitoring technology, a video processing method and a video processing system with high real-time performance and high accuracy are required.
发明内容Contents of the invention
本发明主要解决的技术问题是,提供一种基于目标行为属性的视频结构化处理的方法、系统及存储装置,可以实现对视频进行高实时性和高准确率的处理。The technical problem mainly solved by the present invention is to provide a video structured processing method, system and storage device based on target behavior attributes, which can realize high-real-time and high-accuracy processing of video.
为解决上述技术问题,本发明采用的技术方案是提供一种基于目标行为属性的视频结构化处理的方法,包括以下步骤:In order to solve the above technical problems, the technical solution adopted by the present invention is to provide a method for structured video processing based on target behavior attributes, including the following steps:
对所述单帧图片进行目标检测识别;Perform target detection and recognition on the single frame picture;
对所述目标进行跟踪,以得到跟踪结果;和/或者track the target to obtain a tracking result; and/or
对所述目标进行异常行为检测。Abnormal behavior detection is performed on the target.
为解决上述技术问题,本发明采用的另一个技术方案是:提供基于目标行为属性的视频结构化处理系统,包括相互电性连接的处理器和存储器,所述处理器耦合所述存储器,所述处理器在工作时执行指令以实现上述的视频处理方法,并将所述执行指令产生的处理结果保存在所述存储器中。In order to solve the above-mentioned technical problems, another technical solution adopted by the present invention is to provide a video structured processing system based on target behavior attributes, including a processor and a memory electrically connected to each other, the processor is coupled to the memory, and the When working, the processor executes instructions to implement the above video processing method, and stores the processing results generated by executing the instructions in the memory.
为解决上述技术问题,本发明采用的又一个技术方案是提供一种具有存储功能的装置,存储有程序数据,所述程序数据被执行时实现上述的视频处理方法。以上技术方案的有益效果是:区别于现有技术的情况,本发明通过将视频切分成单帧图片,并对单帧图片进行目标检测识别,对识别到的目标进行跟踪,以得到跟踪结果,并对识别到的目标进行异常行为检测,在此过程中,实现从非结构化的视频中提取结构化的数据,可以有效的实现视频数据传输的高实时性和高准确率的处理。In order to solve the above technical problems, another technical solution adopted by the present invention is to provide a device with a storage function, storing program data, and implementing the above video processing method when the program data is executed. The beneficial effects of the above technical solutions are: different from the situation of the prior art, the present invention divides the video into single-frame pictures, and performs target detection and recognition on the single-frame pictures, and tracks the recognized targets to obtain tracking results. And detect the abnormal behavior of the recognized target. In this process, the structured data can be extracted from the unstructured video, which can effectively realize the high real-time and high accuracy processing of the video data transmission.
附图说明Description of drawings
图1是本申请基于目标行为属性的视频结构化处理的方法的一实施方式的流程示意图;FIG. 1 is a schematic flow diagram of an embodiment of a method for structured video processing based on target behavior attributes in the present application;
图2是本申请基于目标行为属性的视频结构化处理的方法的另一实施例的流程示意图;FIG. 2 is a schematic flow diagram of another embodiment of the method for structured video processing based on target behavior attributes of the present application;
图3是本申请基于目标行为属性的视频结构化处理的方法的又一实施例的流程示意图;FIG. 3 is a schematic flow diagram of another embodiment of the method for structured video processing based on target behavior attributes in the present application;
图4是本申请基于目标行为属性的视频结构化处理的方法的又一实施例的流程示意图;FIG. 4 is a schematic flow diagram of another embodiment of the method for structured video processing based on target behavior attributes in the present application;
图5是本申请基于目标行为属性的视频结构化处理的方法的再一实施例的流程示意图;FIG. 5 is a schematic flow diagram of another embodiment of the method for structured video processing based on target behavior attributes in the present application;
图6是本申请基于目标行为属性的视频结构化处理的方法的又一实施例的流程示意图;FIG. 6 is a schematic flow diagram of another embodiment of the method for structured video processing based on target behavior attributes in the present application;
图7是本申请基于目标行为属性的视频结构化处理的方法的又一实施例的流程示意图;FIG. 7 is a schematic flow diagram of another embodiment of the method for structured video processing based on target behavior attributes in the present application;
图8是图7提供的实施例中步骤S243的一实施例的流程示意图;FIG. 8 is a schematic flowchart of an embodiment of step S243 in the embodiment provided in FIG. 7;
图9是本申请基于目标行为属性的视频结构化处理的方法的一实施例中的运动时空容器的示意图;FIG. 9 is a schematic diagram of a motion space-time container in an embodiment of a method for structured video processing based on target behavior attributes in the present application;
图10是本申请一种具有存储功能的装置的一实施例的示意图;FIG. 10 is a schematic diagram of an embodiment of a device with a storage function in the present application;
图11是本申请基于目标行为属性的视频结构化处理系统的一实施例结构示意图。FIG. 11 is a schematic structural diagram of an embodiment of a video structured processing system based on target behavior attributes of the present application.
具体实施方式Detailed ways
在下文中,将参照附图来描述本申请的示例性实施例。为了清楚和简要的目的,不详细描述公知的功能和构造。考虑到本申请中的功能而限定的下面描述的术语可以根据用户和操作者的意图或实施而不同。因此,应该在整个说明书的公开的基础上来限定所述术语。Hereinafter, exemplary embodiments of the present application will be described with reference to the accompanying drawings. Well-known functions and constructions are not described in detail for clarity and conciseness. Terms described below, which are defined in consideration of functions in the present application, may vary according to user and operator's intention or implementation. Therefore, the terms should be defined on the basis of the disclosure throughout the specification.
请参阅图1,为本发明基于视频结构化数据及深度学习的视频监控方法的方法的第一实施方式的流程示意图。该方法包括:Please refer to FIG. 1 , which is a schematic flowchart of a first embodiment of the video surveillance method based on video structured data and deep learning in the present invention. The method includes:
S10:读取视频。S10: Read the video.
可选地,读取视频包括读取摄像头采集的实时视频和/或预先录制保存的视频的数据。其中,采集实时视频的摄像头,可以是USB摄像头和基于rtsp协议流的网络摄像头其中的一种,或者其他种类的摄像头。Optionally, reading the video includes reading real-time video collected by the camera and/or pre-recorded and saved video data. Wherein, the camera for collecting real-time video may be one of a USB camera and a network camera based on rtsp protocol stream, or other types of cameras.
在一实施例,读取的视频是USB摄像头或者是基于rtsp协议流的网络摄像头实时拍摄采集的视频。In one embodiment, the read video is a video captured by a USB camera or a network camera based on an rtsp protocol stream in real time.
在另一实施例中,读取的视频是预先录制保存的视频,通过从本地存储器或者是如U盘、硬盘等外部存储设备输入读取的,也可以是从网络上调取的视频,在此不一一详述。In another embodiment, the read video is a pre-recorded and saved video, which is input and read from a local memory or external storage devices such as U disk, hard disk, etc., or it can be a video retrieved from the network. This is not detailed one by one.
S20:对视频进行结构化处理,得到结构化数据。S20: Perform structured processing on the video to obtain structured data.
可选地,对视频进行结构化处理,得到结构化数据具体是指,将步骤S10中读取的非结构化的视频数据转化成机构化的数据,具体的,结构化数据是指对于后续分析比较重要的数据。可选地,结构化数据包括目标的位置、目标类别、目标属性、目标运动状态、目标运动轨迹、目标驻留时间中的至少一个信息,其中,可以理解的是,结构化数据也可以包括用户(使用本发明的中所述的方法或系统的人)需要的其他类别的信息。其他数据不是特别重要,或者可以通过结构化数据等相关信息进行挖掘得到。结构化信息的具体包括哪些信息,根据不同需求而定。关于如何将结构化数据处理,以得到结构化数据,下文会做详细的阐述。Optionally, performing structured processing on the video to obtain structured data specifically refers to converting the unstructured video data read in step S10 into structured data. Specifically, structured data refers to more important data. Optionally, the structured data includes at least one information of the target's location, target category, target attribute, target motion state, target motion trajectory, and target dwell time, where it can be understood that the structured data may also include user Other categories of information required by (a person using the method or system described in the present invention). Other data is not particularly important, or can be obtained through mining related information such as structured data. The specific information included in the structured information depends on different needs. How to process structured data to obtain structured data will be described in detail below.
S30:将结构化数据上传至云端服务器,并对结构化数据进行深入分析,以得到预设结果。S30: Upload the structured data to the cloud server, and conduct in-depth analysis on the structured data to obtain preset results.
可选地,在步骤S20将视频结构化处理之后,将所得到的结构化的数据上传至云端服务器,存储到云端服务器的存储区。Optionally, after the structured video is processed in step S20, the obtained structured data is uploaded to the cloud server and stored in the storage area of the cloud server.
在一实施例中,将视频结构化处理的所得到的数据,直接保存到云端服务器的存储区,用以留存档案,也用作完善本系统的数据库。In one embodiment, the data obtained through structured processing of the video is directly stored in the storage area of the cloud server, to preserve files and also to improve the database of the system.
可选地,在步骤S20将视频处理之后,将得到的结构化数据上传至云端服务器,云端服务器对这些结构化数据进行进一步的深入分析。Optionally, after the video is processed in step S20, the obtained structured data is uploaded to the cloud server, and the cloud server performs further in-depth analysis on the structured data.
可选地,云端服务器对从各个监控节点上传的结构化的数据进行的进一步的深入分析,其中,深入分析包括目标轨迹分析和目标流量分析或其他所需的分析,目标包括人、车以及动物等其中的至少一种。Optionally, the cloud server performs further in-depth analysis on the structured data uploaded from each monitoring node, wherein the in-depth analysis includes target trajectory analysis and target traffic analysis or other required analysis, and the targets include people, vehicles and animals Wait for at least one of these.
在一实施例中,云端服务器对对从各个监控节点上传的结构化的数据进行的进一步的深入分析是轨迹分析,根据上传的目标的轨迹的规律、在该场景驻留时间来进一步判定该目标是否可疑,该目标是否是长时间滞留在某一区域,是否发生区域入侵等异常行为。In one embodiment, the further in-depth analysis of the structured data uploaded from each monitoring node by the cloud server is trajectory analysis, and the target is further judged according to the law of the trajectory of the uploaded target and the residence time in the scene Whether it is suspicious, whether the target stays in a certain area for a long time, whether abnormal behavior such as area intrusion occurs.
在另一实施例中,云端服务器对对从各个监控节点上传的结构化的数据进行的进一步的深入分析是目标流量分析,根据各个监控点上传的结构化的数据,对出现在某一监控点的目标进行统计,并通过统计得到该监控节点各个时间段内目标的流量。其中的目标可以是行人和车辆,同时可以得到目标流量的高峰期或者是低峰时期。通过计算目标流量相关数据,用来合理的提示行人和司机的,避开交通高峰期,也可以为公共资源如照明提供参考依据。In another embodiment, the further in-depth analysis performed by the cloud server on the structured data uploaded from each monitoring node is target traffic analysis. According to the structured data uploaded by each monitoring point, the Statistics are made on the target, and the traffic of the target in each time period of the monitoring node is obtained through the statistics. The targets can be pedestrians and vehicles, and the peak or low peak periods of the target traffic can be obtained at the same time. By calculating the relevant data of the target traffic, it can be used to reasonably prompt pedestrians and drivers to avoid traffic peak hours, and can also provide a reference for public resources such as lighting.
本方法通过将视频结构化处理得到对深入分析起到关键性的结构化数据,然后仅将结构化数据上传至云端,而不是将整个视频传输至云端,解决了网络传输压力大、数据流量成本高的问题。This method obtains structured data that is critical for in-depth analysis by structured video processing, and then only uploads the structured data to the cloud instead of transmitting the entire video to the cloud, which solves the problem of heavy network transmission pressure and data traffic costs high question.
在一实施例中,根据预先的设定,当各个监控节点将经过基于目标行为属性的视频结构化处理系统(以下简称:视频处理系统)处理所得结构化数据上传至云端服务器时,云端服务器在保存结构化数据之后,对结构化数据进行深入分析。In one embodiment, according to preset settings, when each monitoring node uploads the structured data processed by the video structured processing system based on the target behavior attribute (hereinafter referred to as: the video processing system) to the cloud server, the cloud server will After saving the structured data, perform in-depth analysis on the structured data.
在另一实施例中,当各个监控节点将经过视频处理系统处理所得的结构化数据上传至云端服务器时,服务器在保存结构化数据之后需要用户选择是否进行深入分析。In another embodiment, when each monitoring node uploads the structured data processed by the video processing system to the cloud server, the server requires the user to choose whether to perform in-depth analysis after saving the structured data.
在又一实施例中,当用户有需要时,可以将已经在最初上传的时候已经完成一次深入分析的结构化数据,再次重新进行设定的深入分析。In yet another embodiment, when the user needs it, the structured data that has already been analyzed in depth at the time of initial upload can be reset for another in-depth analysis.
可选地,对各个监控节点上传的结构化数据进行的深入分析进一步包括:对结构化数据进行统计、分析以得到一个或多个目标的行为类型以及异常行为,并对异常行为进行报警等,或者其他用户需要的分析处理的内容。Optionally, the in-depth analysis of the structured data uploaded by each monitoring node further includes: performing statistics and analysis on the structured data to obtain the behavior type and abnormal behavior of one or more targets, and giving an alarm to the abnormal behavior, etc., Or the content of analysis and processing required by other users.
关于如何将视频结构化数据处理,以得到结构化数据,以下详细阐述,即本申请还提供一种基于目标行为属性的视频结构化处理的方法。在一实施例中,视频结构化数据处理是利用嵌入了深度学习的目标检测识别算法、多目标跟踪算法、基于运动光流特征的异常行为识别等算法的智能分析模块,将步骤S10中读取的非结构化的视频数据转化成结构化的数据。How to process video structured data to obtain structured data will be described in detail below, that is, the present application also provides a video structured processing method based on target behavior attributes. In one embodiment, the video structured data processing is to use the intelligent analysis module embedded in deep learning algorithms such as target detection and recognition algorithms, multi-target tracking algorithms, abnormal behavior recognition based on motion optical flow features, and read in step S10 Convert unstructured video data into structured data.
参见图2,为本申请提供的一种基于目标行为属性的视频结构化处理的方法一实施例的流程示意图,该方法同时也是以上实施例的步骤S20包括步骤S22至步骤S23。Referring to FIG. 2 , it is a schematic flow chart of an embodiment of a method for video structural processing based on target behavior attributes provided by the present application. The method is also step S20 of the above embodiment, including step S22 to step S23.
S22:对单帧图片进行目标检测识别。S22: Perform target detection and recognition on a single frame picture.
可选地,步骤S22是单帧图片进行目标检测识别。其中,目标检测识别对象的包括行人检测识别、车辆检测识别和动物检测识别等。Optionally, step S22 is to perform target detection and recognition on a single frame of pictures. Wherein, target detection and recognition objects include pedestrian detection and recognition, vehicle detection and recognition, and animal detection and recognition.
可选地,步骤S22对单帧图片进行目标检测识别的步骤包括:提取单帧图片中目标的特征信息。提取单帧图片中所有目标的特征信息、目标的类别和目标的位置信息等,其中目标可以是行人、车辆和动物等。Optionally, the step S22 of performing target detection and recognition on a single frame picture includes: extracting feature information of the target in the single frame picture. Extract the feature information, target category and target location information of all targets in a single frame picture, where the targets can be pedestrians, vehicles and animals.
在一实施例中,当单帧图片中只包含行人时,目标检测识别是对行人的检测识别,即提取图片中所有行人的特征信息。In an embodiment, when only pedestrians are included in a single frame of pictures, target detection and recognition is detection and recognition of pedestrians, that is, feature information of all pedestrians in the picture is extracted.
在另一实施例中,当单帧图片中包含行人、车辆等多种类的目标时,目标检测识别是对行人、车辆等多种种类进行检测识别,即提取单帧图片中行人、车辆等的特征信息,可以理解的,目标识别的种类可以由用户的特定指定。In another embodiment, when a single frame of pictures contains multiple types of targets such as pedestrians and vehicles, the target detection and recognition is to detect and identify various types of pedestrians and vehicles, that is, to extract the pedestrians, vehicles, etc. in the single frame of pictures As for the feature information, it can be understood that the type of object recognition can be specified by the user.
可选地,步骤S22对单帧图片进行目标检测识别所采用的算法是优化后的基于深度学习的目标检测算法。具体的,可以采用YOLOV2深度学习目标检测框架进行目标检测识别,该算法的核心是利用整张图像作为作为网络输入,直接在输出层回归bounding box的位置和bounding box所属的类别。Optionally, the algorithm used in step S22 to perform object detection and recognition on a single frame of pictures is an optimized object detection algorithm based on deep learning. Specifically, the YOLOV2 deep learning target detection framework can be used for target detection and recognition. The core of the algorithm is to use the entire image as the network input, and directly return the position of the bounding box and the category to which the bounding box belongs at the output layer.
可选地,目标检测的是由模型训练和模型测试两部分构成。Optionally, target detection consists of two parts: model training and model testing.
在一实施例中,在模型训练的方面,采用的取50%的来自VOC数据集和COCO数据集的行人图像或者车辆图像,剩下的50%的数据取自真实的街道、室内通道、广场等监控数据。可以理解的是,模型训练的中的所采用的公共数据集上(VOC数据集和COCO数据集)数据与真实的监控数据集中的数据的比例可以根据需要进行调整的,其中当公共数据集中的数据所取的比例越高,相对来说,所得数据模型在真是监控场景下的精度就会相对略差,反之,当真实的监控数据集中所取的比例越高,精度会相对来说有提高。In one embodiment, in terms of model training, 50% of the pedestrian images or vehicle images from the VOC dataset and COCO dataset are used, and the remaining 50% of the data are taken from real streets, indoor passages, and squares. and other monitoring data. It can be understood that the ratio of the data in the public data set (VOC data set and COCO data set) used in model training to the data in the real monitoring data set can be adjusted as needed, wherein when the data in the public data set The higher the proportion of the data, relatively speaking, the accuracy of the obtained data model in the real monitoring scene will be relatively slightly worse. Conversely, when the proportion of the real monitoring data set is higher, the accuracy will be relatively improved. .
可选地,在一实施例中,当步骤S22在单帧图片中检测到目标后,将该行人目标放入到跟踪队列中(下文也称跟踪链)中,然后还会采用目标跟踪算法对目标进行预设的跟踪与分析。Optionally, in one embodiment, after step S22 detects the target in the single frame picture, put the pedestrian target into the tracking queue (hereinafter also referred to as the tracking chain), and then use the target tracking algorithm to Goals are tracked and analyzed by default.
可选地,上述提取单帧图片中目标的特征信息的步骤之前进一步包括:构建元数据结构。可选地,目标的特征信息是根据元数据结构进行提取,即根据元数据结构提取单帧图片中的目标的特征信息。在一实施例中,元数据结构包括行人的基本属性单元,如:摄像头地址、目标进出摄像头的时间、目标在当前监控节点的轨迹信息、目标穿着的颜色或者目标的截图中的至少一种。例如,行人的元数据结构可以参见下表1所示,其中元数据结构还可以包括其他用户所需但下表中未包含的信息。Optionally, before the above step of extracting the feature information of the target in the single frame picture, it further includes: constructing a metadata structure. Optionally, the feature information of the target is extracted according to the metadata structure, that is, the feature information of the target in the single frame picture is extracted according to the metadata structure. In one embodiment, the metadata structure includes basic attribute units of pedestrians, such as at least one of: camera address, time when the target enters and exits the camera, track information of the target at the current monitoring node, color worn by the target, or a screenshot of the target. For example, the metadata structure of pedestrians can be seen in Table 1 below, where the metadata structure can also include information required by other users but not included in the table below.
可选地,在一实施例中,为了节省网络传输的资源,元数据结构中只包含一些基本的属性信息,其他属性可以通过目标轨迹等相关信息进行挖掘计算即可得到。Optionally, in an embodiment, in order to save resources for network transmission, the metadata structure only contains some basic attribute information, and other attributes can be obtained by mining and calculating relevant information such as target trajectories.
表1行人的元数据结构Table 1 Pedestrian metadata structure
在另一实施例中,元数据结构还可以包括车辆的基本属性信息,如:摄像头地址、目标进出摄像头的时间、目标在当前监控节点的轨迹信息、目标的外观颜色、目标的车牌号或者是目标的截图中的至少一种。In another embodiment, the metadata structure can also include basic attribute information of the vehicle, such as: camera address, time when the target enters and exits the camera, track information of the target at the current monitoring node, appearance color of the target, license plate number or At least one of the screenshots of the target.
可以理解的是,元数据结构具体包含的信息和元数据的数据类型的定义是根据需要进行初始设定,也可以是在初始设定之后根据用户的需要在已设定的众多信息中特别指定需要获取的特定属性信息。It can be understood that the definition of the information contained in the metadata structure and the data type of the metadata is initially set according to the needs, or it can be specially specified in the numerous information that has been set according to the needs of the user after the initial setting The specific attribute information that needs to be obtained.
在一实施例中,元数据的结构初始设定的是摄像头地址、目标进出摄像头的时间、目标在当前监控节点的轨迹信息、目标穿着的颜色或者目标的截图等类别,在进行目标识别时,用户可以根据自己的需要特别指定获取目标进出摄像头的时间。In one embodiment, the metadata structure is initially set to the camera address, the time when the target enters and exits the camera, the track information of the target at the current monitoring node, the color of the target's clothing, or the screenshot of the target. When performing target recognition, Users can specify the time to acquire the target entering and exiting the camera according to their own needs.
在一实施例中,当单帧图片中的目标是行人时,根据预先设定的行人的元数据的结构进行提取行人的特征信息,即提取行人进出摄像头的时间、行人所处当前摄像头地址、行人进出摄像头的时间、行人在当前监控节点的轨迹信息、行人穿着的颜色或者行人当前的截图中的至少一种,也可以是根据用户的特别指定的其他的目标属性信息,如行人进出摄像头的时间和行人的穿着颜色等。In one embodiment, when the target in the single frame picture is a pedestrian, the feature information of the pedestrian is extracted according to the structure of the metadata of the preset pedestrian, that is, the time when the pedestrian enters and exits the camera, the current camera address where the pedestrian is located, At least one of the time when a pedestrian enters and exits the camera, the trajectory information of the pedestrian at the current monitoring node, the color of the pedestrian's clothing, or the current screenshot of the pedestrian, or other target attribute information specially specified by the user, such as the pedestrian entering and exiting the camera. Time and the color of the pedestrian's clothing, etc.
可选地,当从单帧图片中检测识别到目标,在获取目标的特征信息的同时,从原始的视频帧中截取出目标的图像,然后利用基于yolov2(yolov2是Joseph Redmon在2016年提出的一种基于深度学习的目标检测识别的方法)的框架进行模型训练。Optionally, when the target is detected and recognized from a single frame of pictures, while acquiring the feature information of the target, the image of the target is intercepted from the original video frame, and then using A method of target detection and recognition based on deep learning) for model training.
在一实施例中,当对单帧图片进行目标检测时,所检测到的目标是行人,则从原始的视频帧中截取出检测的行人的图像,然后利用基于yolov2的框架训练好头肩、上半身、下半身检测模型将行人进行部位切分,判断其上下半身部位的衣着颜色信息,并且截取出行人的头肩图片。In one embodiment, when the target detection is performed on a single frame picture, the detected target is a pedestrian, then the image of the detected pedestrian is intercepted from the original video frame, and then the head and shoulders, The upper body and lower body detection model segments the pedestrian, judges the clothing color information of the upper and lower body parts, and intercepts the pedestrian's head and shoulders picture.
在另一实施例中,当对单帧图片进行目标检测时检测到的目标是车辆,则从原始的视频帧中截取出检测的车辆的图像,然后利用基于yolov2的框架训练好车辆的检测模型对车辆进行检测识别,判断其车身外观颜色、识别车牌信息,并且截取出车辆的图片。可以理解的是,因为识别的目标种类可以由用户设定选择,所以对车辆的检测识别由管理者决定是否进行。In another embodiment, when the detected target is a vehicle when performing target detection on a single frame picture, the image of the detected vehicle is intercepted from the original video frame, and then the vehicle detection model is trained using the framework based on yolov2 Detect and recognize the vehicle, judge the appearance color of its body, recognize the license plate information, and intercept the picture of the vehicle. It can be understood that, since the recognized target type can be set and selected by the user, it is up to the manager to decide whether to detect and recognize the vehicle.
在又一实施例中,当对单帧图片进行目标检测时检测到的目标是动物,则从原始的视频帧中截取出检测的动物的图像,然后利用基于yolov2的框架训练好动物的检测模型对动物进行检测识别,判断其外观颜色、品种等信息,并且截取出动物的图片。可以理解的是,因为识别的目标种类可以由用户设定选择,所以对动物的检测识别由用户决定是否进行。In yet another embodiment, when the detected target is an animal when performing target detection on a single frame picture, the image of the detected animal is intercepted from the original video frame, and then the detection model of the animal is trained using a framework based on yolov2 Detect and identify animals, judge their appearance, color, species and other information, and intercept pictures of animals. It can be understood that since the recognized target type can be set and selected by the user, it is up to the user to decide whether to detect and recognize the animal.
可选地,每次目标检测识别的单帧图片可以是一张,也可以是多张单帧图片同时进行。Optionally, one single-frame picture may be recognized for each target detection, or multiple single-frame pictures may be performed simultaneously.
在一实施例中,每次进行目标检测识别的单帧图片是一张,即每次只对一张单帧图片中的目标进行目标检测识别。In an embodiment, one single frame picture is used for target detection and recognition each time, that is, target detection and recognition is only performed on a target in one single frame picture each time.
在另一实施例中,每次可以对多张图片进行目标检测识别,即每次同时对多张单帧图片中的目标进行目标检测识别。In another embodiment, target detection and recognition may be performed on multiple pictures at a time, that is, target detection and recognition may be performed on targets in multiple single-frame pictures at the same time.
可选地,对基于yolov2的框架所进行模型训练后对检测到的目标进行ID(IDentity)标号,以方便在后续跟踪时进行关联。其中,不同的目标的类别的ID号可以预先设定,且ID号的上限是由用户设定。Optionally, after model training based on the yolov2 framework, ID (IDentity) labeling is carried out on the detected targets, so as to facilitate association during follow-up tracking. Wherein, the ID numbers of different target categories can be preset, and the upper limit of the ID numbers is set by the user.
可选地,对检测识别到的目标自动进行ID标号,也可以是人为进行ID标号。Optionally, the ID labeling is automatically performed on the detected and identified targets, or the ID labeling can be performed manually.
在一实施例中,对检测识别到的目标进行标号,其中,根据检测目标的类别的而定,标记的ID号会有差距,例如行人的ID号可以设定为:数字+数字,车辆:大写字母+数字,动物:小写的字母+数字,方便在后续跟踪时进行关联。其中的设定的规则可以根据用户的习惯和喜好设定,在此不一一赘述。In one embodiment, the detected and identified targets are labeled, wherein, depending on the type of the detected target, there will be gaps in the ID numbers of the tags. For example, the ID number of a pedestrian can be set as: number + number, and the vehicle: Uppercase letters + numbers, animals: lowercase letters + numbers, for easy association during follow-up. The setting rules can be set according to the habits and preferences of the user, and will not be repeated here.
在另一实施例中,对检测识别到的目标进行标号,其中,根据检测到的目标的类别的而定,对目标所标记的ID号所属于的区间不同。例如,将所检测的行人目标的ID标号设定在区间1到1000000,将所检测到的车辆目标的ID标号设定在区间1000001到2000000。具体的,可以根据初始设定人员设定而定,也可以根据需要进行调整和改变的。In another embodiment, the detected and identified targets are marked, wherein, depending on the category of the detected targets, the ID numbers marked on the targets belong to different intervals. For example, the ID number of the detected pedestrian object is set in the interval 1 to 1,000,000, and the ID number of the detected vehicle object is set in the interval 1,000001 to 2,000,000. Specifically, it can be determined according to the initial setting personnel setting, and can also be adjusted and changed according to needs.
可选地,对检测的目标进行ID标号,可以是由预先设定由系统自动完成,也可以是由用户进行手动ID标号。Optionally, the ID labeling of the detected targets can be done automatically by the system in advance, or can be done manually by the user.
在一实施例中,当在单帧图片中检测识别到行人或者是车辆的目标时,系统会自动将所检测到的目标,根据检测的目标的类别,并接着之前已经标号的ID号自动进行ID标号。In one embodiment, when a pedestrian or a vehicle target is detected in a single frame picture, the system will automatically convert the detected target according to the type of the detected target, and then automatically carry out the ID number that has been marked before. ID number.
在另一实施例中,用户手动对图片中的目标进行ID标号。可以是对没有经过系统自动ID标号的单帧图片目标进行ID标号,也可以是遗漏的目标或者是其他在预先设定的检测目标类别之外的目标,可以由用户自主进行ID标号。In another embodiment, the user manually IDs the objects in the picture. ID labeling can be performed on single-frame image targets that have not been automatically ID-labeled by the system, or missed targets or other targets outside the preset detection target category, and ID labeling can be performed by the user independently.
可选地,参见图3,在一实施例中,在步骤S22对单帧图片进行目标检测识别之前还包括:Optionally, referring to FIG. 3 , in one embodiment, before performing target detection and recognition on a single frame picture in step S22, the method further includes:
S21:将视频切分成单帧图片。S21: Divide the video into single-frame pictures.
可选地,步骤将视频切分成单帧图片是将步骤S10中读取的视频切分成单帧图片,为步骤S22做准备。Optionally, the step of dividing the video into single-frame pictures is to divide the video read in step S10 into single-frame pictures in preparation for step S22.
可选地,在一实施例中,将视频切分成单帧图片的步骤是将步骤S10中读取的视频等间距跳帧或不等间距跳帧的切分。Optionally, in an embodiment, the step of dividing the video into single-frame pictures is to divide the video read in step S10 into frames skipped at equal intervals or skipped at unequal intervals.
在一实施例中,将视频切分成单帧图片的步骤是将步骤S10中读取的视频等间距的跳帧的切分,所跳过的帧数是相同的,即等间距的跳过相同的帧数进行切分成单帧图片,其中所跳过的帧数是不包含重要信息的帧数,即可以忽略的帧数。例如,等间距的中间跳过1帧,进行视频切分,即取第t帧,第t+2帧,第t+4帧,所跳过的帧数是第t+1帧,第t+3帧,上述所跳过的帧数是经过判断不包含的重要信息的帧数,或者是上述所跳过的帧数的是与所取的帧数重合的帧数或者是重合度很高的帧数。In one embodiment, the step of dividing the video into single-frame pictures is to divide the video read in step S10 with equidistant frame skipping, and the number of skipped frames is the same, that is, the equidistant skipping is the same The number of frames is divided into single-frame pictures, and the number of skipped frames is the number of frames that do not contain important information, that is, the number of frames that can be ignored. For example, one frame is skipped in the middle of the equidistant interval, and the video is divided, that is, the tth frame, the t+2th frame, and the t+4th frame are taken, and the number of skipped frames is the t+1th frame, the t+th frame 3 frames, the number of skipped frames above is the number of frames that are judged not to contain important information, or the number of frames skipped above is the number of frames that coincide with the number of frames taken or have a high degree of overlap number of frames.
在另一实施例中,将视频切分成单帧图片的步骤是将步骤S10中读取的视频不等间距的跳帧的切分,即所跳过的帧数可以是不相同的,不等间距的跳过不同的帧数进行切分成单帧图片,其中所跳过的帧数是不包含重要信息的帧数,即是可以忽略的帧数,其中不包含重要信息的帧数是经过判定,且判定结果确实是不重要的帧数。例如,不等间距的跳帧切分,即取第t帧,然后跳过2帧取t+3帧,再跳过1帧取t+5帧,再跳过3帧取t+9帧,其中,所跳过的帧数分别有t+1帧、t+2帧、t+4帧、t+6帧、t+7帧、t+8帧等帧数,上述跳过的帧数是经过判断没有包含此次分析所需的信息的帧数。In another embodiment, the step of dividing the video into single-frame pictures is to divide the video read in step S10 into skipped frames at unequal intervals, that is, the number of skipped frames may be different, ranging from The number of skipped frames of the spacing is divided into single-frame pictures. The number of skipped frames is the number of frames that do not contain important information, that is, the number of frames that can be ignored. The number of frames that do not contain important information is determined , and the judgment result is indeed an unimportant number of frames. For example, frame skipping at unequal intervals, that is, take the tth frame, then skip 2 frames and get t+3 frames, then skip 1 frame and get t+5 frames, then skip 3 frames and get t+9 frames, Among them, the number of skipped frames includes t+1 frame, t+2 frame, t+4 frame, t+6 frame, t+7 frame, t+8 frame, etc., and the number of skipped frames mentioned above is The number of frames judged not to contain the information required for this analysis.
在不同的实施例中,将视频切分成单帧图片的步骤可以是由系统自动将读取的视频切分成单帧图片,也可以是由用户选择是否将视频切分成单帧图片,还可以是用户手动输入已经预先完成切分的单帧图片。In different embodiments, the step of dividing the video into single-frame pictures may be automatically divided into single-frame pictures by the system, or it may be selected by the user whether to divide the video into single-frame pictures, or it may be The user manually inputs a single-frame image that has been pre-segmented.
可选地,在一实施例中,将视频切分成单帧图片的步骤完成后,即完成对读入的视频切分成单帧图片时,自动对切分得到的单帧图片执行步骤S22,即对切分所得单帧图片进行目标检测识别,也可以是由用户选择决定是否要将切分所得的单帧图片进行步骤S22所述的目标检测识别。Optionally, in one embodiment, after the step of dividing the video into single-frame pictures is completed, that is, when the read-in video is divided into single-frame pictures, step S22 is automatically performed on the single-frame pictures obtained by segmentation, namely The target detection and recognition is performed on the single frame of the segmented picture, or the user may decide whether to perform the target detection and recognition described in step S22 on the single frame of the segmented picture.
可选地,在对目标进行检测识别的过程中,会对各个目标的检测识别的值按照一定的规律进行的统计计算。Optionally, in the process of detecting and identifying the targets, statistical calculations are performed on the detection and identification values of each target according to a certain rule.
在一实施例中,在步骤S22后,对检测到某一目标在当前监控节点中合计帧数(共计出现的帧数),其中检测值为A的帧数、检测值为B的帧数等等的统计(检测值可以有多种或一种,以检测结果为准),并保存统计的结果,以备调用。In one embodiment, after step S22, the total number of frames (the total number of frames that occur) in the current monitoring node is detected for a certain target, wherein the number of frames whose detection value is A, the number of frames whose detection value is B, etc. etc. (the detection value can have multiple or one type, subject to the detection result), and save the statistical results for calling.
可选地,校正的方法主要分为轨迹校正和目标属性校正。Optionally, correction methods are mainly divided into trajectory correction and target attribute correction.
可选地,经过对目标检测得到各个目标的结构化数据后,对所得结构化数据进行校正。即是对结构化数据中的误检数据的进行校正,校正是按照权重比进行投票,最终多数的概率的数据值为准确值,少数结果的数据值为误检值。Optionally, after the structured data of each target is obtained through target detection, the obtained structured data is corrected. That is to correct the false detection data in the structured data. The correction is to vote according to the weight ratio. In the end, the data value of the majority probability is the correct value, and the data value of the minority result is the false detection value.
在一实施例中,经统计计算后(调用上述统计结果),发现步骤S22中检测识别到某一目标的在当前监控节点出现的帧数为200帧,其中有180帧检测出该目标的上衣颜色为红色,20帧中检测出该目标的上衣颜色为黑色,按照权重比进行投票,最终校正该目标的准确值上衣颜色为红色,并将结构化数据中对应的值修改为红色,最终完成校正。In one embodiment, after statistical calculation (calling the above-mentioned statistical results), it is found that in step S22, the number of frames in which a certain target is detected and recognized in the current monitoring node is 200 frames, and the top of the target is detected in 180 frames. The color is red, and the coat color of the target detected in 20 frames is black, and votes are made according to the weight ratio, and the accurate value of the target is finally corrected. The coat color is red, and the corresponding value in the structured data is changed to red, and the final completion Correction.
可选地,轨迹校正具体如下:假设一个目标在某一监控场景下出现时长为T帧,故可得到其轨迹点集合为G={p1,p2,……,pN},计算轨迹点在X轴和Y轴的均值以及偏差,然后剔除异常以及噪声轨迹点,具体表达式为:Optionally, the trajectory correction is specifically as follows: Assume that a target appears in a monitoring scene for T frames, so the set of its trajectory points can be obtained as G={p1,p2,...,p N }, and the calculated trajectory points are at The mean and deviation of the X-axis and Y-axis, and then remove the abnormal and noise track points, the specific expression is:
在一实施例中,轨迹校正中剔除偏差或者均值很小的轨迹点,减少噪声点干扰。In one embodiment, during trajectory correction, trajectory points with small deviations or small average values are eliminated to reduce noise point interference.
可选地,目标属性校正具体如下:目标属性校正是基于加权判定法校正同一个目标的属性值。假设某一个目标的上衣颜色标签为label={“红色”,“黑色”,“白色”,……},即某一个属性值有T个分类。先将其转换为数字编码L=[m1,m2,m3,……,mT];然后求出频率最高的编码值x以及其频次F,最后直接输出目标的属性值Y(准确值)。具体表达式如下:Optionally, the target attribute correction is specifically as follows: the target attribute correction is to correct the attribute value of the same target based on a weighted determination method. Assume that the color label of a target's shirt is label={"red", "black", "white", ...}, that is, there are T categories for a certain attribute value. First convert it into a digital code L=[m 1 ,m 2 ,m 3 ,...,m T ]; then calculate the code value x with the highest frequency and its frequency F, and finally directly output the attribute value Y of the target (accurate value). The specific expression is as follows:
F=T-||M-mx||0 F=T-||Mm x || 0
Y=label[mx]Y=label[m x ]
上式需要满足, The above formula needs to be satisfied,
可选地,在一实施例中,本发明结合YOLO目标检测框架进行目标识别与定位,并使用GoogLeNet网络提取出每一个目标的特征向量,以便后续目标匹配。GoogLeNet是2014年Google公司提出的一个22层深的CNN神经网络,其广泛运用于图像分类、识别等领域。由于深层次深度学习网络提取的特征向量具有较好的鲁棒性、可区分性,所以上述步骤可以较好的提高后续对于目标的跟踪的准确性。Optionally, in one embodiment, the present invention combines the YOLO target detection framework for target recognition and positioning, and uses the GoogLeNet network to extract the feature vector of each target for subsequent target matching. GoogLeNet is a 22-layer deep CNN neural network proposed by Google in 2014, which is widely used in image classification, recognition and other fields. Since the feature vectors extracted by the deep-level deep learning network have better robustness and distinguishability, the above steps can better improve the accuracy of subsequent target tracking.
S23:对目标进行跟踪,以得到跟踪结果。S23: Track the target to obtain a tracking result.
可选地,对检测到的目标进行跟踪,以得到跟踪结果的步骤中,所跟踪的目标是步骤S22检测到的目标或用户特别指定的其他目标,步骤S23进一步包括:对目标进行跟踪,记录目标进入或者离开该监控节点的时间,以及目标经过的各个位置,以得到目标的运动轨迹。具体如何对目标进行跟踪,以得到跟踪结果,本申请基于此提供了一种基于KCF与Kalman的改进型多目标跟踪方法,下文将做详细阐述。Optionally, in the step of tracking the detected target to obtain the tracking result, the tracked target is the target detected in step S22 or other targets specified by the user, and step S23 further includes: tracking the target, recording The time when the target enters or leaves the monitoring node, and the various positions that the target passes through, to obtain the target's trajectory. Specifically, how to track the target to obtain the tracking result. Based on this, the present application provides an improved multi-target tracking method based on KCF and Kalman, which will be described in detail below.
另一实施例中,本申请提供的视频处理方法在以上实施例包括步骤S21、S22和S23的基础之上进一步包括步骤S24,或者该实施例仅包括步骤S21、S22和S24,参见图4和图5。可以理解的是,基于目标行为属性视频结构化处理方法(简称视频结构化处理)实现将视频数据转化为结构化的数据,其中具体的转化过程包括:目标检测识别、目标轨迹跟踪提取和目标异常行为检测。一实施例中,视频结构化处理的包括目标检测识别和目标轨迹提取。在另一实施例中,视频结构化处理包括目标检测识、目标轨迹提取和目标异常行为检测。In another embodiment, the video processing method provided by the present application further includes step S24 on the basis of steps S21, S22 and S23 in the above embodiment, or this embodiment only includes steps S21, S22 and S24, see Fig. 4 and Figure 5. It can be understood that the video structured processing method based on target behavior attributes (referred to as video structured processing) realizes the conversion of video data into structured data, wherein the specific conversion process includes: target detection and recognition, target trajectory tracking and extraction, and target anomalies Behavioral detection. In one embodiment, the video structural processing includes target detection and recognition and target trajectory extraction. In another embodiment, the video structuring process includes object detection, object trajectory extraction, and object abnormal behavior detection.
S24:对目标进行异常行为检测。S24: Perform abnormal behavior detection on the target.
可选地,步骤S24是对上述步骤S21中检测识别出的目标进行异常行为检测的操作。Optionally, step S24 is an operation of performing abnormal behavior detection on the target detected and identified in the above step S21.
可选地,异常行为检测包括行人异常行为检测和车辆异常行为检测,其中行人的异常行为包括:奔跑、打架和骚乱,交通异常行为包括:撞击和超速等。Optionally, the abnormal behavior detection includes pedestrian abnormal behavior detection and vehicle abnormal behavior detection, wherein the abnormal behavior of pedestrians includes: running, fighting and rioting, and the abnormal behavior of traffic includes: collision and speeding, etc.
通过以上方法将视频处理,以得到重要数据,进而能够避免数据量过大,大大减轻网络传输的压力。Through the above methods, the video is processed to obtain important data, which can avoid excessive data volume and greatly reduce the pressure of network transmission.
在一实施例中,当对步骤S21中检测到的行人目标进行异常行为检测时,判定一监控节点中大于等于预设数量的人发生奔跑时,可以判定发生人群骚乱。如:可以设定当步骤S24判定10人发生奔跑异常时,可以判定发生人群骚乱,其他实施例中,判定骚乱的人数阈值根据具体情况而定。In one embodiment, when the abnormal behavior detection is performed on the pedestrian targets detected in step S21, if it is determined that more than or equal to a preset number of people running in a monitoring node, it can be determined that a crowd riot has occurred. For example, it can be set that when step S24 determines that 10 people are running abnormally, it can be determined that crowd riots have occurred. In other embodiments, the threshold for determining riots depends on specific circumstances.
在另一实施例中,可以设定当步骤S24判定2辆车发生撞击异常时,可以以此判定发生交通事故,当步骤S24判定超过3辆车发生撞击异常行为时,可以判定发生重大车祸。可以理解的,判定的关于车的数量是可以的根据需要设定调整的。In another embodiment, it can be set that when step S24 determines that two vehicles have collision abnormalities, it can be determined that a traffic accident has occurred, and when step S24 determines that more than three vehicles have collision abnormal behaviors, it can be determined that a major traffic accident has occurred. It can be understood that the determined number of cars can be set and adjusted according to needs.
在又一实施例中,当步骤S24中检测出车辆的速度超过预设的速度值时,既可以判定该车辆为超速车辆,即可将该车辆的对应的视频进行截图保存,识别的车辆的信息。其中车辆的信息包括车牌号。In yet another embodiment, when it is detected in step S24 that the speed of the vehicle exceeds the preset speed value, it can be determined that the vehicle is a speeding vehicle, and the corresponding video of the vehicle can be screenshotted and saved. information. The vehicle information includes the license plate number.
可选地,一实施例中,当步骤S24检测出异常行为时,监控节点会进行声光报警处理。Optionally, in an embodiment, when an abnormal behavior is detected in step S24, the monitoring node will perform sound and light alarm processing.
在一实施例中,声光报警的内容包括播报语音提示内容:如“请大家不要拥挤,注意安全!”或其他预先设定的语音提示内容;声光报警的内容还包括:打开对应监控节点的警示灯,用以提醒过往人群和车辆,注意安全。In one embodiment, the content of the sound and light alarm includes broadcasting voice prompt content: such as "Please don't be crowded, pay attention to safety!" or other preset voice prompt content; the content of the sound and light alarm also includes: opening the corresponding monitoring node The warning lights are used to remind passing people and vehicles to pay attention to safety.
可选地,根据发生异常行为的人数的多少进行设定异常行为恶劣的等级,不同的恶劣等级对应不同的应急处理措施。异常行为的恶劣等级可以划分为黄色、橙色和红色。黄色等级的异常行为对应的应急措施是进行声光报警,橙色等级的异常行为对应的应急措施是进行声光报警的同时连线监控负责点的安保人员,红色预警的异常行为措施是进行声光报警、连线监控负责点的安保人员同时会及时线上报警。Optionally, the severe level of abnormal behavior is set according to the number of people who have abnormal behavior, and different severe levels correspond to different emergency treatment measures. The severe levels of abnormal behavior can be divided into yellow, orange and red. The emergency measure corresponding to the abnormal behavior of the yellow level is to sound and light alarm, the emergency measure corresponding to the abnormal behavior of the orange level is to perform sound and light alarm while connecting the security personnel of the responsible point for monitoring, and the abnormal behavior measure of the red warning is to carry out sound and light alarm The security personnel at the responsible point for alarm and online monitoring will also timely alarm online.
在一实施例中,当发生异常行为的人数是3人或3人以下时,设定为黄色等级的人群异常行为;当发生异常行为的人数大于3人超过小于等于5人时橙色等级的人群异常行为;当发生异常行为的人数超过5人时设定为红色等级的人群异常行为。其中,具体的设定人数可以根据实际的需要进行调整,在此不一一赘述。In one embodiment, when the number of abnormal behaviors is 3 or less, the abnormal behavior of the crowd is set as a yellow level; when the number of abnormal behaviors is greater than 3 and less than or equal to 5 people, the orange level of the crowd Abnormal behavior; when the number of abnormal behavior exceeds 5 people, the abnormal behavior of the crowd is set as red level. Wherein, the specific set number of people can be adjusted according to actual needs, and will not be repeated here.
可选地,一实施例中,对目标进行异常行为检测的步骤之后还包括以下步骤:若检测出异常行为,则将当前视频帧图像截图保存并和所检测到发生异常行为的目标的特征信息打包发送至云端服务器。Optionally, in one embodiment, after the step of detecting the abnormal behavior of the target, the following steps are further included: if the abnormal behavior is detected, the screenshot of the current video frame image is saved and combined with the detected characteristic information of the target in which the abnormal behavior occurs Packaged and sent to the cloud server.
可选地,对发生异常行为的目标的所对应的特征信息可以包括:摄像头ID,异常事件类型、异常行为发生事件、异常行为截图等等信息,也可以包括所需要的其他类型的信息。其中发送至云端服务器的异常行为的元数据结构所包含的信息包括下表2中的结构,也可以包括包括其他的类别的信息。Optionally, the feature information corresponding to the target with abnormal behavior may include: camera ID, abnormal event type, abnormal behavior occurrence event, abnormal behavior screenshot, etc., and may also include other types of information required. The information contained in the metadata structure of the abnormal behavior sent to the cloud server includes the structure in Table 2 below, and may also include other types of information.
表2异常行为的元数据结构Table 2 Metadata structure of abnormal behavior
在一实施例中,对目标进行异常行为检测时,检测出有行人发送打架的异常行为,则将对应的当前视频帧图像截图保存,并将截图及发生异常行为的目标所对应的结构化数据一起打包发送至云端服务器。在将所检测到的异常行为的截图发送至云端服务器的同时,这一监控节点进行声光报警处理,并根据异常行为的等级启动对应的应急措施。In one embodiment, when an abnormal behavior of a target is detected, if an abnormal behavior of a pedestrian sending a fight is detected, the corresponding screenshot of the current video frame image is saved, and the structured data corresponding to the screenshot and the target with the abnormal behavior Package them together and send them to the cloud server. While sending the screenshots of detected abnormal behaviors to the cloud server, the monitoring node performs sound and light alarm processing, and initiates corresponding emergency measures according to the level of abnormal behaviors.
在另一实施例中,在对目标进行异常行为检测时,检测出发生人群骚乱时,将当前视频帧图像截图保存并发送至云端服务器,以备云端服务器进行进一步的处理,同时监控节点进行声光报警,并根据异常行为的等级启动对应的应急措施。In another embodiment, when the abnormal behavior of the target is detected, when a crowd riot is detected, the screenshot of the current video frame image is saved and sent to the cloud server for further processing by the cloud server. Light alarm, and start corresponding emergency measures according to the level of abnormal behavior.
具体的,在一实施例中,对目标进行异常行为检测的步骤包括:提取一个或多个目标的多个特征点的光流运动信息,并根据光流运动信息进行聚类以及异常行为检测。基于此,本申请还提供一种基于聚类光流特征的异常行为检测方法,下文将做详细阐述。Specifically, in one embodiment, the step of detecting abnormal behavior of the target includes: extracting optical flow motion information of multiple feature points of one or more targets, and performing clustering and abnormal behavior detection according to the optical flow motion information. Based on this, the present application also provides a method for detecting abnormal behavior based on clustering optical flow features, which will be described in detail below.
上述基于目标行为属性视频结构化处理的方法,可以实现将非结构化的视频数据转化成结构化数据,提高食品处理分析的实时性。The above-mentioned structured video processing method based on target behavior attributes can convert unstructured video data into structured data and improve the real-time performance of food processing analysis.
参见图6,为本申请还提供的一种基于KCF与Kalman的改进型多目标跟踪方法一实施例的流程示意图,该方法同时也是以上实施例中的步骤S23,具体包括步骤S231至步骤S234。具体包括以下步骤:Referring to FIG. 6 , it is a schematic flowchart of an embodiment of an improved multi-target tracking method based on KCF and Kalman provided by the present application. This method is also step S23 in the above embodiment, specifically including steps S231 to S234. Specifically include the following steps:
S231:结合跟踪链以及上一帧图片中第一多个目标对应的检测框预测第一多个目标中各个目标在当前帧的跟踪框。S231: Combine the tracking chain and the detection frames corresponding to the first multiple targets in the last frame of the picture to predict the tracking frame of each target in the current frame of the first multiple targets.
可选地,跟踪链是根据对当前帧图片之前的所有从视频中切分所得的单帧图片或部分连续单帧图片中的多个目标跟踪计算所得的,汇集之前所有图片中的多个目标的轨迹信息和经验值。Optionally, the tracking chain is calculated based on tracking multiple targets in all single-frame pictures or partial continuous single-frame pictures before the current frame picture, and gathers multiple targets in all previous pictures The trajectory information and experience value of .
在一实施例中,跟踪链是根据对当前帧图片之前的所有图片的目标跟踪计算所得,包含当前帧图片之前的所有帧图片中的所有目标的所有的信息。In an embodiment, the tracking chain is calculated based on object tracking of all pictures before the current frame picture, and includes all information of all targets in all the frame pictures before the current frame picture.
在另一实施例中,跟踪链是根据对当前帧图片之前的部分连续的图片的目标跟踪计算所得。其中跟踪计算的连续的图片数越多,预算的准确率就越高。In another embodiment, the tracking chain is calculated based on object tracking of partially continuous pictures preceding the current frame picture. Among them, the more consecutive pictures are tracked and calculated, the higher the accuracy of the budget will be.
可选地,结合跟踪链中的目标的特征信息,以及根据上一帧图片中第一多个目标对应的检测框,预测所跟踪的第一多个目标在当前帧图片中的跟踪框,例如预测第一多个目标在当前帧中可能出现的位置。Optionally, combining the feature information of the targets in the tracking chain, and according to the detection frames corresponding to the first multiple targets in the previous frame of pictures, predict the tracking frames of the tracked first multiple targets in the current frame picture, for example Predict where the first multiple objects are likely to appear in the current frame.
在一实施例中,上述步骤可以预测第一多个目标在当前帧中的跟踪框的位置,即得到第一多个目标的预测值。In an embodiment, the above steps may predict the positions of the tracking frames of the first multiple targets in the current frame, that is, obtain the predicted values of the first multiple targets.
在另一实施例中,上述步骤可以预测第一多个目标在当前帧的下一帧中的跟踪框的位置。其中,所预测的第一多个目标在当前帧的下一帧的跟踪框的位置相比于,所预测的第一多个目标在当前帧中的跟踪框的位置的误差要略大。In another embodiment, the above steps may predict the positions of the tracking frames of the first multiple objects in the frame next to the current frame. Wherein, the predicted positions of the tracking frames of the first multiple targets in the next frame of the current frame have a slightly larger error than the predicted positions of the tracking frames of the first multiple targets in the current frame.
可选地,第一多个目标是指上一帧图片中的所有检测到的目标。Optionally, the first multiple targets refer to all detected targets in the previous frame of pictures.
S232:获取上一帧图片中的第一多个目标在当前帧中对应的跟踪框,以及当前帧图片中第二多个目标的检测框。S232: Obtain the tracking frames corresponding to the first plurality of targets in the previous frame of the picture in the current frame, and the detection frames of the second plurality of targets in the current frame of the picture.
具体的,第二多个目标是指当前帧图片中的所检测到的所有目标。Specifically, the second plurality of targets refers to all detected targets in the current frame picture.
可选地,获取上一帧图片中的第一多个目标在当前帧中对应的跟踪框,以及当前帧图片中第二多个目标的检测框。其中跟踪框是在预测第一多个目标在当前帧中将会出现的位置时的矩形框,或者其他形状的框,框中包括一个或多个目标。Optionally, the tracking frames corresponding to the first multiple targets in the previous frame picture in the current frame, and the detection frames of the second multiple targets in the current frame picture are acquired. The tracking frame is a rectangular frame when predicting the position where the first multiple targets will appear in the current frame, or a frame of other shapes, and the frame includes one or more targets.
可选地,获取上一帧图片中的第一多个目标在当前帧中对应的跟踪框,以及当前帧图片中第二多个目标的检测框时,所获取的跟踪框和检测框包含跟踪框和检测框分别对应的目标的特征信息。例如目标的位置信息、颜色特征和纹理特征等。可选地,对应的特征信息可以由用户根据需要进行设定。Optionally, when obtaining the tracking frames corresponding to the first multiple targets in the current frame in the previous frame of pictures, and the detection frames of the second multiple targets in the current frame of pictures, the acquired tracking frames and detection frames contain tracking The feature information of the target corresponding to the frame and the detection frame respectively. For example, the location information, color features and texture features of the target. Optionally, the corresponding feature information can be set by the user according to needs.
S233:建立第一多个目标在当前帧中的跟踪框和当前帧中第二多个目标的检测框的目标关联矩阵。S233: Establish an object correlation matrix of the tracking frames of the first multiple targets in the current frame and the detection frames of the second multiple targets in the current frame.
可选地,根据步骤S232中获取的上一帧图片中的第一多个目标在当前帧中的对应的跟踪框与当前帧图片中所检测到的第二多个目标对应的检测框,建立目标关联矩阵。Optionally, according to the corresponding tracking frames in the current frame of the first multiple targets in the previous frame picture acquired in step S232 and the detection frames corresponding to the second multiple targets detected in the current frame picture, establish target incidence matrix.
在一实施例中,例如上一帧图片中第一多个目标数量为N,当前帧检测到的目标数量为M,则建立一个大小M×N的目标关联矩阵W,其中:In one embodiment, for example, the number of the first plurality of targets in the previous frame picture is N, and the number of targets detected in the current frame is M, then a target correlation matrix W of size M×N is established, wherein:
Aij(0<i≤M;0<j≤N)的值是由dist(i,j)、IOU(i,j)、m(i,j)决定,具体来说,可表示以下公式:The value of A ij (0<i≤M;0<j≤N) is determined by dist(i,j), IOU(i,j), m(i,j), specifically, the following formula can be expressed:
其中,IW、Ih为图像帧的宽度和高度;dist(i,j)为上一帧中得到的跟踪链中第j个目标所预测的下一帧跟踪框与当前帧中检测识别得到的第i个目标的检测框的质心距离,d(i,j)为采用图像帧对角线1/2距离进行归一化后的质心距离,m(i,j)为两个目标特征向量的欧式距离,FMi、FNj为基于GoogLeNet网络所提取的特征向量,该特征向量采用CNN框架的模型进行特征提取相比传统的手工特征提取更加具有鲁棒性和可区分性。其中,归一化的目的主要是为了保证d(i,j)与IOU(i,j)对A(i,j)的影响是一致的。IOU(i,j)表示上一帧的跟踪链中第j个目标预测的在当前帧中跟踪框与当前帧中检测识别得到的第j个目标的检测框的重叠率,即上述跟踪框与检测框的交集比上其并集。IOU具体表达式为:Among them, I W , I h are the width and height of the image frame; dist(i, j) is the tracking frame of the next frame predicted by the jth target in the tracking chain obtained in the previous frame and the detection and identification in the current frame. The centroid distance of the detection frame of the i-th target, d(i, j) is the centroid distance normalized by the diagonal 1/2 distance of the image frame, and m(i, j) is the two target feature vectors The Euclidean distance of , F Mi and F Nj are the feature vectors extracted based on the GoogLeNet network. The feature vectors are extracted using the model of the CNN framework, which is more robust and distinguishable than the traditional manual feature extraction. Among them, the purpose of normalization is mainly to ensure that the influence of d(i, j) and IOU(i, j) on A(i, j) is consistent. IOU(i, j) represents the overlap rate of the tracking frame in the current frame predicted by the jth target in the tracking chain of the previous frame and the detection frame of the jth target detected and recognized in the current frame, that is, the above tracking frame and The intersection of detection boxes is compared to their union. The specific expression of IOU is:
可选地,IOU(i,j)其取值范围为0≤IOU(i,j)≤1,该值越大,表明上述跟踪框与检测框重叠率越大。Optionally, the value range of IOU(i, j) is 0≦IOU(i, j)≦1, and the larger the value is, the larger the overlap rate between the tracking frame and the detection frame is.
在一实施例中,当目标静止的时候,同一目标在前后两帧所检测出的质心位置应该是在同一个点或者偏差很小,因此IOU的值应该近似为1,d(i,j)也应该趋于0,故Aij的值较小,且当目标匹配时,m(i,j)的取值较小,因此在进行匹配的时候跟踪链中ID=j的目标与检测链ID=i的检测目标匹配成功的可能性就越大;若前后两帧同一个目标检测框的位置相差很远,没有重叠,则IOU应该为0,m(i,j)取值较大,故d(i,j)的值就越大,因此跟踪链中ID=j的目标与检测链ID=i的检测目标匹配成功的可能性就越小。In one embodiment, when the target is still, the center of mass position detected by the same target in the two frames before and after should be at the same point or the deviation is small, so the value of IOU should be approximately 1, d(i, j) It should also tend to 0, so the value of A ij is small, and when the target matches, the value of m(i, j) is small, so when matching, the target with ID=j in the tracking chain and the detection chain ID = i is more likely to match the detection target successfully; if the position of the same target detection frame in the two frames before and after is far away and there is no overlap, then the IOU should be 0, and the value of m(i, j) is larger, so The larger the value of d(i, j), the less likely it is that the target with ID=j in the tracking chain will be successfully matched with the detection target with ID=i in the detection chain.
可选地,目标关联矩阵的建立参照质心距离、IOU、以及目标的特征向量欧式距离外,同时还可以参照目标的其他特征信息,如:颜色特征,纹理特征等。可以理解的是,当参照的指标越多时,那么准确率看就越高,但是实时性会相应的因计算量的增加而变略有下降。Optionally, the establishment of the target correlation matrix refers to the centroid distance, IOU, and Euclidean distance of the target feature vector, and can also refer to other feature information of the target, such as: color features, texture features, etc. It is understandable that when more indicators are referenced, the accuracy rate will be higher, but the real-time performance will decrease slightly due to the increase in calculation amount.
可选地,在一实施例中,当需要保证较好的实时性时,多数情况下只参照所取的两帧图像中目标的位置信息建立目标关联矩阵。Optionally, in an embodiment, when a better real-time performance needs to be ensured, in most cases only the position information of the target in the two captured images is used to establish the target correlation matrix.
在一实施例中,参照目标的位置信息和目标的穿着颜色(也可以是目标的外观颜色)建立第一多个目标对应的跟踪框和第二多个目标对应的当前帧的检测框的目标关联矩阵。In one embodiment, the tracking frame corresponding to the first plurality of targets and the detection frame of the current frame corresponding to the second plurality of targets are established with reference to the position information of the target and the clothing color of the target (it may also be the appearance color of the target). Incidence matrix.
S234:利用目标匹配算法进行校正,以得到当前帧第一部分目标对应的实际位置。S234: Use the target matching algorithm to correct, so as to obtain the actual position corresponding to the first part of the target in the current frame.
可选地,利用目标匹配算法,根据实际检测到的目标的观测值和步骤S231中对目标检测框所对应的预测值,对目标值进行校正,以得到当前帧中第一多个目标的实际位置,也即是上一帧中的第一多个目标中同时出现在当前帧的第二多个目标的目标在当前帧中的实际位置。可以理解的,因当前帧中的第二多个目标的观测值会因为切分图片的清晰度等因素会有一定的误差,所以采用结合了跟踪链及上一帧中第一多个目标在上一帧图片中的检测框,所预测的第一多个目标在当前帧中的位置进行校正第二多个目标的实际位置。Optionally, using the target matching algorithm, the target value is corrected according to the observed value of the actually detected target and the predicted value corresponding to the target detection frame in step S231, so as to obtain the actual values of the first multiple targets in the current frame. The positions, that is, the actual positions of the targets in the current frame of the targets of the second multiple targets that simultaneously appear in the current frame among the first multiple targets in the previous frame. It is understandable that the observation value of the second multiple targets in the current frame will have certain errors due to factors such as the clarity of the segmented image, so the combination of the tracking chain and the first multiple targets in the previous frame is adopted. For the detection frame in the previous frame, the predicted positions of the first multiple targets in the current frame are corrected to the actual positions of the second multiple targets.
可选地,目标匹配算法是匈牙利算法(Hungarian),观测值是步骤S22中对目标检测识别时获得目标的特征信息,包括目标的类别和目标的位置信息等,目标的预测值是步骤S231中结合跟踪链及目标在上一帧中的位置所预测的目标在当前帧中的位置值及其他特征信息。其中,以目标的位置信息为主要判断依据,其他特征信息为次要判断依据。Optionally, the target matching algorithm is the Hungarian algorithm (Hungarian), the observed value is the characteristic information of the target obtained when the target is detected and identified in step S22, including the category of the target and the position information of the target, etc., and the predicted value of the target is the target in step S231. Combining the tracking chain and the position of the target in the previous frame to predict the position value of the target in the current frame and other feature information. Among them, the location information of the target is used as the main judgment basis, and other characteristic information is used as the secondary judgment basis.
可选地,一实施例中,将第二多个目标中的检测框,与第一多个目标在当前帧中的跟踪框匹配成功的目标定义为第一部分目标,同时第一多个目标中在当前帧的跟踪框与第二多个目标在当前帧的检测框匹配成功的也定义为第一部分目标,即匹配成功的每组跟踪框与检测框均来自同一个目标。其中,可以理解的是,第二多个目标中的检测框,与第一多个目标在当前帧中的跟踪框匹配成功是指:位置信息及其他的特征信息一一对应,或者对应的项数比较多,即对应的项数概率比较高即为匹配成功。Optionally, in an embodiment, the detection frame of the second plurality of targets matches the tracking frame of the first plurality of targets in the current frame and is defined as the first part of targets, while the first plurality of targets The tracking frame in the current frame is successfully matched with the detection frames of the second multiple targets in the current frame, which is also defined as the first part of targets, that is, each group of tracking frames and detection frames that are successfully matched are from the same target. Wherein, it can be understood that the successful matching of the detection frames in the second plurality of targets with the tracking frames of the first plurality of targets in the current frame means: one-to-one correspondence between position information and other feature information, or the corresponding item The number is relatively large, that is, the probability of the corresponding item number is relatively high, which means the matching is successful.
在另一实施例中,第一部分目标的数量小于第一多个目标,即为第一多个目标在当前帧中的跟踪框只有部分可以与第二多个目标的检测框匹配成功,还有一部分在当前帧中根据匹配依据的特征信息无法匹配成功。In another embodiment, the number of the first part of targets is less than the number of the first multiple targets, that is, only part of the tracking frames of the first multiple targets in the current frame can successfully match with the detection frames of the second multiple targets, and A part cannot be successfully matched according to the feature information of the matching basis in the current frame.
可选地,不同的实施中,当前帧中的第二多个目标的检测框和上一帧中的第一多个目标在当前帧中的跟踪框匹配成功的步骤包括:根据当前帧中的第二多个目标的检测框和上一帧中的第一多个目标在当前帧中的跟踪框的质心距离和/或重叠率判断是否匹配成功。Optionally, in different implementations, the step of successfully matching the detection frames of the second plurality of targets in the current frame with the tracking frames of the first plurality of targets in the previous frame in the current frame includes: according to the The detection frame of the second plurality of targets and the centroid distance and/or overlap rate of the tracking frame of the first plurality of targets in the previous frame in the current frame determine whether the matching is successful.
在一实施例中,当前帧中的第二多个目标中的某一个或多个目标的检测框和上一帧中的第一多个目标中某一个或多个目标在当前帧中的跟踪框的质心距离很近时,且重叠率很高时则判断目标匹配成功。可以理解的相邻两帧图片的切分的时间相隔非常短,即目标在这个相隔的时间里移动的距离很微小,所以此时可以判定两帧图片中的目标匹配成功。In one embodiment, the detection frame of one or more targets among the second multiple targets in the current frame and the tracking of one or more targets among the first multiple targets in the previous frame in the current frame When the centroids of the boxes are very close and the overlap rate is high, it is judged that the target matching is successful. It can be understood that the time interval between the segmentation of two adjacent frames of pictures is very short, that is, the target moves a very small distance during this time interval, so it can be determined that the target matching in the two frames of pictures is successful at this time.
可选地,第二多个目标包括第一部分目标和第二部分目标,其中,由上可知,第一部分目标为:第二多个目标中的检测框与第一多个目标在当前帧中的跟踪框匹配成功的目标。第二部分目标为:第二多个目标中的检测框,与第一多个目标在当前帧中的跟踪框未匹配成功的目标,将第二部分目标中在跟踪链中没有记载的目标定义为新增目标。可以理解的,第二部分目标中,除了新增目标还可能存在另一类目标:在第一多个目标中没有匹配成功但是在跟踪链出现过的目标。Optionally, the second plurality of targets includes a first part of targets and a second part of targets, wherein, as can be seen from the above, the first part of targets is: the detection frame in the second plurality of targets and the detection frame of the first plurality of targets in the current frame The tracking box matches the target successfully. The second part of the target is: the detection frame in the second multiple targets, the target that does not match the tracking frame of the first multiple targets in the current frame, defines the target that is not recorded in the tracking chain in the second part of the target to add a target. It can be understood that in the second part of the target, there may be another type of target besides the newly added target: the target that has not been successfully matched in the first multiple targets but has appeared in the tracking chain.
在一实施例中,第二部分目标的数量可以是0,即当前帧中的第二多个目标的检测框与第一多个目标在当前帧中的跟踪框均可以匹配成功,所以此时的第二部分目标的数量是0。In an embodiment, the number of the second part of targets may be 0, that is, the detection frames of the second multiple targets in the current frame and the tracking frames of the first multiple targets in the current frame can all match successfully, so at this time The number of targets in the second part is 0.
可选地,在利用目标匹配算法进行校正分析,以得到当前帧第一部分目标对应的实际位置的步骤之后包括:筛选出第二部分目标中的新增目标;将新增目标加入跟踪链。另一实施例中还包括:以新增目标的初始位置和/或特征信息初始化对应的滤波跟踪器。一实施例中滤波跟踪器包括卡尔曼滤波器(kalman)、核化相关滤波器(kcf)和卡尔曼滤波器与核化相关滤波器相结合的滤波器。卡尔曼滤波器、核化相关滤波器和卡尔曼滤波器与核化相关滤波器相结合的滤波器均是基于编程实现的多目标跟踪算法。其中,卡尔曼滤波器与核化相关滤波器相结合的滤波器是指结合了卡尔曼滤波器和核化相关滤波器两者的结构的算法结构所实现的滤波器结构。其他实施例中,滤波跟踪器也可以为其他类型的滤波器,只要能够实现相同的功能即可。Optionally, after the step of using the target matching algorithm to perform correction analysis to obtain the actual position corresponding to the first part of the target in the current frame includes: screening out new targets in the second part of targets; adding the new target to the tracking chain. Another embodiment further includes: initializing the corresponding filter tracker with the initial position and/or feature information of the newly added target. In one embodiment, the filter tracker includes a Kalman filter (kalman), a kernelization correlation filter (kcf), and a filter combined with a Kalman filter and a kernelization correlation filter. Kalman filter, correlating correlation filter and the combination of Kalman filter and correlating correlating filter are multi-target tracking algorithms based on programming. Wherein, the filter combining the Kalman filter and the correlating correlation filter refers to a filter structure realized by an algorithm structure combining the structures of the Kalman filter and the correlating correlation filter. In other embodiments, the filter tracker may also be other types of filters, as long as the same function can be realized.
可选地,跟踪链的数据由上一帧以及上一帧以前的所有帧的数据训练计算所得,跟踪链中的目标包括上述描述的第一部分目标以及第三部分目标。具体的,第一部分目标是指:第一多个目标中在当前帧中的跟踪框与第二多个目标中的检测框匹配成功的目标。第三部分目标是指:跟踪链中的目标与第二多个目标未匹配成功的目标。Optionally, the data of the tracking chain is calculated by training data of the previous frame and all frames before the previous frame, and the targets in the tracking chain include the first part of the targets and the third part of the targets described above. Specifically, the first part of targets refers to targets whose tracking frames in the current frame of the first multiple targets successfully match the detection frames of the second multiple targets. The third part of the target refers to: the target in the tracking chain does not successfully match the second multiple targets.
可以理解的,第三部分目标实质上是跟踪链中除去与第二多个目标匹配成功的第一部分目标之外的所有目标。It can be understood that the third part of the targets is essentially all targets in the tracking chain except the first part of targets that successfully match the second plurality of targets.
可选地,在步骤S234利用目标匹配算法进行校正分析,以得到当前帧第一部分目标对应的实际位置的步骤之后包括:第三部分目标对应的目标丢失帧数计数值加1,并在目标丢失帧数计数值大于等于预设阈值时将对应目标从跟踪链移除。可以理解的,丢失帧数计数值的预设阈值是预先设定,且可以根据需要进行调整的。Optionally, after the step of using the target matching algorithm to perform correction analysis in step S234 to obtain the actual position corresponding to the first part of the target in the current frame, it includes: adding 1 to the count value of the number of lost frames of the target corresponding to the third part of the target, and when the target is lost When the frame count value is greater than or equal to the preset threshold, the corresponding target is removed from the tracking chain. It can be understood that the preset threshold of the count value of the number of lost frames is preset and can be adjusted as required.
在一实施例中,第三部分目标中某一目标对应的丢失帧数计数值大于等于预设阈值时,将这一目标从当前的跟踪链中移除。In an embodiment, when the lost frame count value corresponding to an object in the third part of objects is greater than or equal to a preset threshold, this object is removed from the current tracking chain.
可选地,当某一目标从当前的跟踪链中移除,将该目标所对应的结构化数据上传至云端服务器,云端服务器会对结合该目标的结构化数据或者数据库中的经验值,再次对该目标进行轨迹或异常行为的深入分析。Optionally, when a target is removed from the current tracking chain, the structured data corresponding to the target is uploaded to the cloud server, and the cloud server will combine the structured data of the target or the experience value in the database, and again Perform in-depth analysis of the trajectory or anomalous behavior of the target.
其中,可以理解的,当该被从跟踪链中移除的目标所对应的结构化数据发送至云端服务器时,执行该方法的系统可以选择信任,中断云端服务器对该目标的深入分析。Wherein, it can be understood that when the structured data corresponding to the target removed from the tracking chain is sent to the cloud server, the system executing the method can choose to trust and interrupt the in-depth analysis of the target by the cloud server.
可选地,在步骤S234利用目标匹配算法进行校正分析,以得到当前帧第一部分目标对应的实际位置的步骤之后包括:第三部分目标对应的目标丢失帧数计数值加1,并在计数值小于预设阈值时,局部跟踪第三部分目标以得到当前跟踪值。Optionally, after the step S234 of using the target matching algorithm to perform correction analysis to obtain the actual position corresponding to the first part of the target in the current frame, it includes: adding 1 to the count value of the number of lost frames of the target corresponding to the third part of the target, and adding 1 to the count value When it is less than the preset threshold, the third part of the target is partially tracked to obtain the current tracking value.
进一步,一实施例中根据第三部分目标的当前跟踪值和第三部分目标对应的预测值进行校正,以得到第三部分目标的实际位置。具体的,一实施例中,当前跟踪值是由核化相关滤波器和卡尔曼滤波器与核化相关滤波器相结合的滤波器对第三部分目标进行局部跟踪时获得,预测值是卡尔曼滤波器(kalman)预测第三部分目标的位置值。Further, in an embodiment, correction is performed according to the current tracking value of the third part of the target and the corresponding predicted value of the third part of the target, so as to obtain the actual position of the third part of the target. Specifically, in one embodiment, the current tracking value is obtained when the third part of the target is partially tracked by a filter combined with a kernelization correlation filter and a Kalman filter and a kernelization correlation filter, and the predicted value is Kalman The filter (kalman) predicts the location value of the third part of the object.
可选地,对上述步骤S22中检测到的目标进行跟踪,是由卡尔曼滤波跟踪器(kalman)和核化相关滤波跟踪器(kcf)的滤波器相结合的滤波器共同完成。Optionally, the tracking of the target detected in the above step S22 is jointly completed by a filter combined with filters of a Kalman filter tracker (kalman) and a kernelization correlation filter tracker (kcf).
在一实施例中,当跟踪的目标均为可以匹配的目标时,即无疑似丢失目标时,只调用卡尔曼滤波跟踪器(kalman)既可以完成对目标的跟踪工作。In one embodiment, when the tracked targets are all targets that can be matched, that is, when no target is suspected to be lost, only the Kalman filter tracker (kalman) can be used to complete the target tracking work.
在另一实施例中,当跟踪的目标中有出现疑似丢失的目标时,调用尔曼滤波跟踪器(kalman)和核化相关滤波跟踪器(kcf)相结合的滤波器共同配合完成对目标的跟踪工作,也可以是由卡尔曼滤波跟踪器(kalman)和核化相关滤波跟踪器(kcf)先后配合完成。In another embodiment, when there is a target suspected of being lost among the tracked targets, the filter combined with the Kalman filter tracker (kalman) and the kernelization correlation filter tracker (kcf) is called to cooperate to complete the target tracking The tracking work can also be completed successively by cooperation of a Kalman filter tracker (kalman) and a kernelization correlation filter tracker (kcf).
可选地,一实施例中,步骤S234利用目标匹配算法进行校正,以得到当前帧第一部分目标对应的实际位置的步骤包括:对于第一部分目标中的各个目标,根据各个目标对应的当前帧跟踪框对应的预测值以及当前帧检测框对应的观测值进行校正,以得第一部分目标中各个目标的实际位置。Optionally, in one embodiment, step S234 uses the target matching algorithm to perform correction to obtain the actual position corresponding to the first part of the target in the current frame. The step includes: for each target in the first part of targets, tracking The predicted value corresponding to the frame and the observed value corresponding to the detection frame of the current frame are corrected to obtain the actual position of each target in the first part of targets.
在一实施例中,对于第一部分目标中各个目标在当前帧中跟踪框对应的预测值可以理解为:结合跟踪链中的经验值以及上一帧中的位置信息,预测第一部分目标中各个目标的在当前帧中位置信息,然后结合观测所得的第一部分目标在当前帧中的实际位置(即观测值),校正第一部分目标中各个目标的实际位置。这一操作用以减小因预测值或观测值的误差带来的测得各个目标实际值不准确的问题。In an embodiment, the prediction value corresponding to the tracking frame of each target in the first part of the target in the current frame can be understood as: combining the experience value in the tracking chain and the position information in the previous frame to predict each target in the first part of the target position information in the current frame, and then combined with the observed actual positions of the first part of the targets in the current frame (ie observed values), the actual positions of the targets in the first part of the targets are corrected. This operation is used to reduce the problem of inaccurate measurement of the actual value of each target caused by the error of the predicted value or the observed value.
可选地,在一实施例中,上述基于基于KCF与Kalman的改进型多目标跟踪方法可以实现对多个目标进行跟踪分析,记录目标进入该监控节点的出入时间以及在该监控场景下的每一个运动位置,从而生成一条轨迹链,可以具体清楚的反应目标在当前监控节点的运动信息。Optionally, in one embodiment, the above-mentioned improved multi-target tracking method based on KCF and Kalman can realize tracking and analysis of multiple targets, record the time when the target enters the monitoring node and each A movement position, thereby generating a trajectory chain, which can specifically and clearly reflect the movement information of the target at the current monitoring node.
参见图7,为本申请还提供的一种基于聚类光流特征的异常行为检测方法一实施例的流程示意图,该方法同时也是以上实施例的步骤24,包括步骤S241至步骤S245。具体的步骤如下:Referring to FIG. 7 , it is a schematic flowchart of an embodiment of an abnormal behavior detection method based on clustering optical flow features provided by the present application. This method is also step 24 of the above embodiment, including steps S241 to S245. The specific steps are as follows:
S241:对一个或多个目标的检测框所在区域进行光流检测。S241: Perform optical flow detection on areas where the detection frames of one or more targets are located.
可选地,在对目标进行异常行为检测之前,已经基于预设算法完成对目标的检测识别,并获取对单帧图片中的目标进行目标检测时各个目标对应的检测框以及检测框所在的位置,然后对一个或多个目标的检测框进行光流检测。其中,光流包含了目标的运动信息。可选地,预设算法可以是yolov2算法,也可以是其他具有类似功能的算法。Optionally, before detecting the abnormal behavior of the target, the detection and recognition of the target has been completed based on the preset algorithm, and the detection frame corresponding to each target and the position of the detection frame are obtained when the target in the single frame picture is detected , and then perform optical flow detection on the detection boxes of one or more objects. Among them, the optical flow contains the motion information of the target. Optionally, the preset algorithm may be the yolov2 algorithm, or other algorithms with similar functions.
可理解的,所获取的单帧图片中各个目标对应的检测框以及检测框所在的位置,因检测框的中心会和目标的重心接近重合,所以即可以此得到每一帧图像中各个行人目标又或是其他类型目标的位置信息。Understandably, the detection frame corresponding to each target in the acquired single frame picture and the position of the detection frame, because the center of the detection frame will be close to the center of gravity of the target, so it can be used to obtain each pedestrian target in each frame of image Or the location information of other types of targets.
在一实施例中,对一个或多个目标的检测框进行光流检测的实质是获取目标所对应检测框中光流点的运动信息,包括光流点的运动的速度大小和运动方向。In one embodiment, the essence of performing optical flow detection on the detection frames of one or more objects is to obtain the motion information of the optical flow points in the detection frames corresponding to the objects, including the speed and direction of the motion of the optical flow points.
可选地,光流检测是获取光流点的各个运动特征信息,是由LK(Lucas–Kanade)金字塔光流法或者其他具有相同或类似功能的流光法来完成。Optionally, the optical flow detection is to obtain the information of each motion characteristic of the optical flow point, which is completed by the LK (Lucas–Kanade) pyramid optical flow method or other flow methods with the same or similar functions.
可选地,可以每次对每帧图片中的一个目标的检测框进行光流检测,也可以同时对每帧图片中多个的目标的检测框进行光流检测,一般每次进行光流检测的目标数是依据系统初始设定而定。可以理解的是,这一设定可以根据需要进行调整设定,当需要快速的光流检测时,可以设定为同时对每帧图片中多个目标的检测框进行检测。当需要非常精细的光流检测时,可以调整设定为每次对每帧图片中的一个目标的检测框进行光流检测。Optionally, the optical flow detection can be performed on the detection frame of one target in each frame of the picture at a time, or the optical flow detection can be performed on the detection frames of multiple targets in each frame of the picture at the same time, generally the optical flow detection is performed each time The number of targets is based on the initial settings of the system. It can be understood that this setting can be adjusted and set as needed. When fast optical flow detection is required, it can be set to detect the detection frames of multiple targets in each frame of picture at the same time. When very fine optical flow detection is required, it can be adjusted to perform optical flow detection on the detection frame of a target in each frame of picture at a time.
可选地,在一实施例中,每次对连续的多帧图片中的一个目标的检测框进行光流检测,也可以是对单帧图片中的一个目标的检测框进行检测。Optionally, in an embodiment, the optical flow detection is performed on a detection frame of a target in consecutive multiple frames of pictures each time, or may be detected on a detection frame of a target in a single frame of pictures.
可选的,在另一实施例中,每次对连续的多帧图片中的多个或全部目标的检测框进行光流检测,也可以是每次对单帧图片中的多个或全部目标的检测框进行光流检测。Optionally, in another embodiment, the optical flow detection is performed on the detection frames of multiple or all targets in consecutive multi-frame pictures each time, or multiple or all targets in a single frame of pictures each time The detection frame is used for optical flow detection.
可选地,在一实施例中,在对目标进行光流检测之前,在上述步骤中先检测出目标的大致位置区域,然后直接在连续的两帧图像中的有目标出现的区域(可以理解是目标检测区域)进行光流检测。其中,进行光流检测的连续的两帧图像是大小相同的图像。Optionally, in one embodiment, before performing optical flow detection on the target, the approximate location area of the target is detected in the above steps, and then directly in the area where the target appears in the two consecutive frames of images (it can be understood is the target detection area) for optical flow detection. Wherein, two consecutive frames of images for optical flow detection are images of the same size.
可选地,在一实施例中,对目标的检测框所在区域进行光流检测可以是对一帧图片的中目标的检测框所在区域进行光流检测,然后将所得数据和信息保存在本地存储器中,再对下一帧或预设帧中的图片中的目标的检测框所在区域进行光流检测。Optionally, in an embodiment, performing optical flow detection on the region where the detection frame of the target is located may be performing optical flow detection on the region where the detection frame of the target is located in a frame of pictures, and then saving the obtained data and information in the local memory , and then perform optical flow detection on the area where the detection frame of the target in the picture in the next frame or in the preset frame is located.
在一实施例中,每次对一个目标的检测框及所在区域进行光流检测,并逐一对图片中的所有目标的检测框进行光流检测。In one embodiment, the optical flow detection is performed on the detection frame and the region of one object at a time, and the optical flow detection is performed on the detection frames of all objects in the picture one by one.
在另一实施例中,每次对一张图片中的多个目标同时进行光流检测,即可以理解每次对一张单帧图片中的所有目标或者是部分目标的检测框进行光流检测。In another embodiment, the optical flow detection is performed on multiple targets in a picture at the same time, that is, it can be understood that the optical flow detection is performed on all targets or the detection frames of some targets in a single frame of the picture each time.
在又一实施例中,每次对多张单帧图片中的所有目标的检测框的进行光流检测。In yet another embodiment, the optical flow detection is performed on the detection frames of all objects in multiple single-frame pictures each time.
在再一实施例中,每次对多张单帧图片中,特别指定的同一类别的目标检测框进行光流检测。In yet another embodiment, the optical flow detection is performed on specially designated target detection frames of the same category in multiple single-frame pictures each time.
可选地,在步骤S241之后将所得光流信息加入到时空模型中,从而经过统计计算得到前后多帧图像的光流矢量信息。Optionally, after step S241, the obtained optical flow information is added to the space-time model, so that the optical flow vector information of multiple frames of images before and after is obtained through statistical calculation.
S242:提取至少连续两帧图像中检测框对应的特征点的光流运动信息,计算检测框所在区域的信息熵。S242: Extract the optical flow motion information of the feature points corresponding to the detection frame in at least two consecutive frames of images, and calculate the information entropy of the region where the detection frame is located.
可选地,步骤242提取至少连续两帧图像中检测框对应的特征点的光流运动信息,计算检测框所在区域的信息熵,,是对至少连续两帧图像中的检测框区域对应的特征点进行计算,其中光流运动信息是指光流点的运动方向和运动速度的大小,即提取光流点的运动方向和运动的距离,然后计算光流点的运动速度,特征点是可以代表物体特征信息的一个或多个像素点的集合。Optionally, step 242 extracts the optical flow motion information of the feature points corresponding to the detection frame in at least two consecutive frames of images, and calculates the information entropy of the area where the detection frame is located, which is the feature corresponding to the detection frame area in at least two consecutive frames of images points, where the optical flow motion information refers to the movement direction and speed of the optical flow point, that is, to extract the movement direction and distance of the optical flow point, and then calculate the movement speed of the optical flow point. The feature point can represent A collection of one or more pixel points of object feature information.
可选地,在提取出连续两帧图像中检测框对应的特征点的光流运动信息后,并依据所提取的光流运动信息计算检测框所在区域的信息熵,可以理解的是,信息熵时基于目标检测区域内所有光流点的光流信息计算所得。Optionally, after extracting the optical flow motion information of the feature points corresponding to the detection frame in two consecutive frames of images, and calculating the information entropy of the area where the detection frame is located according to the extracted optical flow motion information, it can be understood that the information entropy is calculated based on the optical flow information of all optical flow points in the target detection area.
可选地,步骤242提取至少连续两帧图像中检测框对应的特征点的光流运动信息,计算检测框所在区域的信息熵,是LK(Lucas–Kanade)金字塔光流法(LK金字塔光流法在下文中简称LK光流法)提取相邻帧仅含有行人目标的矩形框区域内的像素光流特征信息并且利用图形处理器(Graphics Processing Unit)对LK光流提取算法进行加速,从而实现实时在线提取像素的光流特征信息。其中,光流特征信息,是指光流矢量信息,可简称光流矢量。Optionally, step 242 extracts the optical flow motion information of the feature points corresponding to the detection frame in at least two consecutive frames of images, and calculates the information entropy of the region where the detection frame is located, which is the LK (Lucas–Kanade) pyramid optical flow method (LK pyramid optical flow method (hereinafter referred to as the LK optical flow method) extracts the pixel optical flow feature information in the rectangular frame area containing only pedestrian targets in adjacent frames In addition, a Graphics Processing Unit is used to accelerate the LK optical flow extraction algorithm, so as to realize real-time online extraction of optical flow feature information of pixels. Wherein, the optical flow characteristic information refers to the optical flow vector information, which may be referred to as the optical flow vector.
可选地,由于光流算法提取的光流矢量是由两个二维矩阵矢量构成,即Optionally, due to the optical flow vector extracted by the optical flow algorithm is composed of two two-dimensional matrix vectors constitute, namely
其中,矩阵中各个点对应于图像中各个像素点位置;代表相邻帧中同一个像素点在X轴移动的像素间隔,代表相邻帧中同一个像素点在Y轴移动的像素间隔。Wherein, each point in the matrix corresponds to the position of each pixel in the image; Represents the pixel interval of the same pixel point moving on the X axis in adjacent frames, Represents the pixel interval of the same pixel moving on the Y axis in adjacent frames.
可选地,像素间隔是指特征点在相邻的两帧图像中移动的距离,可以由LK光流提取算法直接提取获得。Optionally, the pixel interval refers to the moving distance of the feature points in two adjacent frames of images, which can be directly extracted by the LK optical flow extraction algorithm.
在一实施例中,步骤242是对已经完成目标检测的单帧图像,且已经获取到目标检测时的检测框的图像中,各个目标的检测框所对应的特征点的光流运动信息进行计算。其中特征点也可以解释为指的是图像灰度值发生剧烈变化的点或者在图像边缘上曲率较大的点(即两个边缘的交点)。这一操作可以减少计算量,提高计算效率。In one embodiment, step 242 is to calculate the optical flow motion information of the feature points corresponding to the detection frames of each target in the single-frame image that has completed the target detection and has obtained the detection frame of the target detection . The feature point can also be interpreted as the point where the gray value of the image changes drastically or the point with a large curvature on the edge of the image (ie, the intersection point of two edges). This operation can reduce the calculation amount and improve the calculation efficiency.
可选地,步骤S242可以同时计算连续两帧图像中的所有检测框或部分检测框对应的特征点的光流信息,也可以同时计算超过两张的连续图像中所有的检测框对应的特征点的光流信息,每次计算的图像的数量的是由预先在系统的设定的,且可以根据需要设定。Optionally, step S242 can simultaneously calculate the optical flow information of all detection frames or feature points corresponding to some detection frames in two consecutive frames of images, or simultaneously calculate the feature points corresponding to all detection frames in more than two consecutive images For optical flow information, the number of images calculated each time is pre-set in the system and can be set as needed.
在一实施例中,步骤S242同时计算连续两帧图像中的所有检测框对应的特征点的光流信息。In one embodiment, step S242 simultaneously calculates the optical flow information of the feature points corresponding to all the detection frames in two consecutive frames of images.
在另一实施例中,步骤S242同时计算超过两张的连续图像中所有的检测框对应的特征点的光流信息。In another embodiment, step S242 simultaneously calculates the optical flow information of the feature points corresponding to all the detection frames in more than two consecutive images.
可选地,步骤S242可以同时计算至少连续两帧图像中的所有目标对应的检测框的光流信息,也可以是同时计算至少连续两帧图像中特别指定且相对应的目标的检测框的光流信息。Optionally, step S242 may simultaneously calculate the optical flow information of the detection frames corresponding to all targets in at least two consecutive frames of images, or simultaneously calculate the optical flow information of the detection frames of specially designated and corresponding targets in at least two consecutive frames of images stream information.
在一实施例中,步骤S242是同时计算连续至少两帧图像中的所有目标对应的检测框的光流信息,如:第t帧中和第t+1帧图像中所有目标所对应的检测框的光流信息。In one embodiment, step S242 is to simultaneously calculate the optical flow information of the detection frames corresponding to all targets in at least two consecutive frames of images, such as: the detection frames corresponding to all targets in the t-th frame and the t+1-th frame of images optical flow information.
在另一实施例中,步骤S242是同时计算至少连续两帧图像中的特别指定且相对应的目标的检测框,如:第t帧A类目标和第t+1帧图像A’类目标,ID标号为1到3的目标的所对应的检测框的光流信息,即同时提取并计算目标A1、A2、A3和其对应的目标A1’、A2’、A3’的检测框的光流信息。In another embodiment, step S242 is to simultaneously calculate the detection frame of a specially designated and corresponding target in at least two consecutive frames of images, such as: a class A target in the tth frame and a class A' target in the t+1th frame image, The optical flow information of the detection frames corresponding to the targets with ID numbers 1 to 3, that is, to simultaneously extract and calculate the target A 1 , A 2 , A 3 and their corresponding targets A 1 ′, A 2 ′, A 3 ′ The optical flow information of the detection box.
S243:根据光流运动信息和信息熵建立聚类点。S243: Establish clustering points according to the optical flow motion information and information entropy.
可选地,根据步骤S242中提取的光流运动信息和计算所得的信息熵建立聚类点。其中光流运动信息是反应光流的运动特征的信息,包括运动的方向和运动的速度大小,也可以包括其他的相关运动特征信息,信息熵是依据光流运动信息通过计算所得。Optionally, clustering points are established according to the optical flow motion information extracted in step S242 and the calculated information entropy. The optical flow motion information is information that reflects the motion characteristics of the optical flow, including the direction of motion and the speed of motion, and may also include other related motion feature information. The information entropy is calculated based on the optical flow motion information.
在一实施例中,步骤S242中提取的光流运动信息包括运动的方向、运动的距离、运动的速度大小以及其他的相关运动特征信息中至少一种。In one embodiment, the optical flow motion information extracted in step S242 includes at least one of motion direction, motion distance, motion speed, and other related motion feature information.
可选地,步骤S243根据光流运动信息和计算所得的信息熵建立聚类点之前,要先采用K-均值算法(k-mean)对光流进行聚类。其中,聚类点个数可以根据目标检测时的检测框个数确定,对光流进行聚类是依据:将运动方向和运动速度大小相同的光流点建立成聚类点。可选地,在一实施例中,K的取值范围为6~9,当然K值也可以是其他的值,在此不做赘述。Optionally, before step S243 establishes clustering points according to the optical flow motion information and the calculated information entropy, the optical flow should be clustered by using the K-means algorithm (k-mean). Among them, the number of clustering points can be determined according to the number of detection frames when the target is detected, and the basis for clustering the optical flow is to establish the optical flow points with the same motion direction and motion speed as cluster points. Optionally, in an embodiment, the value of K ranges from 6 to 9, of course, the value of K may also be other values, which will not be repeated here.
可选地,聚类点是运动方向和运动速度大小相同或近似相同的光流点的集合。Optionally, the cluster points are a set of optical flow points whose motion directions and motion speeds are the same or approximately the same.
S244:计算聚类点的动能或目标检测框所在区域的动能。具体的,以步骤S243中所建立的聚类点为单位,计算步骤S245中所建立的聚类点的动能,或同时计算目标检测框所在区域的动能。S244: Calculate the kinetic energy of the cluster points or the kinetic energy of the area where the target detection frame is located. Specifically, the kinetic energy of the cluster points established in step S245 is calculated in units of the cluster points established in step S243, or the kinetic energy of the region where the target detection frame is located is calculated at the same time.
在一实施例中,计算步骤S243中所建立的聚类点的动能或目标所在区域的动能中至少一种。可以理解的是,不同的实施例中,根据具体需求可以配置其中一种需要的计算方式,也可以同时配置计算聚类点的动能或目标所在区域的动能两种计算方式,当只需要计算其中一种时,可以手动选择不计算另一种。可选地,根据聚类点的位置利用其前后N帧的运动矢量建立一个运动时空容器,并计算出每一个聚类点所在检测区域的光流直方图(HOF)的信息熵以及聚类点集合的平均动能。In one embodiment, at least one of the kinetic energy of the cluster points established in step S243 or the kinetic energy of the area where the target is located is calculated. It can be understood that in different embodiments, one of the required calculation methods can be configured according to specific needs, and two calculation methods for calculating the kinetic energy of the cluster points or the kinetic energy of the target area can also be configured at the same time. When using one, you can manually choose not to calculate the other. Optionally, according to the position of the clustering point, use the motion vectors of N frames before and after it to establish a motion space-time container, and calculate the information entropy of the optical flow histogram (HOF) of the detection area where each clustering point is located and the clustering point The average kinetic energy of the collection.
可选地,目标检测框所在区域的动能的公式如下:Optionally, the formula for the kinetic energy of the area where the target detection frame is located is as follows:
可选地,i=0,…,k-1表示单个目标检测框所在区域中光流的序号,k表示单个目标区域的聚类后光流总个数,此外,为了方便计算,令m=1。可选地,在一实施例中,K的取值范围为6~9,当然K值也可以是其他的值,在此不做赘述。Optionally, i=0,...,k-1 represents the sequence number of the optical flow in the region where the single target detection frame is located, and k represents the total number of clustered optical flows in the single target region. In addition, for the convenience of calculation, let m= 1. Optionally, in an embodiment, the value of K ranges from 6 to 9, of course, the value of K may also be other values, which will not be repeated here.
S245:根据聚类点的动能和/或信息熵判断异常行为。S245: Judging the abnormal behavior according to the kinetic energy and/or information entropy of the cluster points.
可选地,根据步骤S244中所计算的聚类点的动能或所述目标检测框所在区域的动能判断聚类点所对应的目标是否发生异常行为,其中当目标是行人时,异常行为包括,奔跑、打架和骚乱,当目标是车辆时,异常行为包括撞击和超速。Optionally, according to the kinetic energy of the cluster point calculated in step S244 or the kinetic energy of the area where the target detection frame is located, it is judged whether the target corresponding to the cluster point has abnormal behavior, wherein when the target is a pedestrian, the abnormal behavior includes: Running, fighting and rioting, when the target is a vehicle, abnormal behavior includes hitting and speeding.
具体的,打架和奔跑两种异常行为都与目标检测框所在区域的信息熵与聚类点的动能有关。即异常行为是打架时,目标检测框所在区域的光流信息熵较大,目标所对应的聚类点的动能或目标所在区域的动能也较大。而异常行为是奔跑的时候,目标所对应的聚类点的动能或目标所在区域的动能较大,目标检测框所在区域的光流信息熵较小。当没有发生异常行为时,目标所对应检测框所在区域的光流信息熵较小,目标所对应的聚类点的动能或目标所在区域的动能也较小。Specifically, the two abnormal behaviors of fighting and running are related to the information entropy of the area where the target detection frame is located and the kinetic energy of the cluster points. That is, when the abnormal behavior is fighting, the optical flow information entropy of the area where the target detection frame is located is relatively large, and the kinetic energy of the cluster point corresponding to the target or the kinetic energy of the area where the target is located is also large. When the abnormal behavior is running, the kinetic energy of the cluster point corresponding to the target or the kinetic energy of the area where the target is located is relatively large, and the optical flow information entropy of the area where the target detection frame is located is relatively small. When no abnormal behavior occurs, the optical flow information entropy of the area where the detection frame corresponding to the target is located is small, and the kinetic energy of the cluster point corresponding to the target or the kinetic energy of the area where the target is located is also small.
可选地,一实施例中,S245根据聚类点的动能和/或信息熵判断异常行为的步骤进一步包括:若目标所对应的检测框所在区域的光流信息熵大于等于第一阈值,且目标所对应的聚类点的动能或目标检测框所在区域的动能大于等于第二阈值,则判断异常行为是打架。Optionally, in an embodiment, the step of S245 judging the abnormal behavior according to the kinetic energy and/or information entropy of the cluster points further includes: if the optical flow information entropy of the region where the detection frame corresponding to the target is located is greater than or equal to the first threshold, and If the kinetic energy of the cluster point corresponding to the target or the kinetic energy of the area where the target detection frame is located is greater than or equal to the second threshold, it is judged that the abnormal behavior is fighting.
可选地,另一实施例中,根据聚类点的动能和/或信息熵判断异常行为的步骤进一步包括:若目标所对应的检测框所在区域的信息熵大于等于第三阈值且小于第一阈值,同时目标所对应的聚类点的动能或目标检测框所在区域的动能大于第二阈值。则判断异常行为是奔跑。Optionally, in another embodiment, the step of judging the abnormal behavior according to the kinetic energy and/or information entropy of the cluster points further includes: if the information entropy of the area where the detection frame corresponding to the target is located is greater than or equal to the third threshold and less than the first Threshold, at the same time the kinetic energy of the cluster point corresponding to the target or the kinetic energy of the area where the target detection frame is located is greater than the second threshold. Then it is judged that the abnormal behavior is running.
一实施例中,例如,信息熵用H表示,动能用E表示。In one embodiment, for example, information entropy is represented by H, and kinetic energy is represented by E.
可选地,目标奔跑行为的判断公式如下:Optionally, the judgment formula of the target running behavior is as follows:
在一实施例中,本发明训练得到奔跑行为的取值范围为λ1取值为3000,其中是用来表示目标检测框所在区域的光流信息熵H和目标检测框的所在区域的动能E的比值,λ1是一个预设的动能值。In one embodiment, the present invention trains to get the result of running behavior The value range is The value of λ 1 is 3000, where is used to represent the ratio of the optical flow information entropy H of the area where the target detection frame is located to the kinetic energy E of the area where the target detection frame is located, and λ 1 is a preset kinetic energy value.
可选地,目标打架行为的判断公式:Optionally, the judgment formula for target fighting behavior:
在一实施例中,本发明训练得到打架行为的取值范围为λ2取值为3.0,其中是用来表示信息熵H和动能E的比值,λ2是一个预设的信息熵值。In one embodiment, the present invention trains to obtain the fighting behavior The value range is The value of λ 2 is 3.0, where is used to represent the ratio of information entropy H to kinetic energy E, and λ 2 is a preset information entropy value.
可选地,正常行为的判断公式:Optionally, the judgment formula for normal behavior:
在一实施例中,在本发明中,训练得到的正常行为λ3取1500,λ4取1.85,λ3是一个预设的动能值,且小于λ1,λ4是一个预设的信息熵值,且小于λ2。In one embodiment, in the present invention, the normal behavior λ 3 obtained by training is 1500, λ 4 is 1.85, λ 3 is a preset kinetic energy value, and is smaller than λ 1 , and λ 4 is a preset information entropy value, and less than λ 2 .
在一实施例中,当某一行人目标在奔跑时,该行人目标所对应的聚类点的光流动能较大,光流信息熵较小。In an embodiment, when a pedestrian object is running, the optical flow energy of the cluster point corresponding to the pedestrian object is relatively large, and the optical flow information entropy is relatively small.
可选地,当发生人群骚乱时,首先会在一张单帧图片中检测到多个行人目标,然后在对所检测的多个行人目标进行异常行为检测时,会发现多个目标的均发生了奔跑异常,此时可以判定发生人群骚乱。Optionally, when a crowd riot occurs, multiple pedestrian targets will be detected in a single frame of the picture first, and then when abnormal behavior detection is performed on the detected multiple pedestrian targets, it will be found that all of the multiple targets are running Abnormal, at this time it can be determined that a crowd riot has occurred.
在一实施例中,对一张单帧图片中所检测到的多个目标进行异常行为检测时,当有超过预设阈值数量的目标所对应的聚类点的运动动能较大,光流信息熵较小;此时可以判定可能发生了人群骚乱。In one embodiment, when abnormal behavior detection is performed on a plurality of targets detected in a single frame picture, when the kinetic energy of the cluster points corresponding to the targets exceeding the preset threshold number is relatively large, the optical flow information entropy is relatively large. Small; at this point, it can be determined that a crowd riot may have occurred.
可选地,当目标是车辆时,异常行为的判定同样是基于对于目标所对应检测框中的多数光流方向和所检测的车辆之间的距离的大小(可以从位置信息计算得出),判断是否发生撞击。可以理解的是,当两个车辆目标的检测框的多数光流方向相对,且两辆车的距离很近时,可以判断疑似发生撞击事件。Optionally, when the target is a vehicle, the determination of abnormal behavior is also based on the distance between most optical flow directions in the detection frame corresponding to the target and the detected vehicle (can be calculated from the position information), Determine if a collision has occurred. It can be understood that when most of the optical flow directions of the detection frames of two vehicle targets are opposite, and the distance between the two vehicles is very close, it can be judged that a collision event is suspected.
可选地,将步骤S245判断异常行为的结果保存,并发送至云端服务器。Optionally, the result of judging the abnormal behavior in step S245 is saved and sent to the cloud server.
上述步骤S241至步骤S245所述的方法可以有效的提高异常行为检测的效率和实时性。The methods described in steps S241 to S245 above can effectively improve the efficiency and real-time performance of abnormal behavior detection.
可选地,一实施例中,步骤S242提取至少连续两帧图像中检测框对应的特征点的光流运动信息,计算检测框所在区域的信息熵的步骤之前还包括:提取至少连续两帧图像的特征点。Optionally, in one embodiment, step S242 extracts the optical flow motion information of the feature points corresponding to the detection frame in at least two consecutive frames of images, and before the step of calculating the information entropy of the area where the detection frame is located, it also includes: extracting at least two consecutive frames of images feature points.
可选地,提取至少连续两帧图像的特征点,可以每次提取两帧连续的图像的中目标检测框的特征点,也可以是每次提取多帧(超过两帧)连续的图像中目标检测框的特征点,其中每次提取的图像的数量由初始化系统时设定,且可以根据需要进行调整。其中,特征点指的是图像灰度值发生剧烈变化的点或者在图像边缘上曲率较大的点(即两个边缘的交点)。Optionally, the feature points of at least two consecutive frames of images are extracted, the feature points of the target detection frame of two consecutive frames of images can be extracted each time, or the targets in multiple frames (more than two frames) of continuous images can be extracted each time The feature points of the detection frame, where the number of images extracted each time is set when the system is initialized, and can be adjusted as needed. Among them, the feature point refers to the point where the gray value of the image changes sharply or the point with a large curvature on the edge of the image (ie, the intersection point of two edges).
可选地,一实施例中,步骤S242提取至少连续两帧图像中检测框对应的特征点的光流运动信息,计算检测框所在区域的信息熵的步骤进一步包括:采用预设算法计算连续两帧图像中目标匹配的特征点,去除连续两帧图像中不匹配的特征点。Optionally, in one embodiment, step S242 extracts the optical flow motion information of the feature points corresponding to the detection frame in at least two consecutive frames of images, and the step of calculating the information entropy of the area where the detection frame is located further includes: using a preset algorithm to calculate two consecutive The matching feature points of the target in the frame image are removed, and the unmatched feature points in two consecutive frames of images are removed.
可选地,首先,调用图像处理函数(goodFeaturesToTrack())提取上一帧图像中已经检测到的目标区域中的特征点(也可称作Shi-Tomasi角点),然后调用LK-pyramid光流提取算法中的函数calcOpticalFlowPyrLK()计算当前帧检测到的目标与上一帧匹配的特征点,去除前后两帧中未移动的特征点,从而得到像素点的光流运动信息。其中,本实施例中的特征点可以是Shi-Tomasi角点,又或是简称角点。Optionally, first, call the image processing function (goodFeaturesToTrack()) to extract the feature points (also called Shi-Tomasi corner points) in the target area that have been detected in the previous frame image, and then call the LK-pyramid optical flow The function calcOpticalFlowPyrLK() in the extraction algorithm calculates the feature points that match the target detected in the current frame and the previous frame, and removes the feature points that have not moved in the two frames before and after, so as to obtain the optical flow motion information of the pixels. Wherein, the feature points in this embodiment may be Shi-Tomasi corner points, or corner points for short.
可选地,一实施例中,步骤S245根据光流运动信息建立聚类点的步骤之前还包括:在图像中画出特征点的光流运动方向。Optionally, in an embodiment, before the step of establishing cluster points according to the optical flow motion information in step S245, the step further includes: drawing the optical flow motion direction of the feature points in the image.
在一实施例中,根据光流运动信息建立聚类点的步骤之前还包括,在每一帧图像中画出各个特征点的光流运动方向。In an embodiment, before the step of establishing the cluster points according to the optical flow motion information, drawing the optical flow motion direction of each feature point in each frame of image.
可选的,参见图8,一实施例中,步骤S243根据光流运动信息建立聚类点的步骤之后还包括步骤S2431和步骤S2432:Optionally, referring to FIG. 8 , in an embodiment, step S243 further includes step S2431 and step S2432 after the step of establishing cluster points according to the optical flow motion information:
S2431:基于目标检测区域的位置和运动矢量建立时空容器。S2431: Establish a space-time container based on the position and motion vector of the target detection area.
可选地,基于目标检测区域即目标检测框所在的位置信息和检测框中的聚类点在前后多帧的运动矢量关系建立时空容器。Optionally, the space-time container is established based on the location information of the target detection area, that is, the position information of the target detection frame, and the motion vector relationship of the cluster points in the detection frame in multiple frames before and after.
可选地,参见图9是一实施例中的运动时空容器的示意图,其中AB是该时空容器的二维高度,BC是该时空容器的二维宽度,CE是该时空容器的深度。其中,时空容器的深度CE是视频帧数,ABCD代表时空容器的二维大小,二维大小代表目标检测时目标检测框的大小。可以理解的,时空容器的模型可以是其他的图形,当目标检测框的图形改变时,时空容器的模型会相应改变。Optionally, see FIG. 9 , which is a schematic diagram of a moving space-time container in an embodiment, where AB is the two-dimensional height of the space-time container, BC is the two-dimensional width of the space-time container, and CE is the depth of the space-time container. Among them, the depth CE of the space-time container is the number of video frames, ABCD represents the two-dimensional size of the space-time container, and the two-dimensional size represents the size of the target detection frame during target detection. It can be understood that the model of the space-time container may be other graphs, and when the graph of the target detection frame changes, the model of the space-time container will change accordingly.
可选地,在一实施例中,当目标检测框的图形发生变化,则相对应的所建立的时空容器会依据目标检测框的图形变化发生变化。Optionally, in an embodiment, when the graphic of the target detection frame changes, the corresponding created space-time container will change according to the graphic change of the target detection frame.
S2432:计算各个聚类点所对应的检测框的光流直方图的平均信息熵与平均运动动能。S2432: Calculate the average information entropy and average motion kinetic energy of the optical flow histogram of the detection frame corresponding to each cluster point.
可选地,计算各个聚类点所对应的检测框的光流直方图的平均信息熵和平均动能。光流直方图HOF(Histogram of Oriented Optical Flow)来统计光流点在某一特定方向分布的概率的示意图。Optionally, calculate the average information entropy and average kinetic energy of the optical flow histogram of the detection frame corresponding to each cluster point. Optical flow histogram HOF (Histogram of Oriented Optical Flow) is a schematic diagram of the probability of statistical distribution of optical flow points in a specific direction.
可选地,HOF的基本思想是根据各个光流点的方向值将其投影到所对应的直方图bin中,并根据该光流的幅值进行加权,在本发明中,bin的取值大小为12,其中各个光流点的运动速度大小和方向的计算公式如下所示,T是指相邻的两帧图像间隔的时间。Optionally, the basic idea of HOF is to project the direction value of each optical flow point into the corresponding histogram bin, and weight it according to the magnitude of the optical flow. In the present invention, the value of the bin is is 12, where the calculation formula of the motion speed and direction of each optical flow point is as follows, and T refers to the time interval between two adjacent frames of images.
其中,采用光流直方图,可以减少目标的尺寸、目标运动方向以及视频中的噪声等因素对目标像素的光流特征的影响。Among them, the use of the optical flow histogram can reduce the influence of factors such as the size of the target, the moving direction of the target, and the noise in the video on the optical flow characteristics of the target pixel.
可选地,不同的实施例中异常行为的种类包括打架奔跑、骚乱或者交通异常中的一种。Optionally, the type of abnormal behavior in different embodiments includes one of fighting and running, rioting or abnormal traffic.
在一实施例中,当目标是行人时,异常行为包括:打架、奔跑和骚乱。In one embodiment, when the target is a pedestrian, the abnormal behavior includes: fighting, running and rioting.
在另一实施例中,当目标是车辆时,异常行为为例如:撞击和超速。In another embodiment, when the target is a vehicle, the abnormal behaviors are, for example: crashing and speeding.
可选地,在一实施例中,计算各个聚类点所对应的检测框的光流直方图的平均信息熵与平均动能,实质上是计算前后N帧图像中各个聚类中心的光流的平均信息熵和平均动能。Optionally, in one embodiment, calculating the average information entropy and average kinetic energy of the optical flow histogram of the detection frame corresponding to each cluster point is essentially calculating the optical flow of each cluster center in the N frame images before and after Average information entropy and average kinetic energy.
上述的异常行为检测的方法,可以有效的提高现在安防的智能化,同时还可以有效的减少在异常行为检测过程中的计算量,提高系统对目标进行异常行为检测的效率、实时性和准确率。The above-mentioned abnormal behavior detection method can effectively improve the intelligence of current security, and at the same time, it can effectively reduce the amount of calculation in the process of abnormal behavior detection, and improve the efficiency, real-time and accuracy of the system for abnormal behavior detection of targets. .
可选地,对目标进行跟踪,以得到跟踪结果的步骤之后进一步包括:将已离开当前监控节点的目标对象的结构化数据发送至云端服务器。Optionally, after the step of tracking the target to obtain the tracking result, the step further includes: sending the structured data of the target object that has left the current monitoring node to the cloud server.
可选地,对目标进行跟踪时,当某一目标的特征信息尤其是位置信息在预设时间内没有进行更新,即可判定该目标已经离开当前的监控节点,将该目标的结构化数据发送至云端服务器。其中预设时间可以由用户设定,如设定5分钟或者是10分钟等,在此不一一赘述。Optionally, when tracking a target, if the characteristic information of a certain target, especially the location information, is not updated within a preset time, it can be determined that the target has left the current monitoring node, and the structured data of the target is sent to to the cloud server. The preset time can be set by the user, such as 5 minutes or 10 minutes, etc., which will not be described here.
在一实施例中,在对目标进行跟踪时,当发现某行人的位置信息即坐标值在一定的预设时间内没有进行更新,即可以判定这个行人已经离开当前的监控节点,将该行人对应的结构化数据发送至云端服务器。In one embodiment, when the target is tracked, when it is found that the position information of a pedestrian, that is, the coordinate value, has not been updated within a certain preset time, it can be determined that the pedestrian has left the current monitoring node, and the pedestrian corresponds to The structured data is sent to the cloud server.
在另一实施例中,在对目标进行跟踪时,当发现某行人或某车辆的位置坐标一直停留在监控节点的视角边缘时,即可以判定该行人或者车辆已经离开当前的监控节点,将该行人或车辆的结构化数据发送至云端服务器。In another embodiment, when the target is tracked, when the position coordinates of a pedestrian or a vehicle are found to stay at the edge of the viewing angle of the monitoring node, it can be determined that the pedestrian or vehicle has left the current monitoring node, and the The structured data of pedestrians or vehicles is sent to the cloud server.
可选地,将被判定离开当前监控节点的目标的预设特征信息(如目标属性值、运动轨迹、目标截图等及其他所需的信息)进行打包成预设的元数据结构,然后编码成预设格式发送至云端服务器,云端服务器对所接收到的打包数据进行解析,提取出目标的元数据并保存至数据库。Optionally, the preset feature information (such as target attribute value, motion track, target screenshot, etc. and other required information) of the target determined to leave the current monitoring node is packaged into a preset metadata structure, and then encoded into The preset format is sent to the cloud server, and the cloud server analyzes the received packaged data, extracts the metadata of the target and saves it to the database.
在一实施例中,将被判定离开当前节点的目标的预设特征信息打包成为预设的元数据结构,然后编码成JSON数据格式通过网络发送至云端服务器,云端服务器对接收到的JSON数据包进行解析,提取出元数据结构,然后保存至云端服务器的数据库。可以理解的,预设的特征信息可以根据需要进行调整设定,在此不做一一赘述。In one embodiment, the preset feature information of the target that is determined to leave the current node is packaged into a preset metadata structure, and then encoded into a JSON data format and sent to the cloud server through the network, and the cloud server processes the received JSON data packet Analyze, extract the metadata structure, and then save it to the database of the cloud server. It can be understood that the preset feature information can be adjusted and set according to needs, and details will not be described here.
参见图10,本发明还提供一种具有存储功能的装置400,存储有程序数据,该程序数据被执行时实现如上所述的种基于目标行为属性的视频结构化处理的方法及实施方式所描述的方法。具体的,上述具有存储功能的装置可以是存储器、个人计算机、服务器、网络设备,或者U盘等其中的一种。Referring to FIG. 10 , the present invention also provides a device 400 with a storage function, which stores program data. When the program data is executed, the above-mentioned method and implementation of video structured processing based on target behavior attributes are described Methods. Specifically, the above-mentioned device with a storage function may be one of a memory, a personal computer, a server, a network device, or a USB disk.
请参阅图11,图11是本发明一种基于目标行为属性的视频结构化处理系统的一实施例示意图,本实施例中,视频处理系统400包括:一与处理器402耦合的存储器404,处理器402在工作时执行指令以实现如以上如上所述的一种视频处理的方法及实施方式所描述的方法,并将执行指令产生的处理结果保存在存储器404中。Please refer to FIG. 11. FIG. 11 is a schematic diagram of an embodiment of a video structured processing system based on target behavior attributes in the present invention. In this embodiment, the video processing system 400 includes: a memory 404 coupled with a processor 402, processing The processor 402 executes instructions during operation to implement the method described in the above-mentioned video processing method and implementation manner, and stores the processing results generated by executing the instructions in the memory 404 .
可选地,步骤S23对目标进行跟踪,以得到跟踪结果和步骤S24对目标进行异常行为检测,均基于步骤S22对单帧图片进行目标检测识别的基础之上,才可以进行对目标的跟踪和对目标异常行为进行检测。Optionally, step S23 tracks the target to obtain the tracking result and step S24 detects the abnormal behavior of the target, all based on the target detection and recognition of the single frame picture in step S22, the target can be tracked and Detect abnormal behavior of the target.
可选地,步骤S24对目标进行异常行为检测可以在步骤S22完成之后直接进行,也可以是和步骤S23同时进行,又或者是在步骤S23之后,并基于步骤S23跟踪的结果之上进行。Optionally, the abnormal behavior detection of the target in step S24 can be performed directly after step S22 is completed, or can be performed simultaneously with step S23, or can be performed after step S23 and based on the tracking results of step S23.
可选地,当步骤S24对目标进行异常行为检测基于步骤S23对目标进行跟踪,以得到跟踪结果,对目标的异常行为的检测会更加精确。Optionally, when step S24 detects the abnormal behavior of the target and tracks the target based on step S23 to obtain a tracking result, the detection of the abnormal behavior of the target will be more accurate.
其中,步骤S21至步骤S24所述的一种基于目标行为属性的视频结构化处理的方法,可以有效的减小监控视频的网络传输的压力,有效地提高监控系统的实时性,大幅度削减数据流量费。Among them, the video structure processing method based on the target behavior attribute described in step S21 to step S24 can effectively reduce the pressure of network transmission of surveillance video, effectively improve the real-time performance of the surveillance system, and greatly reduce data Traffic fee.
可选地,对所述单帧图片进行目标检测识别的步骤,进一步包括提取出单帧图片中的目标的特征信息。可以理解的是,将读取的视频切分成多张单帧图片后,要对切分之后的单帧图片进行目标检测识别。Optionally, the step of performing target detection and recognition on the single-frame picture further includes extracting feature information of the target in the single-frame picture. It can be understood that after the read video is divided into multiple single-frame pictures, target detection and recognition should be performed on the divided single-frame pictures.
可选地,对将视频切分所得到的单帧图片中的目标的特征信息进行提取,其中目标包括行人、车辆和动物,根据需要也可以提取建筑物或者道路桥梁的特征信息。Optionally, the feature information of the target in the single-frame picture obtained by segmenting the video is extracted, where the target includes pedestrians, vehicles and animals, and the feature information of buildings or roads and bridges can also be extracted as required.
在一实施例中,当目标是行人时,提取的特征信息包括:行人的位置、行人衣着颜色、行人的性别、运动状态、运动轨迹、驻留时间等特征化信息以及其他可获取的信息。In an embodiment, when the target is a pedestrian, the extracted feature information includes: pedestrian position, pedestrian clothing color, pedestrian gender, motion state, motion trajectory, residence time and other characteristic information and other available information.
在另一实施例中,当目标是车辆时,提取的特征信息包括:车辆的型号、车身的颜色、车辆的行驶速度以及车辆的车牌号等。In another embodiment, when the target is a vehicle, the extracted feature information includes: the model of the vehicle, the color of the body, the speed of the vehicle, and the license plate number of the vehicle.
在又一实施例中,当目标是建筑物时,提取的特征信息包括:建筑物的基本信息:如建筑层高、建筑的高度、建筑的外观颜色等。In yet another embodiment, when the target is a building, the extracted feature information includes: basic information of the building: such as building floor height, building height, building appearance color, and the like.
在再一实施例中,当目标是道路桥梁时,提取的特征信息包括:道路的宽度、道路的名称、道路的限速值等信息。In yet another embodiment, when the target is a road bridge, the extracted feature information includes information such as the width of the road, the name of the road, and the speed limit of the road.
可选地,对目标进行异常行为检测的步骤包括:提取一个或多个目标的多像素点的运动矢量,并根据运动矢量之间的关系进行异常行为检测。Optionally, the step of detecting the abnormal behavior of the target includes: extracting multi-pixel motion vectors of one or more targets, and performing abnormal behavior detection according to the relationship between the motion vectors.
在一实施例中,具体细节参见,如上所述的一种异常行为检测的方法。In an embodiment, for specific details, refer to the above-mentioned method for abnormal behavior detection.
在一实施例中,初始设定在视频处理阶段获取的结构化数据包括目标的位置、目标类别、目标属性、目标运动状态、目标运动轨迹、目标驻留时间中至少一个信息。其中,可以根据用户需要调整,在视频处理阶段只获取目标的位置信息,或者是同时获得目标的位置和目标类别。可以理解的是,视频处理阶段获取信息,可以由用户来选择所需要在视频处理阶段获取的信息类别。In one embodiment, the structured data initially set to be acquired in the video processing stage includes at least one information of the target's position, target category, target attribute, target motion state, target motion trajectory, and target dwell time. Among them, it can be adjusted according to the user's needs, and only the position information of the target is obtained in the video processing stage, or the position and the target category of the target are obtained at the same time. It can be understood that the information acquired in the video processing stage may be selected by the user as the type of information to be acquired in the video processing stage.
可选地,在对视频结构化处理结束之后,将所获得的结构化数据上传至云端服务器,云端服务器会保存各个监控节点所上传的结构化数据,并对各个监控节点所上传的结构化数据进行深入分析,以得到预设的结果。Optionally, after the video structured processing is completed, the obtained structured data is uploaded to the cloud server, and the cloud server will save the structured data uploaded by each monitoring node, and the structured data uploaded by each monitoring node Perform in-depth analysis to get preset results.
可选地,云端服务器对各个监控节点所上传的结构化数据进行深入分析的步骤可以是设定由系统自动进行,也可以是由用户手动进行。Optionally, the step of the cloud server performing an in-depth analysis of the structured data uploaded by each monitoring node may be set to be performed automatically by the system, or manually by the user.
在一实施例中,预先设定云端服务器的深入分析所包括的基础分析内容,如统计行人的数量、目标轨迹分析、目标是否有异常行为发生、发生异常行为的目标的数量,同时深入分析还包括需要用户特别选择的其他内容,如目标的各个时段的比例、目标的速度等。In one embodiment, the basic analysis content included in the in-depth analysis of the cloud server is preset, such as counting the number of pedestrians, target trajectory analysis, whether the target has abnormal behavior, and the number of targets with abnormal behavior. Including other content that needs to be specially selected by the user, such as the proportion of each time period of the target, the speed of the target, and so on.
以上所述仅为本发明的实施方式,并非因此限制本发明的专利范围,凡是利用本发明说明书及附图内容所作的等效结构或等效流程变换,或直接或间接运用在其他相关的技术领域,均同理包括在本发明的专利保护范围内。The above is only the embodiment of the present invention, and does not limit the patent scope of the present invention. Any equivalent structure or equivalent process conversion made by using the description of the present invention and the contents of the accompanying drawings, or directly or indirectly used in other related technologies fields, all of which are equally included in the scope of patent protection of the present invention.
Claims (10)
- A kind of 1. method of the video structural processing based on goal behavior attribute, it is characterised in that including:Target detection identification is carried out to single frames picture;To the target into line trace, to obtain tracking result;And/orUnusual checking is carried out to the target.
- 2. the method for the video structural processing based on goal behavior attribute according to claim 1, it is characterised in that described The step of carrying out target detection identification to the single frames picture includes:Extract clarification of objective information described in the single frames picture.
- 3. the method for the video structural processing based on goal behavior attribute according to claim 2, it is characterised in that described Further comprise before the step of extracting clarification of objective information described in the single frames picture:Build metadata structure;Wherein, the clarification of objective information is extracted according to metadata structure.
- 4. the method for the video structural processing according to claim 1 based on goal behavior attribute, it is characterised in that institute State to the target into line trace, to obtain tracking result the step of further comprises:To the target into line trace, record the time of the into or out monitoring node of target, and target pass through it is each A position, to form the movement locus of the target.
- 5. the method for the video structural processing based on goal behavior attribute according to claim 1, it is characterised in that described To the target into line trace, the step of to obtain tracking result after further comprise:Current monitor node will be left The structural data of the destination object is sent to cloud server.
- 6. the method for the video structural processing based on goal behavior attribute according to claim 1, it is characterised in that described The step of carrying out unusual checking to the target includes:The light stream movable information of multiple characteristic points of the one or more targets of extraction, and according to the light stream movable information into Row cluster and unusual checking.
- 7. the method for the video structural processing based on goal behavior attribute according to claim 1, it is characterised in that described Abnormal behaviour further comprises:Run, fight, at least one of riot or traffic abnormity.
- 8. the method for the video structural processing based on goal behavior attribute according to claim 1, it is characterised in that described It is further comprising the steps of after the step of carrying out unusual checking to the target:If it is detected that the abnormal behaviour, then will Current video two field picture sectional drawing is preserved and sent to cloud server.
- 9. a kind of video structural processing system based on goal behavior attribute, it is characterised in that including what is be electrically connected with each other Processor and memory, the processor couple the memory, the processor at work execute instruction to realize such as power Profit requires 1~8 any one of them method, and the handling result that the execute instruction is produced is stored in the memory.
- 10. a kind of device with store function, it is characterised in that have program stored therein data, and described program data are performed Realize such as claim 1~8 any one of them method.
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201711055281.0A CN108009473B (en) | 2017-10-31 | 2017-10-31 | Video structured processing method, system and storage device based on target behavior attribute |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201711055281.0A CN108009473B (en) | 2017-10-31 | 2017-10-31 | Video structured processing method, system and storage device based on target behavior attribute |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| CN108009473A true CN108009473A (en) | 2018-05-08 |
| CN108009473B CN108009473B (en) | 2021-08-24 |
Family
ID=62051189
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN201711055281.0A Expired - Fee Related CN108009473B (en) | 2017-10-31 | 2017-10-31 | Video structured processing method, system and storage device based on target behavior attribute |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN108009473B (en) |
Cited By (42)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN108810616A (en) * | 2018-05-31 | 2018-11-13 | 广州虎牙信息科技有限公司 | Object localization method, image display method, device, equipment and storage medium |
| CN108900898A (en) * | 2018-08-21 | 2018-11-27 | 北京深瞐科技有限公司 | Video structural method, apparatus and system |
| CN108984799A (en) * | 2018-08-21 | 2018-12-11 | 北京深瞐科技有限公司 | A kind of video data handling procedure and device |
| CN109034124A (en) * | 2018-08-30 | 2018-12-18 | 成都考拉悠然科技有限公司 | A kind of intelligent control method and system |
| CN109146910A (en) * | 2018-08-27 | 2019-01-04 | 公安部第研究所 | A kind of video content analysis index Evaluation Method based on target positioning |
| CN109124782A (en) * | 2018-08-29 | 2019-01-04 | 合肥工业大学 | Intelligent integral cavity mirror system |
| CN109124702A (en) * | 2018-08-29 | 2019-01-04 | 合肥工业大学 | Configure the intelligent cavity mirror system of pneumoperitoneum control and central control module |
| CN109188390A (en) * | 2018-08-14 | 2019-01-11 | 苏州大学张家港工业技术研究院 | A kind of detection of moving target high-precision and method for tracing |
| CN109171605A (en) * | 2018-08-29 | 2019-01-11 | 合肥工业大学 | Intelligent edge calculations system with target positioning and hysteroscope video enhancing processing function |
| CN109171606A (en) * | 2018-08-29 | 2019-01-11 | 合肥德易电子有限公司 | Intelligent integral robot cavity mirror system |
| CN109241898A (en) * | 2018-08-29 | 2019-01-18 | 合肥工业大学 | Object localization method and system, the storage medium of hysteroscope video |
| CN109350239A (en) * | 2018-08-29 | 2019-02-19 | 合肥德易电子有限公司 | Intelligent integrated robotic endoscopy system with target positioning function |
| CN109363614A (en) * | 2018-08-29 | 2019-02-22 | 合肥德易电子有限公司 | Intelligent integrated robotic endoscopy system with high-definition video enhancement processing function |
| CN109389543A (en) * | 2018-09-11 | 2019-02-26 | 深圳大学 | Statistical method, system, computing device and storage medium for bus operation data |
| CN109658128A (en) * | 2018-11-19 | 2019-04-19 | 浙江工业大学 | A kind of shops based on yolo and centroid tracking enters shop rate statistical method |
| CN109743497A (en) * | 2018-12-21 | 2019-05-10 | 创新奇智(重庆)科技有限公司 | A kind of dataset acquisition method, system and electronic device |
| CN109784173A (en) * | 2018-12-14 | 2019-05-21 | 合肥阿巴赛信息科技有限公司 | A kind of shop guest's on-line tracking of single camera |
| CN109859250A (en) * | 2018-11-20 | 2019-06-07 | 北京悦图遥感科技发展有限公司 | A kind of aerial infrared video multi-target detection and tracking method and device |
| CN109903312A (en) * | 2019-01-25 | 2019-06-18 | 北京工业大学 | A kind of football sportsman based on video multi-target tracking runs distance statistics method |
| CN110135317A (en) * | 2019-05-08 | 2019-08-16 | 深圳达实智能股份有限公司 | Behavior monitoring and management system and method based on cooperated computing system |
| CN110163103A (en) * | 2019-04-18 | 2019-08-23 | 中国农业大学 | A kind of live pig Activity recognition method and apparatus based on video image |
| CN110378200A (en) * | 2019-06-03 | 2019-10-25 | 特斯联(北京)科技有限公司 | A kind of intelligent security guard prompt apparatus and method for of Behavior-based control feature clustering |
| CN110706247A (en) * | 2018-06-22 | 2020-01-17 | 杭州海康威视数字技术股份有限公司 | Target tracking method, device and system |
| CN110706266A (en) * | 2019-12-11 | 2020-01-17 | 北京中星时代科技有限公司 | Aerial target tracking method based on YOLOv3 |
| CN110795595A (en) * | 2019-09-10 | 2020-02-14 | 安徽南瑞继远电网技术有限公司 | Video structured storage method, device, device and medium based on edge computing |
| WO2020056903A1 (en) * | 2018-09-21 | 2020-03-26 | 北京字节跳动网络技术有限公司 | Information generating method and device |
| CN111126152A (en) * | 2019-11-25 | 2020-05-08 | 国网信通亿力科技有限责任公司 | Video-based multi-target pedestrian detection and tracking method |
| CN111209807A (en) * | 2019-12-25 | 2020-05-29 | 航天信息股份有限公司 | Yolov 3-based video structuring method and system |
| CN111277816A (en) * | 2018-12-05 | 2020-06-12 | 北京奇虎科技有限公司 | Testing method and device of video detection system |
| CN111524113A (en) * | 2020-04-17 | 2020-08-11 | 中冶赛迪重庆信息技术有限公司 | Lifting chain abnormity identification method, system, equipment and medium |
| CN111753587A (en) * | 2019-03-28 | 2020-10-09 | 杭州海康威视数字技术股份有限公司 | A fall detection method and device |
| CN111800507A (en) * | 2020-07-06 | 2020-10-20 | 湖北经济学院 | Traffic monitoring method and traffic monitoring system |
| CN111814783A (en) * | 2020-06-08 | 2020-10-23 | 三峡大学 | Accurate license plate location method based on license plate vertex offset estimation |
| CN111914839A (en) * | 2020-07-28 | 2020-11-10 | 三峡大学 | Synchronous end-to-end license plate positioning and identifying method based on YOLOv3 |
| WO2021027632A1 (en) * | 2019-08-09 | 2021-02-18 | 北京字节跳动网络技术有限公司 | Image special effect processing method, apparatus, electronic device, and computer-readable storage medium |
| CN113705643A (en) * | 2021-08-17 | 2021-11-26 | 荣耀终端有限公司 | Target detection method and device and electronic equipment |
| CN114663648A (en) * | 2022-03-11 | 2022-06-24 | 深圳市商汤科技有限公司 | Structured target detection method, device, device and storage medium |
| CN115376068A (en) * | 2022-08-19 | 2022-11-22 | 上海浦东发展银行股份有限公司 | Tracking target alarm method, device, equipment and medium |
| CN115376042A (en) * | 2022-07-27 | 2022-11-22 | 交控科技股份有限公司 | Multi-target behavior detection method, equipment and storage medium |
| CN115761896A (en) * | 2022-11-30 | 2023-03-07 | 华智生物技术有限公司 | Abnormal behavior identification method, system, equipment and medium for live pigs |
| CN118656516A (en) * | 2024-08-13 | 2024-09-17 | 浩神科技(北京)有限公司 | A video data structured processing method and system for intelligent video generation |
| CN118968044A (en) * | 2024-10-18 | 2024-11-15 | 厦门四信通信科技有限公司 | A multi-target detection and tracking statistical method and device based on edge computing |
Citations (8)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN101719216A (en) * | 2009-12-21 | 2010-06-02 | 西安电子科技大学 | Movement human abnormal behavior identification method based on template matching |
| CN101859436A (en) * | 2010-06-09 | 2010-10-13 | 王巍 | Large-amplitude regular movement background intelligent analysis and control system |
| CN101902617A (en) * | 2010-06-11 | 2010-12-01 | 公安部第三研究所 | A Device and Method for Realizing Video Structured Description Using DSP and FPGA |
| CN103941866A (en) * | 2014-04-08 | 2014-07-23 | 河海大学常州校区 | Three-dimensional gesture recognizing method based on Kinect depth image |
| CN104301697A (en) * | 2014-07-15 | 2015-01-21 | 广州大学 | An automatic detection system and method for violent incidents in public places |
| CN104915655A (en) * | 2015-06-15 | 2015-09-16 | 西安电子科技大学 | Multi-path monitor video management method and device |
| US20160021390A1 (en) * | 2014-07-15 | 2016-01-21 | Alcatel-Lucent Usa, Inc. | Method and system for modifying compressive sensing block sizes for video monitoring using distance information |
| CN106683121A (en) * | 2016-11-29 | 2017-05-17 | 广东工业大学 | Robust object tracking method in fusion detection process |
-
2017
- 2017-10-31 CN CN201711055281.0A patent/CN108009473B/en not_active Expired - Fee Related
Patent Citations (8)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN101719216A (en) * | 2009-12-21 | 2010-06-02 | 西安电子科技大学 | Movement human abnormal behavior identification method based on template matching |
| CN101859436A (en) * | 2010-06-09 | 2010-10-13 | 王巍 | Large-amplitude regular movement background intelligent analysis and control system |
| CN101902617A (en) * | 2010-06-11 | 2010-12-01 | 公安部第三研究所 | A Device and Method for Realizing Video Structured Description Using DSP and FPGA |
| CN103941866A (en) * | 2014-04-08 | 2014-07-23 | 河海大学常州校区 | Three-dimensional gesture recognizing method based on Kinect depth image |
| CN104301697A (en) * | 2014-07-15 | 2015-01-21 | 广州大学 | An automatic detection system and method for violent incidents in public places |
| US20160021390A1 (en) * | 2014-07-15 | 2016-01-21 | Alcatel-Lucent Usa, Inc. | Method and system for modifying compressive sensing block sizes for video monitoring using distance information |
| CN104915655A (en) * | 2015-06-15 | 2015-09-16 | 西安电子科技大学 | Multi-path monitor video management method and device |
| CN106683121A (en) * | 2016-11-29 | 2017-05-17 | 广东工业大学 | Robust object tracking method in fusion detection process |
Non-Patent Citations (2)
| Title |
|---|
| JOE_QUAN: "YOLO9000: Better,Faster,Stronger(YOLO9000:更好,更快,更强)", 《CSDN》 * |
| XUAN-PHUNG HUYNH: "Tracking a Human Fast and Reliably Against", 《PSIVT2015:IMAGE AND VIDEO TECHNOLOGY》 * |
Cited By (61)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN108810616A (en) * | 2018-05-31 | 2018-11-13 | 广州虎牙信息科技有限公司 | Object localization method, image display method, device, equipment and storage medium |
| US11284128B2 (en) | 2018-05-31 | 2022-03-22 | Guangzhou Huya Information Technology Co., Ltd. | Object positioning method, video display method, apparatus, device, and storage medium |
| WO2019228387A1 (en) * | 2018-05-31 | 2019-12-05 | 广州虎牙信息科技有限公司 | Target positioning method and apparatus, video display method and apparatus, device, and storage medium |
| CN110706247A (en) * | 2018-06-22 | 2020-01-17 | 杭州海康威视数字技术股份有限公司 | Target tracking method, device and system |
| CN109188390B (en) * | 2018-08-14 | 2023-05-23 | 苏州大学张家港工业技术研究院 | A high-precision detection and tracking method for moving targets |
| CN109188390A (en) * | 2018-08-14 | 2019-01-11 | 苏州大学张家港工业技术研究院 | A kind of detection of moving target high-precision and method for tracing |
| CN108900898A (en) * | 2018-08-21 | 2018-11-27 | 北京深瞐科技有限公司 | Video structural method, apparatus and system |
| CN108984799A (en) * | 2018-08-21 | 2018-12-11 | 北京深瞐科技有限公司 | A kind of video data handling procedure and device |
| CN109146910A (en) * | 2018-08-27 | 2019-01-04 | 公安部第研究所 | A kind of video content analysis index Evaluation Method based on target positioning |
| CN109146910B (en) * | 2018-08-27 | 2021-07-06 | 公安部第一研究所 | An evaluation method of video content analysis index based on target positioning |
| CN109241898A (en) * | 2018-08-29 | 2019-01-18 | 合肥工业大学 | Object localization method and system, the storage medium of hysteroscope video |
| CN109350239A (en) * | 2018-08-29 | 2019-02-19 | 合肥德易电子有限公司 | Intelligent integrated robotic endoscopy system with target positioning function |
| CN109363614A (en) * | 2018-08-29 | 2019-02-22 | 合肥德易电子有限公司 | Intelligent integrated robotic endoscopy system with high-definition video enhancement processing function |
| CN109363614B (en) * | 2018-08-29 | 2020-09-01 | 合肥德易电子有限公司 | Intelligent integrated robot cavity mirror system with high-definition video enhancement processing function |
| CN109124702B (en) * | 2018-08-29 | 2020-09-01 | 合肥工业大学 | Intelligent endoscope system with pneumoperitoneum control and central control module |
| CN109171605B (en) * | 2018-08-29 | 2020-09-01 | 合肥工业大学 | Intelligent edge computing system with target localization and endoscopic image enhancement processing |
| CN109171606A (en) * | 2018-08-29 | 2019-01-11 | 合肥德易电子有限公司 | Intelligent integral robot cavity mirror system |
| CN109171605A (en) * | 2018-08-29 | 2019-01-11 | 合肥工业大学 | Intelligent edge calculations system with target positioning and hysteroscope video enhancing processing function |
| CN109124702A (en) * | 2018-08-29 | 2019-01-04 | 合肥工业大学 | Configure the intelligent cavity mirror system of pneumoperitoneum control and central control module |
| CN109124782A (en) * | 2018-08-29 | 2019-01-04 | 合肥工业大学 | Intelligent integral cavity mirror system |
| CN109034124A (en) * | 2018-08-30 | 2018-12-18 | 成都考拉悠然科技有限公司 | A kind of intelligent control method and system |
| CN109389543B (en) * | 2018-09-11 | 2022-03-04 | 深圳大学 | Bus operation data statistical method, system, computing device and storage medium |
| CN109389543A (en) * | 2018-09-11 | 2019-02-26 | 深圳大学 | Statistical method, system, computing device and storage medium for bus operation data |
| WO2020056903A1 (en) * | 2018-09-21 | 2020-03-26 | 北京字节跳动网络技术有限公司 | Information generating method and device |
| CN109658128A (en) * | 2018-11-19 | 2019-04-19 | 浙江工业大学 | A kind of shops based on yolo and centroid tracking enters shop rate statistical method |
| CN109859250B (en) * | 2018-11-20 | 2023-08-18 | 北京悦图遥感科技发展有限公司 | Aviation infrared video multi-target detection and tracking method and device |
| CN109859250A (en) * | 2018-11-20 | 2019-06-07 | 北京悦图遥感科技发展有限公司 | A kind of aerial infrared video multi-target detection and tracking method and device |
| CN111277816B (en) * | 2018-12-05 | 2024-05-14 | 北京奇虎科技有限公司 | Method and device for testing video detection system |
| CN111277816A (en) * | 2018-12-05 | 2020-06-12 | 北京奇虎科技有限公司 | Testing method and device of video detection system |
| CN109784173A (en) * | 2018-12-14 | 2019-05-21 | 合肥阿巴赛信息科技有限公司 | A kind of shop guest's on-line tracking of single camera |
| CN109743497B (en) * | 2018-12-21 | 2020-06-30 | 创新奇智(重庆)科技有限公司 | Data set acquisition method and system and electronic device |
| CN109743497A (en) * | 2018-12-21 | 2019-05-10 | 创新奇智(重庆)科技有限公司 | A kind of dataset acquisition method, system and electronic device |
| CN109903312A (en) * | 2019-01-25 | 2019-06-18 | 北京工业大学 | A kind of football sportsman based on video multi-target tracking runs distance statistics method |
| CN111753587A (en) * | 2019-03-28 | 2020-10-09 | 杭州海康威视数字技术股份有限公司 | A fall detection method and device |
| CN111753587B (en) * | 2019-03-28 | 2023-09-29 | 杭州海康威视数字技术股份有限公司 | A method and device for detecting falling to the ground |
| CN110163103A (en) * | 2019-04-18 | 2019-08-23 | 中国农业大学 | A kind of live pig Activity recognition method and apparatus based on video image |
| CN110135317A (en) * | 2019-05-08 | 2019-08-16 | 深圳达实智能股份有限公司 | Behavior monitoring and management system and method based on cooperated computing system |
| CN110378200A (en) * | 2019-06-03 | 2019-10-25 | 特斯联(北京)科技有限公司 | A kind of intelligent security guard prompt apparatus and method for of Behavior-based control feature clustering |
| WO2021027632A1 (en) * | 2019-08-09 | 2021-02-18 | 北京字节跳动网络技术有限公司 | Image special effect processing method, apparatus, electronic device, and computer-readable storage medium |
| CN110795595B (en) * | 2019-09-10 | 2024-03-05 | 安徽南瑞继远电网技术有限公司 | Video structured storage method, device, equipment and medium based on edge calculation |
| CN110795595A (en) * | 2019-09-10 | 2020-02-14 | 安徽南瑞继远电网技术有限公司 | Video structured storage method, device, device and medium based on edge computing |
| CN111126152B (en) * | 2019-11-25 | 2023-04-11 | 国网信通亿力科技有限责任公司 | Multi-target pedestrian detection and tracking method based on video |
| CN111126152A (en) * | 2019-11-25 | 2020-05-08 | 国网信通亿力科技有限责任公司 | Video-based multi-target pedestrian detection and tracking method |
| CN110706266A (en) * | 2019-12-11 | 2020-01-17 | 北京中星时代科技有限公司 | Aerial target tracking method based on YOLOv3 |
| CN110706266B (en) * | 2019-12-11 | 2020-09-15 | 北京中星时代科技有限公司 | Aerial target tracking method based on YOLOv3 |
| CN111209807A (en) * | 2019-12-25 | 2020-05-29 | 航天信息股份有限公司 | Yolov 3-based video structuring method and system |
| CN111524113A (en) * | 2020-04-17 | 2020-08-11 | 中冶赛迪重庆信息技术有限公司 | Lifting chain abnormity identification method, system, equipment and medium |
| CN111814783A (en) * | 2020-06-08 | 2020-10-23 | 三峡大学 | Accurate license plate location method based on license plate vertex offset estimation |
| CN111814783B (en) * | 2020-06-08 | 2024-05-24 | 深圳市富浩鹏电子有限公司 | Accurate license plate location method based on license plate vertex offset estimation |
| CN111800507A (en) * | 2020-07-06 | 2020-10-20 | 湖北经济学院 | Traffic monitoring method and traffic monitoring system |
| CN111914839A (en) * | 2020-07-28 | 2020-11-10 | 三峡大学 | Synchronous end-to-end license plate positioning and identifying method based on YOLOv3 |
| CN111914839B (en) * | 2020-07-28 | 2024-03-19 | 特微乐行(广州)技术有限公司 | Synchronous end-to-end license plate positioning and recognition method based on YOLOv3 |
| CN113705643A (en) * | 2021-08-17 | 2021-11-26 | 荣耀终端有限公司 | Target detection method and device and electronic equipment |
| CN114663648A (en) * | 2022-03-11 | 2022-06-24 | 深圳市商汤科技有限公司 | Structured target detection method, device, device and storage medium |
| CN114663648B (en) * | 2022-03-11 | 2025-08-19 | 深圳市商汤科技有限公司 | Structured target detection method, device, equipment and storage medium |
| CN115376042A (en) * | 2022-07-27 | 2022-11-22 | 交控科技股份有限公司 | Multi-target behavior detection method, equipment and storage medium |
| CN115376068A (en) * | 2022-08-19 | 2022-11-22 | 上海浦东发展银行股份有限公司 | Tracking target alarm method, device, equipment and medium |
| CN115761896A (en) * | 2022-11-30 | 2023-03-07 | 华智生物技术有限公司 | Abnormal behavior identification method, system, equipment and medium for live pigs |
| CN118656516A (en) * | 2024-08-13 | 2024-09-17 | 浩神科技(北京)有限公司 | A video data structured processing method and system for intelligent video generation |
| CN118656516B (en) * | 2024-08-13 | 2024-11-26 | 浩神科技(北京)有限公司 | A video data structured processing method and system for intelligent video generation |
| CN118968044A (en) * | 2024-10-18 | 2024-11-15 | 厦门四信通信科技有限公司 | A multi-target detection and tracking statistical method and device based on edge computing |
Also Published As
| Publication number | Publication date |
|---|---|
| CN108009473B (en) | 2021-08-24 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN108009473A (en) | Based on goal behavior attribute video structural processing method, system and storage device | |
| CN108053427B (en) | An improved multi-target tracking method, system and device based on KCF and Kalman | |
| CN108052859B (en) | A method, system and device for abnormal behavior detection based on clustered optical flow features | |
| CN108062349B (en) | Video surveillance method and system based on video structured data and deep learning | |
| CN109948582B (en) | A vehicle retrograde intelligent detection method based on tracking trajectory analysis | |
| Bertini et al. | Multi-scale and real-time non-parametric approach for anomaly detection and localization | |
| CN102163290B (en) | Method for modeling abnormal events in multi-visual angle video monitoring based on temporal-spatial correlation information | |
| CN105023008B (en) | The pedestrian of view-based access control model conspicuousness and multiple features recognition methods again | |
| CN104751136B (en) | A kind of multi-camera video event back jump tracking method based on recognition of face | |
| CN111353338B (en) | A method for improving energy efficiency based on video surveillance in business halls | |
| CN113393679A (en) | Regional traffic guidance method and system based on traffic intersection traffic flow identification and statistics | |
| CN102065275B (en) | Multi-target tracking method in intelligent video monitoring system | |
| CN108230364A (en) | A kind of foreground object motion state analysis method based on neural network | |
| CN115512293B (en) | A computer vision-based intelligent analysis method and system for spatial behavior | |
| US12100214B2 (en) | Video-based public safety incident prediction system and method therefor | |
| CN115841649A (en) | Multi-scale people counting method for urban complex scene | |
| CN111091057A (en) | Information processing method and device and computer readable storage medium | |
| CN117437599B (en) | Pedestrian abnormal event detection method and system for monitoring scene | |
| Yang et al. | Traffic flow estimation and vehicle‐type classification using vision‐based spatial–temporal profile analysis | |
| CN113920585A (en) | Behavior recognition method and device, equipment and storage medium | |
| CN119672933A (en) | Real-time warning method and system for abnormal behavior based on intelligent security | |
| CN117911965B (en) | Expressway traffic accident identification method and equipment based on aerial image | |
| CN109241950A (en) | A kind of crowd panic state identification method based on enthalpy Distribution Entropy | |
| CN112001280B (en) | Real-time and online optimized face recognition system and method | |
| CN111950507B (en) | Data processing and model training method, device, equipment and medium |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| GR01 | Patent grant | ||
| GR01 | Patent grant | ||
| CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20210824 |
|
| CF01 | Termination of patent right due to non-payment of annual fee |