[go: up one dir, main page]

WO2023147376A1 - Geo-spatial context for full-motion video - Google Patents

Geo-spatial context for full-motion video Download PDF

Info

Publication number
WO2023147376A1
WO2023147376A1 PCT/US2023/061293 US2023061293W WO2023147376A1 WO 2023147376 A1 WO2023147376 A1 WO 2023147376A1 US 2023061293 W US2023061293 W US 2023061293W WO 2023147376 A1 WO2023147376 A1 WO 2023147376A1
Authority
WO
WIPO (PCT)
Prior art keywords
video
geographical area
image
area
generating
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/US2023/061293
Other languages
French (fr)
Inventor
Dimitrije MIŠIĆ
Nemanja Grujić
Milan Marković
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Cubic Corp
Original Assignee
Cubic Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US18/158,886 external-priority patent/US20230237796A1/en
Application filed by Cubic Corp filed Critical Cubic Corp
Publication of WO2023147376A1 publication Critical patent/WO2023147376A1/en
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/787Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using geographical or spatial information, e.g. location
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
    • H04N7/183Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a single remote source
    • H04N7/186Video door telephones
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10032Satellite or aerial image; Remote sensing
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30244Camera pose

Definitions

  • This disclosure generally relates to a system and a method and, not by way of limitation, to generate a composite video feed for a geographical area.
  • a method and a system for generating a composite video feed for a geographical area are disclosed.
  • a video of the geographical area, captured by a camera, of an aerial platform is received.
  • the video includes metadata indicative of location information, which is used to identify the coordinates of the geographical area.
  • An image that is adjacent to the geographical area is received from the geographical information system and is transformed according to the metadata.
  • the coordinates of the geographical area are used to determine an area with the image.
  • the video is embedded in the area by matching the area with the coordinates of the geographical area, where the edges of the video correspond to the boundaries of the area.
  • a composite video feed, including the video embedded along with the image is generated and a video player displays the composite video feed.
  • a method for generating a composite video feed for a geographical area is disclosed.
  • a video of the geographical area captured by an aerial platform is received.
  • Metadata indicative of location information associated with the video, is processed. Coordinates of the geographical area are identified based on the location information.
  • An image is received from a geographical information system (GIS), where the image is adjacent to the geographical area.
  • GIS geographical information system
  • the image is transformed according to the metadata.
  • An area within the image is determined based on the coordinates of the geographical area.
  • the video is embedded in the area by matching the area and the coordinates of the geographical area, where edges of the video correspond to boundaries of the area.
  • the composite video feed with the video embedded along with the image is generated.
  • a video player is caused to display the composite video feed.
  • the system further comprises a feed generator that is configured to:
  • FIG. 1 illustrates a schematic representation of a military environment according to an embodiment of the present disclosure
  • FIG. 2 illustrates a block diagram showing a stream processor according to an embodiment of the present disclosure
  • FIG. 4 illustrates a 3D video feed of a geographical area according to an embodiment of the present disclosure
  • FIG. 5 illustrates a composite video feed of the geographical area according to an embodiment of the present disclosure
  • FIG. 7 illustrates a flowchart of a method for generating a composite video feed according to another embodiment of the present disclosure.
  • the stream processor 110 works in combination with the GIS 150 for performing video processing and analysis in real-time
  • the GIS 150 utilizes various geospatial tools for collaborating with the stream processor 110 in order to provide GIS image and Geo-tag data.
  • the GIS 150 further includes an integrated database to provide the GIS image and the Geo-tag data in accordance with the geographical area of the video feed.
  • the communication network 120 can be a wide area network (WAN).
  • the WAN may comprise one or more public and/or private data communication networks, including the Internet, one or more cellular, radio, or other wireless WANs (WWANs), and/or the like.
  • the aerial platforms 130 includes multiple sensors to create metadata indicative of the location of the geographical area captured in the video.
  • the metadata provides geographical coordinates of the physical objects within the frame of the video.
  • the sensors are designed to generate raw sensor data.
  • the sensors are selected from a Global Navigation Satellite System GNSS) receiver (e.g., a Global Positioning System (GPS) receiver), magnetometer, altimeter, gyroscope, accelerometer, and/or the like, and may be generally indicative of a location of the geographical area in the video as well as elevation, azimuth, and orientation of the camera capturing video.
  • GNSS Global Navigation Satellite System
  • GPS Global Positioning System
  • the stream processor 110 is designed to identify an error in the metadata of the video feed.
  • the stream processor 110 is configured to receive the video from the aerial platform 130 and GIS data from the GIS 150 and forward the video and the GIS data to the user equipment 140 to initiate further processing and generation of composite video and displaying of the composite video on the user equipment 140.
  • the video processor 206 corrects the metadata of the video to counter distortion generated due to obstructions within an optical path of the camera and due to the transformation.
  • the area locator 222 determines an area within the image based on the coordinates of the geographical area.
  • the embedder 224 receives the information from the area locator 222 and geo function 214 to embed the video in the area by matching the area and the coordinates of the geographical area.
  • the video is embedded in the area by replacing pixels of the area with pixels of the video. The edges of the video correspond to the boundaries of the area.
  • the embedder 224 forwards the embedded video in the image to the feed generator 230 to generate the composite video feed with the video embedded along with the image.
  • the video processor 206 causes a video player to display the composite video feed. In some embodiments, the video player function in the user equipment 140.
  • FIG. 3 illustrates a top view of a landscape according to an embodiment of the present disclosure.
  • the view shows the aerial platform 130 capturing the video of the geographical area at an angled overhead view of the geographical area comprising a grid of city streets.
  • the aerial platform 130 is depicted along with scan lines (dashed lines) to mark the geographical area 300.
  • the aerial platform 130 includes a drone, airplane, etc.
  • Sensors on the aerial platform 130 are used to capture metadata about the geographical area.
  • the metadata is embedded in the video.
  • the metadata includes geographical coordinates corresponding to a center and/or the four corners of the video.
  • the GIS 150 is configured to generate map information that is overlay ed as a graphical representation of the video.
  • the landscape comprises a graphical representation corresponding to physical features in the geographical area, such as roads and intersections.
  • graphical representations of other types of physical structures in the geographical area may be depicted additionally or alternatively.
  • the aerial platform 130 and a camera associated with the aerial platform 130 are calibrated such that the aerial platform 130 flies over a geographical area and the camera is able to capture the video of the geographical area.
  • a video of the geographical area is captured.
  • the captured video includes multiple frames and associated metadata.
  • the metadata may include location information of the geographical area and location information of the physical objects present in the geographical area.
  • the metadata may also include information related to azimuth, elevation, and information related to other angles of the camera.
  • the video undergoes two sub-processes (610a and 610b).
  • One of the subprocesses is named distortion correction stage 610a and the other subprocess is named camera monitoring stage 610b.
  • the received video is checked for distortion due to improper focus of the camera on the geographical region or distortion due to obstruction in an optical path of the camera towards the geographical area.
  • the obstruction may occur due to climatic conditions, such as rain or fog.
  • the camera is calibrated, and the video is captured again after calibration.
  • the user may provide instruction to the aerial platform 130 for displacement to a nearby flight space that may also be capable to provide video of the geographical area but maybe with a different camera angle. In this manner, the video is dynamically updated for correcting any distortions that may have occurred at the time of capturing the video.
  • a change in the camera perspective is identified based on a change in qualifying parameters as compared to the calibration parameters associated with the aerial platform 130 and the camera is stored in the database.
  • a video player is caused to display the composite video feed.
  • the video player functions in the user equipment 140.
  • the video player shows the composite video feed with a graphical representation of physical features, such as roads, intersections, rivers, and residential societies overlaid on the video.
  • FIG. 7 illustrates a flowchart of method 700 for generating a composite video feed according to another embodiment of the present disclosure. Some blocks of method 700 may be performed by the stream processor 110 and by utilizing processing resources through any suitable hardware, non-transitory machine-readable medium, or a combination thereof.
  • the geo player is connected to the GIS 150 to communicate with the integrated database of the GIS 150. Simultaneous with connecting to the GIS 150, a stream processing stage 708 is enabled.
  • the geo-player is connected to the stream processor 110.
  • the received video is checked for distortion due to improper focus of the camera on the geographical region or distortion due to obstruction in an optical path of the camera towards the geographical area.
  • the obstruction may occur due to climatic conditions, such as rain or fog.
  • a change in the camera perspective is identified based on a change in qualifying parameters as compared to the calibration parameters associated with the aerial platform 130 and the camera is stored in the database.
  • a notification is generated for the stream processor 110 to receive a new video and upon identifying no change in the camera perspective, the metadata is retrieved by the user equipment 140.
  • geo-tag data is retrieved from the GIS 150 based on the retrieved metadata.
  • an image from the GIS 150 is received. The image is received based on the location coordinates of the geographical area present in the metadata, where the image is of the surrounding area of the geographical area.
  • pan and zoom operation is performed on the image to identify the geographical area according to the metadata.
  • the image is transformed according to the metadata.
  • the image is transformed by rotating the image, skewing or de-skewing the image, changing orientation, and scaling the image to match a particular size of the video.
  • the image is transformed from 2-D to 3-D by applying a homographic matrix for viewing the image according to an optical perspective of the camera. Further, an area corresponding to the geographical area is located in the image.
  • the video is embedded in the area by matching the area and the coordinates of the geographical area.
  • the video is embedded in the area by replacing pixels of the area with pixels of the video. The edges of the video correspond to the boundaries of the area.
  • a graphical representation of physical features such as roads, intersections, rivers, and residential societies overlaid on the video.
  • a composite video feed is generated with the video embedded along with the image and the graphical representation.
  • the geo player is caused to display the composite video feed.
  • the present embodiments described herein related to the system and the method for generating the composite video feed for the geographical area enables a user to watch the video feed on any video player, where the video feed is provided along with geo-tag information in a same frame rate as it was recorded. This enables the user to easily track asynchronous events happening in the geographical area in real-time. The user or a trainee can therefore have great situational awareness, especially when functioning in the military environment.
  • Implementation of the techniques, blocks, steps and means described above may be done in various ways. For example, these techniques, blocks, steps and means may be implemented in hardware, software, or a combination thereof.
  • the processing units may be implemented within one or more application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field programmable gate arrays (FPGAs), processors, controllers, micro-controllers, microprocessors, other electronic units designed to perform the functions described above, and/or a combination thereof.
  • ASICs application specific integrated circuits
  • DSPs digital signal processors
  • DSPDs digital signal processing devices
  • PLDs programmable logic devices
  • FPGAs field programmable gate arrays
  • processors controllers, micro-controllers, microprocessors, other electronic units designed to perform the functions described above, and/or a combination thereof.
  • the embodiments may be described as a process which is depicted as a flowchart, a flow diagram, a swim diagram, a data flow diagram, a structure diagram, or a block diagram. Although a depiction may describe the operations as a sequential process, many of the operations can be performed in parallel or concurrently. In addition, the order of the operations may be re-arranged.
  • a process is terminated when its operations are completed, but could have additional steps not included in the figure.
  • a process may correspond to a method, a function, a procedure, a subroutine, a subprogram, etc. When a process corresponds to a function, its termination corresponds to a return of the function to the calling function or the main function.
  • embodiments may be implemented by hardware, software, scripting languages, firmware, middleware, microcode, hardware description languages, and/or any combination thereof.
  • the program code or code segments to perform the necessary tasks may be stored in a machine readable medium such as a storage medium.
  • a code segment or machineexecutable instruction may represent a procedure, a function, a subprogram, a program, a routine, a subroutine, a module, a software package, a script, a class, or any combination of instructions, data structures, and/or program statements.
  • a code segment may be coupled to another code segment or a hardware circuit by passing and/or receiving information, data, arguments, parameters, and/or memory contents.
  • Information, arguments, parameters, data, etc. may be passed, forwarded, or transmitted via any suitable means including memory sharing, message passing, token passing, network transmission, etc.
  • the methodologies may be implemented with modules (e.g., procedures, functions, and so on) that perform the functions described herein. Any machine-readable medium tangibly embodying instructions may be used in implementing the methodologies described herein.
  • software codes may be stored in a memory.
  • Memory may be implemented within the processor or external to the processor.
  • the term “memory” refers to any type of long term, short term, volatile, nonvolatile, or other storage medium and is not to be limited to any particular type of memory or number of memories, or type of media upon which memory is stored.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Signal Processing (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Library & Information Science (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Processing Or Creating Images (AREA)

Abstract

A method and a system for generating a composite video feed for a geographical area are disclosed. A video of the geographical area, captured by a camera, of an aerial platform is received. The video includes metadata indicative of location information, which is used to identify the coordinates of the geographical area. An image that is adjacent to the geographical area is received from the geographical information system and is transformed according to the metadata. The coordinates of the geographical area are used to determine an area with the image. The video is embedded in the area by matching the area with the coordinates of the geographical area, where the edges of the video correspond to the boundaries of the area. A composite video feed, including the video embedded along with the image, is generated and a video player displays the composite video feed.

Description

GEO-SPATIAL CONTEXT FOR FULL-MOTION VIDEO
[0001] This application claims the benefit of U.S. Patent Application No. 18/158,886, filed January 24, 2023, and to U.S. Provisional Application No. 63/302,972 filed on January 25, 2022, which is hereby expressly incorporated by reference in its entirety for all purposes.
BACKGROUND
[0002] This disclosure generally relates to a system and a method and, not by way of limitation, to generate a composite video feed for a geographical area.
[0003] For providing a web mapping service, various platforms are designed to serve georeferenced images over the Internet. The geo-referenced images are then displayed on a display device connected to the platforms. The geo-referenced images are typically retrieved from a data center associated with a Geographical Information System (GIS). Further, the platforms are designed to provide a 3D representation of imagery stored in the data center. To achieve this, satellite images, aerial photography, and GIS data are superimposed onto a 3D geographical region. Furthermore, the platforms provide options to modify views of the 3D geographical region to allow users to see landscapes on the display device from various angles or positions.
SUMMARY
[0004] In one embodiment, a method and a system for generating a composite video feed for a geographical area are disclosed. A video of the geographical area, captured by a camera, of an aerial platform is received. The video includes metadata indicative of location information, which is used to identify the coordinates of the geographical area. An image that is adjacent to the geographical area is received from the geographical information system and is transformed according to the metadata. The coordinates of the geographical area are used to determine an area with the image. The video is embedded in the area by matching the area with the coordinates of the geographical area, where the edges of the video correspond to the boundaries of the area. A composite video feed, including the video embedded along with the image, is generated and a video player displays the composite video feed.
[0005] In another embodiment, a method for generating a composite video feed for a geographical area is disclosed. A video of the geographical area captured by an aerial platform is received. Metadata, indicative of location information associated with the video, is processed. Coordinates of the geographical area are identified based on the location information. An image is received from a geographical information system (GIS), where the image is adjacent to the geographical area. The image is transformed according to the metadata. An area within the image is determined based on the coordinates of the geographical area. The video is embedded in the area by matching the area and the coordinates of the geographical area, where edges of the video correspond to boundaries of the area. The composite video feed with the video embedded along with the image is generated. A video player is caused to display the composite video feed.
[0006] In still embodiment, a system for generating a composite video feed for a geographical area is disclosed. The system comprises a video processor configured to:
• receive a video of the geographical area captured by an aerial platform;
• process metadata indicative of location information associated with the video; and
• identify coordinates of the geographical area based on the location information;
The system further comprises an integrator that is configured to receive an image from a geographical information system (GIS), wherein the image is adjacent to the geographical area. The video processor is further configured to:
• transform the image according to the metadata;
• determine an area within the image based on the coordinates of the geographical area; and
• embed the video in the area by matching the area and the coordinates of the geographical area, wherein edges of the video correspond to boundaries of the area; and
The system further comprises a feed generator that is configured to:
• generate the composite video feed with the video embedded along with the image; and
• cause a video player to display the composite video feed.
[0007] Further areas of applicability of the present disclosure will become apparent from the detailed description provided hereinafter. It should be understood that the detailed description and specific examples, while indicating various embodiments, are intended for purposes of illustration only and are not intended to necessarily limit the scope of the disclosure.
BRIEF DESCRIPTION OF THE DRAWINGS
[0008] The present disclosure is described in conjunction with the appended figures: FIG. 1 illustrates a schematic representation of a military environment according to an embodiment of the present disclosure; FIG. 2 illustrates a block diagram showing a stream processor according to an embodiment of the present disclosure;
FIG. 3 illustrates a top view of a landscape according to an embodiment of the present disclosure;
FIG. 4 illustrates a 3D video feed of a geographical area according to an embodiment of the present disclosure;
FIG. 5 illustrates a composite video feed of the geographical area according to an embodiment of the present disclosure;
FIG. 6 illustrates a flowchart of a method for generating a composite video feed according to an embodiment of the present disclosure; and
FIG. 7 illustrates a flowchart of a method for generating a composite video feed according to another embodiment of the present disclosure.
[0009] In the appended figures, similar components and/or features may have the same reference label. Further, various components of the same type may be distinguished by following the reference label by a second alphabetical label that distinguishes among the similar components. If only the first reference label is used in the specification, the description is applicable to any one of the similar components having the same first reference label irrespective of the second reference label.
DETAILED DESCRIPTION
[0010] The ensuing description provides preferred exemplary embodiment(s) only, and is not intended to limit the scope, applicability or configuration of the disclosure. Rather, the ensuing description of the preferred exemplary embodiment(s) will provide those skilled in the art with an enabling description for implementing a preferred exemplary embodiment. It is understood that various changes may be made in the function and arrangement of elements without departing from the spirit and scope as set forth in the appended claims.
[0011] Embodiments described herein are generally related to a method and a system for generating a composite video feed for a geographical area. In particular, some embodiments of the disclosure describe processes for capturing a video via an aerial platform in a military environment. The disclosure specifically indicates receiving the video with metadata and an image from geographical information system (GIS) corresponding to the metadata. The metadata includes location information, which is either or together the location information of the geographical area and the location information of the aerial platform. The disclosure further describes about embedding of the captured video in the image along with overlaying geo-tag data on the video. For performing the embedding process a specific area on the image is determined based on the location information of the geographical area in the video. Further, the pixels of the area are replaced with the pixels of the video in the image. Upon embedding a video player plays a composite video feed at a full frame rate.
[0012] The presentation of the video helps a user, in particular a trainee soldier, to train in a simulating environment designed according to a battlefield to accurately identify physical objects in the geographical area. As the video is played at the full frame rate, a greater situational awareness is provided to the user to enable swift responses related to any asynchronous events in the geographical area.
[0013] Referring to FIG. 1, illustrates a military environment 100 that includes elements for generating a composite video feed for a geographical area. The military environment 100 includes a stream processor 110, three aerial platforms 130-1, 130-2, 130-3, a communication network 120, four user equipment 140-1, 140-2, 140-3, 140-4, and a Geographical Information System (GIS) 150.
[0014] In the military environment 100 various programs may be dedicated for training soldiers or other civic personnel. One of the programs may relate to identification of a geographical area with specific details of physical objects present in the geographical area. For this, a video is captured via the three aerial platforms 130-1, 130-2, and 130-3. The aerial platforms 130-1, 130-2, and 130-3 are moving vehicles, such as airplanes, drones, or the like. The aerial platforms 130-1, 130-2, and 130-3 use a camera and other onboard electrical components to capture video of the geographical area and transmit video feed via the communication network 120 to various user equipment 140-1, 140-2, 140-3 for display. The video feed is transmitted to a stream processor 110 for analysis. The stream processor 110 is a command center, operations room, control outpost, or the like. The stream processor 110 works in combination with the GIS 150 for performing video processing and analysis in real-time The GIS 150 utilizes various geospatial tools for collaborating with the stream processor 110 in order to provide GIS image and Geo-tag data. The GIS 150 further includes an integrated database to provide the GIS image and the Geo-tag data in accordance with the geographical area of the video feed. The communication network 120 can be a wide area network (WAN). The WAN may comprise one or more public and/or private data communication networks, including the Internet, one or more cellular, radio, or other wireless WANs (WWANs), and/or the like.
[0015] The aerial platforms 130 includes multiple sensors to create metadata indicative of the location of the geographical area captured in the video. The metadata provides geographical coordinates of the physical objects within the frame of the video. Further, the sensors are designed to generate raw sensor data. The sensors are selected from a Global Navigation Satellite System GNSS) receiver (e.g., a Global Positioning System (GPS) receiver), magnetometer, altimeter, gyroscope, accelerometer, and/or the like, and may be generally indicative of a location of the geographical area in the video as well as elevation, azimuth, and orientation of the camera capturing video.
[0016] In some embodiments, the sensor data is embedded into the video feed in accordance with governing standards and protocols. The standards are protocols are dependent on various jurisdictions. For example, a Motion Imagery Standards Board (MISB) standard 0601 is one such standard indicating how sensor data is embedded into the video (e.g., as a real-time, synchronous MISB key-length-value (KLV) stream). In some other embodiments, alternative standards may be used depending on functionality.
[0017] In some embodiments, the stream processor 110 is designed to identify an error in the metadata of the video feed. In some embodiments, the stream processor 110 is configured to receive the video from the aerial platform 130 and GIS data from the GIS 150 and forward the video and the GIS data to the user equipment 140 to initiate further processing and generation of composite video and displaying of the composite video on the user equipment 140.
[0018] Referring to FIG. 2, illustrates a block diagram of the stream processor 110 as described in FIG. 1. The stream processor 110 includes a network interface 200, a source controller 202, an application manager 204, a video processor 206, a video receiver 208, a metadata processor 210, a decoder 212, a geo function 214, an integrator 216, an image receiver 218, an image transformer 220, an embedder 224, a data receiver 226, an overlay engine 228, a feed generator 230, and a recorder 240. [0019] The aerial platform 130 communicates with the stream processor 110 via a network interface, where the source controller 202 is configured to select the aerial platform 130 from a plurality of aerial platforms 130-1, 130-2, and 130-2. The source controller 202 is further configured to calibrate the aerial platform with respect to altitude, speed, and location of the aerial platform, and the field of view (FOV) of the camera. The calibration is performed either on the ground or in flight prior to capturing the video. The source controller 202 used herein a source controller is a built-in computer or chipset or processor, which functions in combination with the application manager 204. The application manager 204 is a user interface that executes processes of both stream processor 110 and the aerial platform 130. In some embodiments, the application manager 204 works external to the stream processor 110, for example, at the user equipment 140. The application manager 204 provides communication connectivity between various applications at the stream processor 110 and the aerial platform 130. The network interface 200 also communicates with the video processor 206. The video processor 206 comprises the video receiver 208 to receive the video of the geographical area captured by the aerial platform 130. The video processor 206 further comprises the metadata processor 210 to process metadata indicative of location information associated with the video. In some embodiments, the metadata includes location information of the aerial platform 130 of the geographical area captured in the video, and elevation and azimuth with respect to the FOV of the camera. The aerial platform 130 captures the video and embeds the metadata and decodes the video in a particular video format, for which the decoder 212 is utilized in the stream processor 110 to decode the video. The geo function 214 receives location information from the metadata processor 210 and the decoder to identify coordinates of the geographical area based on the location information.
[0020] The integrator 216 is configured to receive GIS image and GIS data from the GIS 150. The GIS image is adjacent to the geographical area. In some embodiments, the image is a 2- D image of area surrounding the geographical area captured in the video. The video processor 206 further comprises the image transformer 220 to transform the image according to the metadata. In some embodiments, the image is transformed by rotating the image, skewing or deskewing the image, changing orientation, and scaling the image to match a particular size of the video. In some other embodiments, the image is transformed from 2-D to 3-D by applying a homographic matrix for viewing the image according to an optical perspective of the camera. In some other embodiments, the video processor 206 corrects the metadata of the video to counter distortion generated due to obstructions within an optical path of the camera and due to the transformation. The area locator 222 determines an area within the image based on the coordinates of the geographical area. The embedder 224 receives the information from the area locator 222 and geo function 214 to embed the video in the area by matching the area and the coordinates of the geographical area. In some embodiments, the video is embedded in the area by replacing pixels of the area with pixels of the video. The edges of the video correspond to the boundaries of the area. The embedder 224 forwards the embedded video in the image to the feed generator 230 to generate the composite video feed with the video embedded along with the image. The video processor 206 causes a video player to display the composite video feed. In some embodiments, the video player function in the user equipment 140.
[0021] The data receiver 226 is communicatively coupled to the integrator 216 to receive the GIS data. The GIS data includes the geo-tagging information that is forwarded to the overlay engine 228. The geo function 214 uses the geo-tagging information from the overlay engine 228 to overlay geo-tagging information on the composite video feed, where the feed generator 230 is further configured to generate a composite video feed with the geo-tagging information. The video processor is further configured to cause the video player to display the composite video feed at a full-frame rate. The full-frame rate is the upper limit of the frame rate supported by the user equipment 140. In some embodiments, the recorder 240 is configured to record the composite video feed for a pre-defined duration.
[0022] Referring to FIG. 3, illustrates a top view of a landscape according to an embodiment of the present disclosure. The view shows the aerial platform 130 capturing the video of the geographical area at an angled overhead view of the geographical area comprising a grid of city streets. As illustrated, the aerial platform 130 is depicted along with scan lines (dashed lines) to mark the geographical area 300. The aerial platform 130 includes a drone, airplane, etc. Sensors on the aerial platform 130 are used to capture metadata about the geographical area. The metadata is embedded in the video. In some embodiments, the metadata includes geographical coordinates corresponding to a center and/or the four corners of the video. [0023] The GIS 150 is configured to generate map information that is overlay ed as a graphical representation of the video. As illustrated, the landscape comprises a graphical representation corresponding to physical features in the geographical area, such as roads and intersections. Depending on functionality, available map information, and/or other factors, graphical representations of other types of physical structures in the geographical area may be depicted additionally or alternatively.
[0024] Referring to FIG. 4, illustrates a 3D video feed 400 of a geographical area according to an embodiment of the present disclosure. According to some embodiments, the 3D video feed 400 comprises a live video at a full frame rate, which is dynamically updated based on new camera angles or perspectives. In other words, stream processor 110 obtains location information from the metadata for an image within the video feed to determine the geographical area falling within the frame of that image. The stream processor 110 can then retrieve the map data relevant to the geographical area and create overlay geo-tag information on the video.
[0025] Referring to FIG. 5, illustrates a composite video feed 500 of the geographical area according to an embodiment of the present disclosure. According to embodiments, the user is able to view the combined visualization of the video 400 and a GIS image 520. The GIS image 520 is transformed according to the metadata. In some embodiments, the GIS image 520 is transformed by rotating the image, skewing or de- skewing the image, changing orientation and scaling of the image to match a particular size of the video. In some other embodiments, GIS image 520 is transformed from 2-D to 3-D by applying a homographic matrix for viewing the GIS image 520 according to an optical perspective of the camera. Further, an area corresponding to the geographical area is located in the GIS image 520. The video 400 is embedded in the area by matching the area and the coordinates of the geographical area. In some embodiments, the video is embedded in the area by replacing pixels of the area with pixels of video 510. The edges of video 400 correspond to the boundaries of the area. The composite video feed 500 is generated with the video 400 embedded along with the GIS image 520 and the video player is caused to display the composite video feed. In some embodiments, the video player functions in the user equipment 140. In some embodiments, the video player shows the composite video feed with a graphical representation of physical features, such as roads, intersections, rivers, and residential societies overlaid on video 400. [0026] Referring to FIG. 6, illustrates a flowchart of method 600 for generating a composite video feed according to an embodiment of the present disclosure. Some blocks of method 600 may be performed by the stream processor 110 and by utilizing processing resources through any suitable hardware, non-transitory machine-readable medium, or a combination thereof.
[0027] At block 602, an aerial platform 130 is selected out of a plurality of aerial platforms as shown in FIG. 1. The aerial platform 130 is selected based on the current location and flight path of the aerial platform 130. In some embodiments, a user may provide instructions to the aerial platform for moving to different paths. In some other embodiments, the user may operate the aerial platform via the user equipment 160. In other words, a user may provide instructions to the stream processor 110 to enable the source controller 202 to operate and control the aerial platform 130.
[0028] At block 604, a control program, designed to operate the selected aerial platform 130, is executed. The control program has a series of instructions that may be selected by the user in order to control and operate the aerial platform 130.
[0029] At block 606, the aerial platform 130 and a camera associated with the aerial platform 130 are calibrated such that the aerial platform 130 flies over a geographical area and the camera is able to capture the video of the geographical area.
[0030] At block 608, calibration parameters associated with the aerial platform 130 and the camera are stored in a database, so that the parameters can be checked and can be referred to at a later stage of the video-capturing process. In some embodiments, the user may provide instructions to the aerial platform 130 to change the flight path and move to a different location. In some other embodiments, the user may provide instructions to change the camera focus or camera angle in order to counter distortion generated in an optical path of the camera.
[0031] At block 610, a video of the geographical area is captured. The captured video includes multiple frames and associated metadata. The metadata may include location information of the geographical area and location information of the physical objects present in the geographical area. In some embodiments, the metadata may also include information related to azimuth, elevation, and information related to other angles of the camera. Upon video capture, the video undergoes two sub-processes (610a and 610b). One of the subprocesses is named distortion correction stage 610a and the other subprocess is named camera monitoring stage 610b.
[0032] At block 612, the captured video is received by the video receiver 208 of the stream processor 110.
[0033] At block 614, the received video is checked for distortion due to improper focus of the camera on the geographical region or distortion due to obstruction in an optical path of the camera towards the geographical area. The obstruction may occur due to climatic conditions, such as rain or fog.
[0034] At block 616, upon identifying distortion in the video, the camera is calibrated, and the video is captured again after calibration. In some embodiments, the user may provide instruction to the aerial platform 130 for displacement to a nearby flight space that may also be capable to provide video of the geographical area but maybe with a different camera angle. In this manner, the video is dynamically updated for correcting any distortions that may have occurred at the time of capturing the video.
[0035] At block 618, upon identifying no distortion in the video, the metadata is retrieved by the stream processor 110.
[0036] At block 620, and referring to the camera monitoring stage 610b, the perspective of the camera is continually monitored while the video is received by the video receiver 208. For example, the aerial platform while flying may provide a live video feed from multiple camera perspectives In this case, individual camera angle may provide a different view of the geographical area captured in the video.
[0037] At block 622, a change in the camera perspective is identified based on a change in qualifying parameters as compared to the calibration parameters associated with the aerial platform 130 and the camera is stored in the database.
[0038] At block 624, upon identifying the change in the perspective of the camera, the video is re-captured. In this manner, the video is dynamically updated for correcting any distortions that may have occurred at the time of capturing the video due to changes in the camera perspective. [0039] At block 626, upon identifying no change in the camera perspective, the retrieved metadata is processed. The metadata is inclusive of the location coordinates of the geographical area.
[0040] At block 628, an image from a geographical information system (GIS) is received. The image is received based on the location coordinates of the geographical area, where the image is of the surrounding area of the geographical area.
[0041] At block 630, the image is transformed according to the metadata. In some embodiments, the image is transformed by rotating the image, skewing or de-skewing the image, changing orientation, and scaling the image to match a particular size of the video. In some other embodiments, the image is transformed from 2-D to 3-D by applying a homographic matrix for viewing the image according to an optical perspective of the camera. Further, an area corresponding to the geographical area is located in the image.
[0042] At block 632, the video is embedded in the area by matching the area and the coordinates of the geographical area. In some embodiments, the video is embedded in the area by replacing pixels of the area with pixels of the video. The edges of the video correspond to the boundaries of the area.
[0043] At block 634, a composite video feed is generated with the video embedded along with the image.
[0044] At block 636, a video player is caused to display the composite video feed. In some embodiments, the video player functions in the user equipment 140. In some embodiments, the video player shows the composite video feed with a graphical representation of physical features, such as roads, intersections, rivers, and residential societies overlaid on the video.
[0045] Referring to FIG. 7, illustrates a flowchart of method 700 for generating a composite video feed according to another embodiment of the present disclosure. Some blocks of method 700 may be performed by the stream processor 110 and by utilizing processing resources through any suitable hardware, non-transitory machine-readable medium, or a combination thereof.
[0046] At block 702, a geo player is installed on the user equipment 140. In some embodiments, the geo-player is associated with the GIS 150, where the geo-player is configured to receive the composite video feed along with geo-tag information associated with the geographical area. In some other embodiments, an integrated database of the GIS 150 provides processed geo-tag information corresponding to the metadata of the video received by the stream processor 110. In the embodiments of FIG. 7, the stream processor 110 functions externally to the user equipment 140. However, in some other embodiments, the stream processor 110 is integrated into the user equipment 140. In some embodiments, the geo-player is similar to the video player of FIG. 2. In an example, the geo-player may be loaded directly on a web browser. [0047] At block 704, the geo-player is executed on the user equipment to enable a user interface of the geo-player.
[0048] At block 706, the geo player is connected to the GIS 150 to communicate with the integrated database of the GIS 150. Simultaneous with connecting to the GIS 150, a stream processing stage 708 is enabled.
[0049] At block 710, the geo-player is connected to the stream processor 110.
[0050] At block 712, the captured video is received by the geo-player.
[0051] At block 714, the received video is checked for distortion due to improper focus of the camera on the geographical region or distortion due to obstruction in an optical path of the camera towards the geographical area. The obstruction may occur due to climatic conditions, such as rain or fog.
[0052] At block 716, upon identifying distortion in the video, a report is generated for the stream processor 110 to correct parameters for countering distortions and receive a new video. [0053] At block 718, upon identifying no distortion in the video, the metadata is retrieved by the user equipment 140.
[0054] At block 720, a perspective of the camera is continually monitored while the video is received by the geo player.
[0055] At block 722, a change in the camera perspective is identified based on a change in qualifying parameters as compared to the calibration parameters associated with the aerial platform 130 and the camera is stored in the database. Upon identifying the change in the perspective of the camera, a notification is generated for the stream processor 110 to receive a new video and upon identifying no change in the camera perspective, the metadata is retrieved by the user equipment 140.
[0056] At block 723, geo-tag data is retrieved from the GIS 150 based on the retrieved metadata. [0057] At block 724, an image from the GIS 150 is received. The image is received based on the location coordinates of the geographical area present in the metadata, where the image is of the surrounding area of the geographical area.
[0058] At block 726, pan and zoom operation is performed on the image to identify the geographical area according to the metadata.
[0059] At block 728, the image is transformed according to the metadata. In some embodiments, the image is transformed by rotating the image, skewing or de-skewing the image, changing orientation, and scaling the image to match a particular size of the video. In some other embodiments, the image is transformed from 2-D to 3-D by applying a homographic matrix for viewing the image according to an optical perspective of the camera. Further, an area corresponding to the geographical area is located in the image.
[0060] At block 730, the video is embedded in the area by matching the area and the coordinates of the geographical area. In some embodiments, the video is embedded in the area by replacing pixels of the area with pixels of the video. The edges of the video correspond to the boundaries of the area.
[0061] At block 732, In some embodiments, a graphical representation of physical features, such as roads, intersections, rivers, and residential societies overlaid on the video.
[0062] At block 734, a composite video feed is generated with the video embedded along with the image and the graphical representation.
[0063] At block 736, the geo player is caused to display the composite video feed.
[0064] The present embodiments described herein related to the system and the method for generating the composite video feed for the geographical area enables a user to watch the video feed on any video player, where the video feed is provided along with geo-tag information in a same frame rate as it was recorded. This enables the user to easily track asynchronous events happening in the geographical area in real-time. The user or a trainee can therefore have great situational awareness, especially when functioning in the military environment.
[0065] Specific details are given in the above description to provide a thorough understanding of the embodiments. However, it is understood that the embodiments may be practiced without these specific details. For example, circuits may be shown in block diagrams in order not to obscure the embodiments in unnecessary detail. In other instances, well-known circuits, processes, algorithms, structures, and techniques may be shown without unnecessary detail in order to avoid obscuring the embodiments.
[0066] Implementation of the techniques, blocks, steps and means described above may be done in various ways. For example, these techniques, blocks, steps and means may be implemented in hardware, software, or a combination thereof. For a hardware implementation, the processing units may be implemented within one or more application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field programmable gate arrays (FPGAs), processors, controllers, micro-controllers, microprocessors, other electronic units designed to perform the functions described above, and/or a combination thereof.
[0067] Also, it is noted that the embodiments may be described as a process which is depicted as a flowchart, a flow diagram, a swim diagram, a data flow diagram, a structure diagram, or a block diagram. Although a depiction may describe the operations as a sequential process, many of the operations can be performed in parallel or concurrently. In addition, the order of the operations may be re-arranged. A process is terminated when its operations are completed, but could have additional steps not included in the figure. A process may correspond to a method, a function, a procedure, a subroutine, a subprogram, etc. When a process corresponds to a function, its termination corresponds to a return of the function to the calling function or the main function.
[0068] Furthermore, embodiments may be implemented by hardware, software, scripting languages, firmware, middleware, microcode, hardware description languages, and/or any combination thereof. When implemented in software, firmware, middleware, scripting language, and/or microcode, the program code or code segments to perform the necessary tasks may be stored in a machine readable medium such as a storage medium. A code segment or machineexecutable instruction may represent a procedure, a function, a subprogram, a program, a routine, a subroutine, a module, a software package, a script, a class, or any combination of instructions, data structures, and/or program statements. A code segment may be coupled to another code segment or a hardware circuit by passing and/or receiving information, data, arguments, parameters, and/or memory contents. Information, arguments, parameters, data, etc. may be passed, forwarded, or transmitted via any suitable means including memory sharing, message passing, token passing, network transmission, etc.
[0069] For a firmware and/or software implementation, the methodologies may be implemented with modules (e.g., procedures, functions, and so on) that perform the functions described herein. Any machine-readable medium tangibly embodying instructions may be used in implementing the methodologies described herein. For example, software codes may be stored in a memory. Memory may be implemented within the processor or external to the processor. As used herein the term “memory” refers to any type of long term, short term, volatile, nonvolatile, or other storage medium and is not to be limited to any particular type of memory or number of memories, or type of media upon which memory is stored.
[0070] Moreover, as disclosed herein, the term "storage medium" may represent one or more memories for storing data, including read only memory (ROM), random access memory (RAM), magnetic RAM, core memory, magnetic disk storage mediums, optical storage mediums, flash memory devices and/or other machine readable mediums for storing information. The term "machine-readable medium" includes, but is not limited to portable or fixed storage devices, optical storage devices, and/or various other storage mediums capable of storing that contain or carry instruction(s) and/or data.
[0071] While the principles of the disclosure have been described above in connection with specific apparatuses and methods, it is to be clearly understood that this description is made only by way of example and not as limitation on the scope of the disclosure.

Claims

CLAIMS WHAT IS CLAIMED IS:
1. A method for generating a composite video feed for a geographical area, the method comprising: receiving a video of the geographical area captured by an aerial platform; processing metadata indicative of location information associated with the video; identifying coordinates of the geographical area based on the location information; receiving an image from a geographical information system (GIS), wherein the image is adjacent to the geographical area; transforming the image according to the metadata; determining an area within the image based on the coordinates of the geographical area; embedding the video in the area by matching the area and the coordinates of the geographical area, wherein edges of the video correspond to boundaries of the area; generating the composite video feed with the video embedded along with the image; and causing a video player to display the composite video feed.
2. The method for generating the composite video feed for the geographical area as claimed in claim 1, the method further comprises: selecting an aerial platform from a plurality of aerial platforms, wherein the aerial platform has a camera for capturing the video of the geographical area; and prior to capturing the video, calibrating the aerial platform with respect to: altitude, speed, and location of the aerial platform, and a field of view (FOV) of the camera.
3. The method for generating the composite video feed for the geographical area as claimed in claim 2, wherein the metadata includes: location information of the aerial platform of the geographical area captured in the video, and elevation and azimuth with respect to the FOV of the camera.
4. The method for generating the composite video feed for the geographical area as claimed in claim 1, wherein the image is a 2-D image of area surrounding the geographical area captured in the video.
5. The method for generating the composite video feed for the geographical area as claimed in claim 1, wherein transforming the image includes rotating the image, skewing or de-skewing the image, changing orientation and scaling of the image to match a particular size of the video.
6. The method for generating the composite video feed for the geographical area as claimed in claim 1, wherein the method further comprises: transforming the image from 2-D to 3-D by applying a homographic matrix for viewing the image according to an optical perspective of a camera; and correcting metadata of the video to counter distortion generated due to obstructions within an optical path of the camera and due to the transformation.
7. The method for generating the composite video feed for the geographical area as claimed in claim 1, wherein the method further comprises: generating geo-tagging information using the metadata; overlaying the video with the geo-tagging information; and generating a composite video feed with geo -tagging information.
8. The method for generating the composite video feed for the geographical area as claimed in claim 1, wherein embedding the video in the area includes replacing pixels of the area with pixels of the video.
9. The method for generating the composite video feed for the geographical area as claimed in claim 1, wherein the method further comprises causing the video player to display the composite video feed, at a full-frame rate.
10. The method for generating the composite video feed for the geographical area as claimed in claim 1, wherein the method further includes recording the composite video feed for a predefined duration.
11. A system for generating a composite video feed for a geographical area, the system comprising: a video processor configured to: receive a video of the geographical area captured by an aerial platform; process metadata indicative of location information associated with the video; and identify coordinates of the geographical area based on the location information; an integrator configured to receive an image from a geographical information system (GIS), wherein the image is adjacent to the geographical area; wherein the video processor is further configured to: transform the image according to the metadata; determine an area within the image based on the coordinates of the geographical area; and embed the video in the area by matching the area and the coordinates of the geographical area, wherein edges of the video correspond to boundaries of the area; and a feed generator configured to: generate the composite video feed with the video embedded along with the image; and cause a video player to display the composite video feed.
12. The system for generating the composite video feed for the geographical area of claim 11, wherein the video processor is further configured to: select an aerial platform from a plurality of aerial platforms, wherein the aerial platform has a camera for capturing the video of the geographical area; and prior to capturing the video, calibrate the aerial platform with respect to: altitude, speed, and location of the aerial platform, and; and a field of view (FOV) of the camera.
13. The system for generating the composite video feed for the geographical area of claim 12, wherein the metadata includes: location information of the aerial platform of the geographical area captured in the video, and; and elevation and azimuth with respect to the FOV of the camera.
14. The system for generating the composite video feed for the geographical area of claim 11, wherein the image is a 2-D image of area surrounding the geographical area captured in the video.
15. The system for generating the composite video feed for the geographical area of claim 11, wherein transforming the image includes rotating the image, skewing or de- skewing the image, changing orientation and scaling of the image to match a particular size of the video.
16. The system for generating the composite video feed for the geographical area of claim 11, wherein the video processor is further configured to: transform the image from 2-D to 3-D by applying a homographic matrix for viewing the image according to an optical perspective of a camera; and correct metadata of the image to counter distortion generated due to obstructions within an optical path of the camera and due to the transformation.
17. The system for generating the composite video feed for the geographical area of claim 11, wherein the video processor is further configured to: generate geo-tagging information using the metadata; and overlay the video with the geo-tagging information; wherein the feed generator is further configured to generate a composite video feed with the geotagging information.
18. The system for generating the composite video feed for the geographical area of claim 11, wherein embedding the video in the area includes replacing pixels of the area with pixels of the video.
19. The system for generating the composite video feed for the geographical area of claim 11, wherein the video processor is further configured to cause the video player to display the composite video feed, at a full-frame rate.
20. The system for generating the composite video feed for the geographical area of claim 11, wherein the system further comprises a recorder to record the composite video feed for a predefined duration.
PCT/US2023/061293 2022-01-25 2023-01-25 Geo-spatial context for full-motion video Ceased WO2023147376A1 (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US202263302972P 2022-01-25 2022-01-25
US63/302,972 2022-01-25
US18/158,886 2023-01-24
US18/158,886 US20230237796A1 (en) 2022-01-25 2023-01-24 Geo-spatial context for full-motion video

Publications (1)

Publication Number Publication Date
WO2023147376A1 true WO2023147376A1 (en) 2023-08-03

Family

ID=85283961

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2023/061293 Ceased WO2023147376A1 (en) 2022-01-25 2023-01-25 Geo-spatial context for full-motion video

Country Status (1)

Country Link
WO (1) WO2023147376A1 (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170052035A1 (en) * 2015-08-21 2017-02-23 Nokia Technologies Oy Location based service tools for video illustration, selection, and synchronization
CN109523579A (en) * 2018-11-12 2019-03-26 北京联海信息系统有限公司 A kind of matching process and device of UAV Video image and three-dimensional map
CN111385603A (en) * 2020-03-27 2020-07-07 宜宾中星技术智能系统有限公司 Method for embedding video into two-dimensional map
CN111444385A (en) * 2020-03-27 2020-07-24 西安应用光学研究所 Electronic map real-time video mosaic method based on image corner matching
CN112422927A (en) * 2020-12-01 2021-02-26 深圳高度创新技术有限公司 Real-time combination method and system for unmanned aerial vehicle shooting video and map

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170052035A1 (en) * 2015-08-21 2017-02-23 Nokia Technologies Oy Location based service tools for video illustration, selection, and synchronization
CN109523579A (en) * 2018-11-12 2019-03-26 北京联海信息系统有限公司 A kind of matching process and device of UAV Video image and three-dimensional map
CN111385603A (en) * 2020-03-27 2020-07-07 宜宾中星技术智能系统有限公司 Method for embedding video into two-dimensional map
CN111444385A (en) * 2020-03-27 2020-07-24 西安应用光学研究所 Electronic map real-time video mosaic method based on image corner matching
CN112422927A (en) * 2020-12-01 2021-02-26 深圳高度创新技术有限公司 Real-time combination method and system for unmanned aerial vehicle shooting video and map

Similar Documents

Publication Publication Date Title
US20200082563A1 (en) Digitally encoded marker-based augmented reality (ar)
US9743046B2 (en) Real-time moving platform management system
CN111540059B (en) Enhanced video systems providing enhanced environmental awareness
JP2011515760A (en) Visualizing camera feeds on a map
US10818055B2 (en) Computer readable media, information processing apparatus and information processing method
CN115439528B (en) Method and equipment for acquiring image position information of target object
US9467620B2 (en) Synthetic camera lenses
JPH1042282A (en) Video presentation system
WO2021035756A1 (en) Aircraft-based patrol inspection method and device, and storage medium
CN106095774A (en) A kind of unmanned plane image panorama methods of exhibiting
US20230237796A1 (en) Geo-spatial context for full-motion video
CN115760964A (en) Method and equipment for acquiring screen position information of target object
WO2023147376A1 (en) Geo-spatial context for full-motion video
CN111757185B (en) Video playing method and device
CA3063007A1 (en) Human-aided geo-rectification of geospatial metadata in video using a graphical interface
CN110602456A (en) Display method and system of aerial photography focus
CN115390725B (en) Video overlay method, apparatus, electronic device and readable storage medium
CN113066193B (en) Method for enhancing reality on live-action three-dimensional map
WO2025069841A1 (en) Image processing device, image processing method, and program
KR101643024B1 (en) Apparatus and method for providing augmented reality based on time
IL206371A (en) Systems and methods for distributing and archiving video with geographic information-based content

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23706271

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 23706271

Country of ref document: EP

Kind code of ref document: A1