US20170090481A1 - Enhanced distance detection system - Google Patents
Enhanced distance detection system Download PDFInfo
- Publication number
- US20170090481A1 US20170090481A1 US14/864,508 US201514864508A US2017090481A1 US 20170090481 A1 US20170090481 A1 US 20170090481A1 US 201514864508 A US201514864508 A US 201514864508A US 2017090481 A1 US2017090481 A1 US 2017090481A1
- Authority
- US
- United States
- Prior art keywords
- distance detector
- distance
- vehicle
- unmanned aerial
- aerial vehicle
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- E—FIXED CONSTRUCTIONS
- E04—BUILDING
- E04H—BUILDINGS OR LIKE STRUCTURES FOR PARTICULAR PURPOSES; SWIMMING OR SPLASH BATHS OR POOLS; MASTS; FENCING; TENTS OR CANOPIES, IN GENERAL
- E04H15/00—Tents or canopies, in general
- E04H15/32—Parts, components, construction details, accessories, interior equipment, specially adapted for tents, e.g. guy-line equipment, skirts, thresholds
- E04H15/34—Supporting means, e.g. frames
- E04H15/44—Supporting means, e.g. frames collapsible, e.g. breakdown type
- E04H15/48—Supporting means, e.g. frames collapsible, e.g. breakdown type foldable, i.e. having pivoted or hinged means
- E04H15/50—Supporting means, e.g. frames collapsible, e.g. breakdown type foldable, i.e. having pivoted or hinged means lazy-tongs type
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/10—Simultaneous control of position or course in three dimensions
- G05D1/101—Simultaneous control of position or course in three dimensions specially adapted for aircraft
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B64—AIRCRAFT; AVIATION; COSMONAUTICS
- B64C—AEROPLANES; HELICOPTERS
- B64C39/00—Aircraft not otherwise provided for
- B64C39/02—Aircraft not otherwise provided for characterised by special use
- B64C39/024—Aircraft not otherwise provided for characterised by special use of the remote controlled vehicle type, i.e. RPV
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B64—AIRCRAFT; AVIATION; COSMONAUTICS
- B64D—EQUIPMENT FOR FITTING IN OR TO AIRCRAFT; FLIGHT SUITS; PARACHUTES; ARRANGEMENT OR MOUNTING OF POWER PLANTS OR PROPULSION TRANSMISSIONS IN AIRCRAFT
- B64D47/00—Equipment not otherwise provided for
- B64D47/08—Arrangements of cameras
-
- E—FIXED CONSTRUCTIONS
- E04—BUILDING
- E04H—BUILDINGS OR LIKE STRUCTURES FOR PARTICULAR PURPOSES; SWIMMING OR SPLASH BATHS OR POOLS; MASTS; FENCING; TENTS OR CANOPIES, IN GENERAL
- E04H15/00—Tents or canopies, in general
- E04H15/32—Parts, components, construction details, accessories, interior equipment, specially adapted for tents, e.g. guy-line equipment, skirts, thresholds
- E04H15/34—Supporting means, e.g. frames
- E04H15/44—Supporting means, e.g. frames collapsible, e.g. breakdown type
- E04H15/46—Supporting means, e.g. frames collapsible, e.g. breakdown type telescoping and foldable
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/02—Systems using the reflection of electromagnetic waves other than radio waves
- G01S17/06—Systems determining position data of a target
- G01S17/08—Systems determining position data of a target for measuring distance only
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/86—Combinations of lidar systems with systems other than lidar, radar or sonar, e.g. with direction finders
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/88—Lidar systems specially adapted for specific applications
- G01S17/89—Lidar systems specially adapted for specific applications for mapping or imaging
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/88—Lidar systems specially adapted for specific applications
- G01S17/93—Lidar systems specially adapted for specific applications for anti-collision purposes
- G01S17/933—Lidar systems specially adapted for specific applications for anti-collision purposes of aircraft or spacecraft
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/48—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
- G01S7/481—Constructional features, e.g. arrangements of optical elements
- G01S7/4811—Constructional features, e.g. arrangements of optical elements common to transmitter and receiver
- G01S7/4813—Housing arrangements
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/48—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
- G01S7/481—Constructional features, e.g. arrangements of optical elements
- G01S7/4817—Constructional features, e.g. arrangements of optical elements relating to scanning
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/10—Simultaneous control of position or course in three dimensions
- G05D1/101—Simultaneous control of position or course in three dimensions specially adapted for aircraft
- G05D1/102—Simultaneous control of position or course in three dimensions specially adapted for aircraft specially adapted for vertical take-off of aircraft
-
- B64C2201/024—
-
- B64C2201/108—
-
- B64C2201/141—
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S13/00—Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
- G01S13/88—Radar or analogous systems specially adapted for specific applications
- G01S13/93—Radar or analogous systems specially adapted for specific applications for anti-collision purposes
- G01S13/933—Radar or analogous systems specially adapted for specific applications for anti-collision purposes of aircraft or spacecraft
- G01S13/935—Radar or analogous systems specially adapted for specific applications for anti-collision purposes of aircraft or spacecraft for terrain-avoidance
Definitions
- the described technology generally relates to unmanned aerial vehicles and, more specifically, to enhanced distance detection systems of unmanned aerial vehicles.
- An unmanned aerial vehicle also commonly referred to as drone, can travel through a variety of environments, such as indoor, outdoor, and/or mixed indoor and outdoor environments.
- an unmanned aerial vehicle can be configured to conduct surveillance, mapping, delivery, monitoring, or other tasks that can comprise combining movement and data collection.
- the unmanned aerial vehicle can travel over surfaces on which the unmanned aerial vehicle cannot safely land (e.g., water).
- an unmanned aerial vehicle includes a distance detector, a directional controller coupled to the distance detector, and one or more processors.
- the one or more processors are configured to execute a mission of the unmanned aerial vehicle, determine direction priorities based at least in part on current travel instructions of the mission, and acquire and/or process distance data generated by the distance detector based at least in part on the determined direction priorities.
- a method for targeted sensing for an unmanned aerial vehicle includes initiating a mission, determining direction priorities based at least in part on current travel instructions of the mission, controlling an orientation of a distance detector, and acquiring and/or processing distance data from the distance detector based at least in part on the determined direction priorities.
- FIG. 1 is a distance detector system according to one embodiment.
- FIG. 2 is an illustration of an unmanned aerial vehicle having a distance detector system according to one embodiment.
- FIG. 3 is an illustration of an example operation of an unmanned aerial vehicle having a distance detector system according to one embodiment.
- FIG. 4 is a block diagram of an unmanned aerial vehicle according to one embodiment.
- FIG. 5 is a flowchart for enhanced detection according to one embodiment.
- autonomous vehicle or “semi-autonomous vehicle,” as used herein, generally refers to a vehicle that is configured to operate without substantial or any involvement from an on-board operator (e.g., a driver or pilot).
- An “unmanned aerial vehicle,” or “UAV,” as used herein, can denote a type of autonomous or semi-autonomous vehicle whose physical operational capabilities include aerial travel or flight.
- An unmanned aerial vehicle can be an aircraft that is configured to take off and land on a surface. In some cases, an unmanned aerial vehicle can automatically travel from one location to another without any operator involvement. In some cases, an unmanned aerial vehicle can travel a far distance from a starting point. The distance can be far enough that the unmanned aerial vehicle cannot return to a starting point without refueling or recharging at an intermediate location.
- An unmanned aerial vehicle can be configured to land on a landing pad and/or charge at a charging station.
- the distance detection system includes a distance detector, which may have a limited scope of distance detection, and a directional controller, which allows extending the dimension or scope of the distance detector as the vehicle travels and performs missions.
- the directional controller can change the detection direction of the distance detector with a motorized gimbal or functionally similar system, and the change in the detection direction can be integrated with the status of and/or instructions executed or to be executed by the vehicle.
- FIG. 1 is a distance detector system according to one embodiment.
- the illustrated distance detector system 101 includes a distance detector 105 and a motorized directional controller 109 . Also shown in FIG. 1 are reference Cartesian and spherical coordinates, which are shown only for ease of explanation and are not considered limiting the implementation of the distance detector system 101 .
- the distance detector 105 can be implemented using at least one LIDAR, such as a one-, two-, or three-dimensional LIDAR, which has a field of detection 103 .
- the distance detector 105 can be implemented with technologies using other suitable form of signal or energy, such as radar, structured light, ultrasound, etc.
- the “field of detection” as used herein generally refers to one or more lines or paths of emission and detection of laser (or any other suitable form of signal or energy for distance detection) from and to the distance detector (e.g., 105 ) as defined by the orientation and internal configurations native or inherent in the design of the distance detector as one unit.
- a one-dimensional LIDAR may have one emitter-detector pair side by side pointing at the same direction in parallel, and the field of detection of the one-dimensional LIDAR can be a unit-resolution line defined by a unit-resolution point in the space at which the emitter-detector pair is pointing.
- a higher dimensional LIDAR may have more than one emitter-detector pair and/or have internal mechanical and/or optical mechanisms to emit and detect radiation in more than one direction.
- the field of detection can be an accumulation of the lines of emission and detection defined by multiple points in the space (e.g., a perimeter of a surface or a surface of a volume).
- the motorized directional controller 109 can be implemented with a generally L-shaped arm gimbal system providing additional degrees of freedom (e.g., in ⁇ and ⁇ in FIG. 1 ) about additional axes 102 , 104 to allow expanding upon or augmenting the field of detection 103 native to the distance detector 105 .
- the gimbal system illustrated in FIG. 1 can be implemented with a different mechanical structure, such as a U-shaped arm, that can similarly provide additional degrees of freedom.
- the illustrated directional controller 109 includes two hubs allowing the distance detector 105 to rotate about the axes 102 and 104 . Rotations about the axes 102 and 104 can be controlled by motors (not shown) and motor controllers (not shown).
- a brushless motor with a slip ring can be implemented to allow rotations about the axis 102
- another brushless motor can be implemented to allow rotations about the axis 104
- the directional controller 109 can provide continual rotations about one or more of the axes 102 , 104
- the directional controller 109 can provide limited or non-continual rotations about one or more of the axes 102 , 104 .
- the directional controller 109 can allow unlimited rotation about the axis 102 (i.e., unlimited in ⁇ in FIG.
- the range of rotations about the axes 102 , 104 can vary depending on, for example, the purpose, scope, and suitable level of redundancy in expanding the native field of detection 103 .
- the angle of orientation about the axes 102 and 104 at a given moment can each be measured at each hub by one or more sensors (e.g., optical encoder or angle sensors, such as Hall effect sensors or magneto-resistive sensors in combination with a dipole or ring magnet, etc.) and can be communicated to a processor 110 ( FIG. 4 ).
- the sensors, motors, and motor controllers of the distance detector system 101 can be dedicated sensors (not shown), motors (not shown), and motor controllers (not shown) specific to the distance detector system 101 .
- one or more of the sensors, motors, and motor controllers in the distance detector system 101 can be at least partially integrated with sensors 115 ( FIG. 4 ), motors 112 ( FIG. 4 ), and motor controllers 120 ( FIG. 4 ) configured to function with other parts of an autonomous vehicle 202 ( FIG. 4 ).
- the directional controller 109 can be motorless and implemented without dedicated motors controlling the added degrees of freedom discussed above.
- the directional controller 109 can be implemented with other mechanical systems, such as one or more ball and socket joints, hinges, gears, chain, belts, pulleys, or other mechanical couplings, to otherwise add additional directional dimensions to the distance detector 105 , or to expand the scope of coverage of the distance detector 105 .
- the added additional degrees of freedom can be more or less depending on the native scope of the distance detector 105 (e.g., the field of detection 103 ).
- the distance detector 105 can be a two-dimensional LIDAR, and the directional controller 109 may only provide one additional rotational degree of freedom.
- the distance detector 105 can be a two- or three-dimensional LIDAR having a limited scope or breath of distance detection (e.g., incomplete circumference or sphere), and one or more degrees of freedom about the additional axes (e.g., 102 , 104 ) can be added with the directional controller 109 to expand upon the scope of distance detection (e.g., the field of detection 103 ) native to the distance detector 105 .
- a limited scope or breath of distance detection e.g., incomplete circumference or sphere
- the additional axes e.g., 102 , 104
- the directional controller 109 can be implemented to provide additional one or more translational degrees of freedom (e.g., x, y, z, r, or any linear combination thereof) in addition to or instead of the rotational degrees of freedom, for example, to address mechanical blockage in the field of detection 103 , or for ease of stowage of all or part of the distance detector system 101 .
- additional one or more translational degrees of freedom e.g., x, y, z, r, or any linear combination thereof
- the added degrees of freedom can be orthogonal, or fully independent, to each other and to the native field of detection (e.g., 103 ) for efficiency, but in other embodiments, the added degrees of freedom can be at least partially overlapping or redundant or not fully independent or orthogonal, depending on, for example, the physical constraints of the autonomous vehicle 202 and the purpose of the added degrees of freedom.
- Example embodiments of mechanical integration of the distance detector system 101 and the autonomous vehicle 202 are further discussed in connection with FIG. 2 below.
- the directional controller 109 can be configured to rotate the distance detector 105 at a constant speed in at least one direction, and in other instances the directional controller 109 can be configured to rotate the distance detector 105 in a directed, targeted, and non-constant manner that is integrated with execution of a mission or the status of the autonomous vehicle 202 ( FIG. 2 ).
- the directional controller 109 can be configured to rotate the distance detector 105 about the first axis 102 constantly at about 3 Hz (i.e., three full-360° rotations in ⁇ per second) for an unmanned aerial vehicle moving at about 10 m/s or less.
- the directional controller 109 can be further configured to rotate the distance detector 105 about the second axis 104 back and forth depending on the area of interest as the unmanned aerial vehicle operates (e.g., 90° ⁇ with ⁇ being a small angle such as 5° if interested in detecting the surroundings at level with the vehicle altitude). Rotating about the axis 102 at a constant speed while slightly varying the angle (e.g., ⁇ ) about the axis 104 allows sensing and data collection of a three dimensional space of a partial sphere around the vehicle.
- the directional controller 109 rotates the distance detector 105 at 3 Hz and varies ⁇ to three values (e.g., 85°, 90°, and 95°), the data from the partial cone space defined by the location of the distance detector 105 and the range of ⁇ are gathered three times per second.
- the angle of interest in ⁇ can be different depending on, for example, relative altitudes of the vehicle 202 at its current location and the direction to which it is traveling (e.g., if traveling to a higher altitude, ⁇ can be less than 90°; if traveling to a lower altitude, ⁇ can be greater than) 90°.
- the variation of angle of rotation about the second axis 104 e.g., ⁇ or any other range of ⁇ which is not necessarily symmetric about the angle based on the relative altitudes
- the number of datapoints to gather at a given ⁇ and a range of ⁇ can be determined based on the speed and/or status of the vehicle 202 , the objective of the data collection, detection of obstacle(s), electromechanical constraints of the vehicle 202 , or any other suitable parameters or conditions.
- the unmanned aerial vehicle may travel at a faster or slower speed than the example discussed above, and the directional controller 109 can be configured to rotate or move the distance detector 105 at a higher or lower speed as appropriate.
- the appropriate speed (or frequency) of the rotation can also depend on, for example, mechanical constraints within the motorized directional controller 109 , such as speed tolerance of the slip ring of the motor.
- the directional controller 109 can be configured to enable and disable, start and stop, or accelerate and decelerate rotation or movements of the distance detector 105 depending on, for example, mission objectives, mission stage (e.g., takeoff, travel, landing), emergency situations, environmental conditions, power level, duration of the mission, etc.
- Example operations of the distance detector system 101 are further discussed in connection with FIG. 3 below.
- the distance detector system 101 can be advantageous to implement to augment or expand upon the native scope (e.g., the line of detection 103 ) of the distance detector (e.g., 105 ) since the added degrees of freedom by the directional controller 109 may allow effective and efficient implementation of the distance detector system 101 with reduced weight, cost, or complexity associated with a more comprehensive or exhaustive distance detector (e.g., three-dimensional LIDAR with full spherical scope of detection).
- a more comprehensive or exhaustive distance detector e.g., three-dimensional LIDAR with full spherical scope of detection.
- the internal mechanisms of a two-dimensional LIDAR may operate by spreading out transmission power over a larger area, which may render outdoor performance and range worse than a one-dimensional LIDAR at a given point in space, and a high-precision one-dimensional LIDAR rotated as needed can provide more advantages in performance over a standalone two-dimensional LIDAR.
- the distance detector system 101 as disclosed herein can achieve the resolution of, for example, 1 cm +/ ⁇ in detecting the location of one or more objects in the relevant partial space surrounding the vehicle 202 during takeoff, travel, and landing with a cost-effective, lightweight one-dimensional LIDAR. As further described below in connection with FIG.
- data collected from the distance detector 105 can be stored in a data structure involving variable resolution (e.g., resolution as low as 1 m) regardless of the full resolution or data collection capabilities of the distance detector 105 in order to improve efficiency in computation, memory, processing time, etc.
- variable resolution e.g., resolution as low as 1 m
- the distance detector system 101 allows detection, collection, and sensing operations of the distance detector 105 to be efficient and highly integrated with the status and operation of the vehicle 202 ( FIGS. 2-4 ) in contrast to an unmanned aerial vehicle equipped with a more comprehensive distance detector and without a directional controller (e.g., 109 ) performing nondiscriminatory or rote data collection operations.
- the added degrees of freedom using the directional controller 109 can provide better integration between the distance detector system 101 and the autonomous vehicle 202 , allow targeted or tailored scope of detection, and compensate for inherent limits of the distance detector or physical limits of the unmanned aerial vehicle.
- FIG. 2 is an illustration of an unmanned aerial vehicle having a distance detector system according to one embodiment.
- FIG. 2 shows the autonomous or semi-autonomous vehicle 202 having the distance detector system 101 mounted on the bottom of a main body 204 of the vehicle 202 .
- one or more of the distance detector system 101 can be mounted on other parts of the main body 204 , such as the top or the side(s), or the distance detector system 101 can be otherwise attached to, coupled to, or incorporated into the vehicle 202 as part of the vehicle 202 .
- the distance detector system 101 can be at least partly internal to the main body 204 and/or configured to be at least partially stowed in and unstowed from the main body 204 .
- the autonomous or semi-autonomous vehicle 202 can be an unmanned aerial vehicle (UAV), and in other embodiments the vehicle 202 can be a terrestrial or part-terrestrial, part-aerial, and/or amphibious vehicle operating autonomously or semi-autonomously.
- UAV unmanned aerial vehicle
- 2-3 is a quadcopter having four arms with their respective rotors
- the disclosure herein can be implemented in other types of autonomous or semi-autonomous aerial and/or terrestrial vehicles such as a multirotor helicopter having a different number of arms and/or rotors, an aerial vehicle other than a multirotor helicopter such as a fixed wing aerial vehicle, or other types of autonomous or semi-autonomous vehicles such as cars or robots requiring distance detection of the surroundings.
- FIG. 3 is an illustration of an example operation of an unmanned aerial vehicle having a distance detector system according to one embodiment.
- FIG. 3 shows the vehicle 202 with the distance detector system 101 attached traveling along a travel path 302 based on travel instructions.
- FIG. 3 also shows an obstacle 306 in the vicinity of the vehicle 202 near or on the travel path 302 , and a zone or space of interest 304 for navigation by the vehicle 202 .
- the vehicle 202 can determine or be provided with the travel path 302 based on, for example, the overall mission of the travel.
- the vehicle 202 may execute current travel instructions, which can be understood as instructions for the vehicle 202 to travel from its current location (e.g., a point on the travel path 302 ) to immediately following series of locations (e.g., traveling through a series of points on the travel path 302 ) for a certain duration of time and/or distance in space.
- current travel instructions can be understood as instructions for the vehicle 202 to travel from its current location (e.g., a point on the travel path 302 ) to immediately following series of locations (e.g., traveling through a series of points on the travel path 302 ) for a certain duration of time and/or distance in space.
- one or more travel instructions to follow a part of the travel path 302 from the current position of the vehicle 202 to a point on the travel path 302 the vehicle 202 is instructed or expected to be in the next few seconds (e.g., 3 seconds) can constitute current travel instructions.
- the travel path 302 can be predetermined, and the vehicle 202 may be provided with an a priori knowledge of the full expected travel path, and in other instances, the travel path 302 can be determined as the travel occurs on a case-by-case or ad hoc basis. In some instances, the travel path 302 can be tentatively pre-computed based on the real time information gathered by the vehicle 202 or the status of the vehicle 202 , and the vehicle 202 can be configured to obtain, determine, and/or adjust its travel path based on the combination of known or predetermined information and on-going gathering of data.
- This short forward-looking time or distance can be understood as a clearance time or distance for the vehicle 202 to detect obstructions, if any, along the short expected path of travel.
- the illustrated example in FIG. 3 shows that the vehicle 202 plans to turn around soon as it travels along the travel path 302 , which may have been preplanned before the start of the mission.
- the distance detector system 101 can orient the distance detector 105 by rotating the distance detector 105 about the axes 102 , 104 in a manner discussed above in connection with FIG. 1 to observe the space of interest 304 around the planned travel path 302 before the vehicle 202 reaches the space of interest 304 .
- the vehicle 202 could identify that there is the obstacle 306 partially blocking the space of interest 304 and may determine whether it should revise or alter the travel path 302 .
- the distance detector system 101 may be observing ahead of the vehicle 202 along the travel path 302 by a few seconds or a few meters in accordance with current travel instructions. For example, before the vehicle 202 takes off, the distance detector system 101 may have the distance detector 105 pointed to the direction of the takeoff so that the vehicle 202 can determine that its takeoff space is clear of obstacles. As the vehicle 202 takes off the distance detector system 101 can continue to observe its travel path ahead of the vehicle 202 throughout the travel, which can be completed with a landing. The duration and/or distance of advanced observation can be determined based, for example, on the speed of the vehicle 202 , the unexpected nature of the environment, sensing of changes in the environment, detection of target object(s), emergency situation(s), or any combination thereof.
- the vehicle 202 may travel at a slower speed and use a shorter clearance time than it would otherwise.
- the environment the vehicle 202 is traveling includes one or more animate objects (e.g., other vehicles, animals, etc.)
- the vehicle 202 may travel at a slower speed and use a shorter clearance time than it would otherwise as sudden change or movements by the animate objects can create unexpected obstructions to the vehicle 202 .
- the vehicle 202 may travel faster and the length of the clearance time or distance may not be as critical for its travel.
- the vehicle 202 can also or alternatively look ahead based on detected objects or surroundings or the mission objectives.
- the vehicle 202 may be provided with a high-level mission without detailed specifications as to the travel path 302 (i.e., lacking a priori knowledge at least in part).
- the vehicle 202 may constantly detect its surroundings a short time or distance ahead to ensure a clear path of travel as the vehicle 202 performs its mission. Even for travels involving a predetermined travel path, the vehicle 202 nonetheless may in part engage in on-the-fly determination of its own path in various situations, such as emergency, sudden encountering of unexpected obstructions, etc., until the vehicle 202 gets situated in a better known environment, for example.
- the vehicle 202 may follow along a wall at a certain distance, either for guidance (i.e., path planning) or for a specific objective (e.g., identify defects of the wall). As the vehicle 202 follows the wall in this example, the vehicle 202 may look ahead along the tangent lines of the surface of the wall to detect if there are any obstructions on or near the wall or any protrusions from the wall. Similarly, the vehicle 202 may detect other objects in the surroundings and look ahead based on detected edges, corners, surfaces or other features of the objects either to avoid, follow side-by-side, or perform other functions.
- the vehicle 202 may perform tangent-line look ahead based on the physical or geometric features of the objects according to, for example, statistical estimation, modeling, extrapolation, or other similar mechanisms to determine in what direction the distance detector system 101 should be configured to detect as the vehicle 202 travels along the surfaces.
- the vehicle 202 may store, compare, and/or otherwise process distance data that may or may not be on its path forward using the distance detector system 101 .
- the vehicle 202 may not only gather distance data in the direction of takeoff and the travel path, but also the area below the vehicle 202 as it takes off from ground if, for instance, the vehicle 202 plans to return to the same spot for landing.
- the vehicle 202 may store the distance data around the planned landing or return area gathered during takeoff, and the stored distance data information can be compared with the forward looking (i.e., along the travel path) distance data as the vehicle 202 prepares to land upon completion of a mission, for example.
- the vehicle 202 may determine the change of environment, appearance or movements of objects, and other relevant distance information along a timeline, in order to, for example, avoid obstacles, alter its course, abort or delay landing, and find a safe and flat landing spot.
- the vehicle 202 may be configured to collect distance data as payload data for various purposes. For instance, the vehicle 202 may travel for the specific purpose of gathering distance information with respect to the terrains, environment, or a particular object. In some instances, the vehicle 202 can use distance data not only to clear or plan its travel path, but also to locate itself either relatively or absolutely based in part on known or previously detected distance information. For example, the vehicle 202 may travel a closed or narrow space, and the vehicle 202 may be able to locate itself relative to the walls or other boundaries of the closed or narrow space. In another example, the vehicle 202 may gather distance information in relation to objects or terrains of known absolute locations so that the vehicle 202 can determine its absolute location. In other instances, the vehicle 202 can use the distance information gathered using the distance detector system 101 to aid other functionalities or performances, such as magnetometer or other sensor calibration, vehicle state estimation, or emergency landing, that can utilize and benefit from the distance data.
- the vehicle 202 can use the distance information gathered using the distance detector system 101 to aid other functionalities or performances,
- one or more processors may direct the directional controller 109 ( FIGS. 1, 4 ) of the distance detector system 101 according to direction priorities, such as clearing a travel path, detecting and avoiding obstructions, gathering payload data, spotting an object of interest, locating the vehicle 202 itself, performing targeted sensing objectives, aiding or configuring other vehicle functionalities, etc. for the distance detector 105 as discussed herein.
- direction priorities such as clearing a travel path, detecting and avoiding obstructions, gathering payload data, spotting an object of interest, locating the vehicle 202 itself, performing targeted sensing objectives, aiding or configuring other vehicle functionalities, etc.
- obstruction avoidance along the travel path can be based on current travel instructions, which may allow the one or more processors to determine and update the direction priorities.
- the instructions for the distance detector system 101 can be highly integrated with other instructions.
- the distance detector system 101 in conjunction with the one or more processors can be configured to interleave or alternate between the multiple functionalities as the vehicle 202 travels.
- same distance data can be stored, processed, and serve multiple purposes for a variety of unrelated functionalities.
- the direction priorities can be determined periodically or aperiodically. For instance, determination of direction priorities, and thus configuring of the directional controller 109 ( FIGS. 1, 4 ) to move or rotate the distance detector 105 , can range from a predictable scenario of regular priority updates merely confirming no obstruction along a pre-planned travel path to an unpredictable scenario of interruptive priority updates (e.g., forgoing payload data gathering to plan an emergency travel path).
- the disclosed technology provides advantages in utilizing the distance detector 105 in an effective and intelligent manner in addition to the advantages in extending physical scope of detection capabilities described above in connection with FIG. 1 .
- the vehicle 202 may not acquire or process some distance data.
- the directional controller 109 FIGS. 1, 4 ) allows the extended use of the distance detector 105 to be integrated with the execution of various instructions by one or more processors (e.g., 110 in FIG. 4 ), which in turn allows high efficiency, agility, and responsiveness to high-level processes and determinations performed by the vehicle 202 .
- a vehicle 202 using a comprehensive distance detector can benefit from the disclosure herein as it allows efficient use of the comprehensive distance detector avoiding rote data collection and handling or processing of unnecessary data.
- the added intelligence from the vehicle 202 incorporated into the distance detector system 101 may allow targeted sensing or gathering and efficient storing of the gathered data.
- FIG. 4 is a diagram showing components of an example unmanned aerial vehicle according to one embodiment.
- the vehicle 202 illustrated in FIG. 2 includes one or more processor(s) 110 in communication with a state estimator which may be an inertial measurement unit (IMU) 112 .
- the processor 110 is in further communication with one or more transceivers 108 , sensors 115 , a camera 111 , a global positioning system (GPS) module 114 , memory 124 , and motor controllers 120 , which are in communication with motors 122 .
- the processor 110 is in further communication with the distance detector system 101 , which may include the distance detector 105 (e.g., LIDAR), and the directional controller 109 as discussed above in connection with FIGS. 1-3 .
- the distance detector system 101 may include the distance detector 105 (e.g., LIDAR), and the directional controller 109 as discussed above in connection with FIGS. 1-3 .
- the vehicle 202 further includes a power supply 116 and a battery 118 , which provides power to one or more modules of the vehicle 202 , including the processor 110 .
- the transceivers 108 and the GPS module 114 may be in further communication with their respective antennas (not shown).
- the memory 124 may store one or more of mission instructions, travel instructions, pre-mission routines, payload data, flight data and/or telemetry, settings (e.g., calibration settings), parameters, or other similarly relevant information or data.
- the vehicle 202 may also include a recovery system 106 , which may be in communication with one or more modules in the vehicle 202 , including the processor 110 .
- the recovery system 106 may include a dedicated recovery processor (not shown) in communication with a recovery state estimator (not shown), which may also be an additional IMU (not shown).
- the vehicle 202 may include additional or intermediate modules, drivers, controllers, circuitries, lines of communication, and/or signals not illustrated in FIG. 4 .
- the vehicle 202 can perform its regular operation according to instructions executed by the processor 110 to, for example, take a course of action for a mission.
- the processor 110 can be a microprocessor capable of communicating with various modules illustrated in FIG. 4 executing instructions pre-programmed and/or received during a mission, for example.
- the processor 110 may receive settings, values, or parameters stored in the memory 124 and data from the sensors 115 , the distance detector system 101 , the camera 111 , the transceivers 108 , the GPS module 114 , the IMU 112 , and the motor controllers 120 to evaluate the status of the vehicle 202 and determine a course of action.
- the status of the vehicle 202 can be determined based on data received through the sensors 115 , the distance detector system 101 , and/or preloaded data stored in the memory 124 and accessed by the processor 110 .
- the altitude of the vehicle 202 above ground can be determined by the processor 108 based on a digital elevation model (DEM) of a world elevation map or with a LIDAR (e.g., from the distance detector system 101 ), a barometer, or ultrasound.
- the vehicle 202 may include multiple processors of varying levels of computing power and reliability to execute low-level instructions or run high-level application code or a virtual machine.
- one or more of the functionalities of the processor(s) 110 described herein may instead be performed by another processor in the vehicle 202 .
- the transceivers 108 can be devices capable of transmitting and receiving data to and from a system, device, or module external to the vehicle 202 .
- the transceivers 108 may include radio frequency (RF) transceivers capable of communicating data over a Wi-Fi network or any other suitable network in various frequency bands or channels, such as 900 MHz, 2.4 GHz, 5 GHz, etc.
- RF radio frequency
- the transceivers 108 may be implemented with a combination of separate transmitters and receivers.
- the motor controllers 120 may include a controller device or circuit configured to interface between the processor 110 and the motors 122 for regulating and controlling speed, velocity, torque, or other operational parameters of their respective, coupled motors 122 .
- one or more motor control schemes may be implemented with the processor 110 and/or the motor controllers 120 .
- the motors 122 may include electrical or any other suitable motors coupled to their respective rotors of the vehicle 202 to control their propellers, for example.
- the memory 124 can be a memory storage device (e.g., random-access memory, read-only memory, flash memory, or solid state drive (SSD) storage) to store data collected from the sensors 115 , data processed in the processor 110 , or preloaded data, parameters, or instructions.
- the memory 124 may store data gathered from the distance detector 105 using various computationally efficient data structures. For example, in some cases, the distance data from the distance detector 105 can be stored using a three-dimensional occupancy grid mapping, with the gathered data grouped into cube-shaped bins of variable resolution in space.
- the resolution of the occupancy grid can be determined to indicate whether each variable resolution bin within the reach of the distance detector is free or occupied based on the gathered distance data.
- the three-dimensional occupancy mapping values can be estimated using probabilistic approaches based on the gathered distance data.
- the IMU 112 may include a stand-alone IMU chip containing one or more magnetometers, gyroscopes, accelerometers, and/or barometers. In some embodiments, the IMU 112 may be implemented using a combination of multiple chips or modules configured to perform, for example, measuring of magnetic fields and vehicle orientation and acceleration and to generate related data for further processing with the processor 110 . Regardless of integrated or multi-module implementation of the IMU 112 , the term “magnetometer” as used herein, generally refers to the part(s) of the IMU 112 responsible for measuring the magnetic field at the location of the vehicle 202 .
- the recovery system 106 can be responsible for recovery operation of the vehicle 202 to, for example, safely deploy a parachute and land the vehicle 202 .
- the recovery system 106 may include a parachute (not shown) and an electromechanical deployment mechanism (not shown).
- the power supply 116 may include circuitry such as voltage regulators with outputs directly powering various modules of the vehicle 202 with Vcc_vehicle, and the battery 118 can provide power to the power supply 116 .
- the battery 118 can be a multi-cell lithium battery or any other suitable battery capable of powering the vehicle 202 .
- the battery 118 of the vehicle 202 can be removable for easy swapping and charging.
- the sensors 115 may include one or more proximity sensors using, for example, infrared, radar, sonar, ultrasound, LIDAR, barometer, and/or optical technology.
- the sensors 115 may also include other types of sensors gathering data regarding visual fields, auditory signals, and/or environmental conditions (e.g., temperature, humidity, pressure, etc.).
- the GPS module 114 may include a GPS transceiver and/or a GPS driver configured to receive raw and/or processed GPS data such as ephemerides for further processing within the GPS module 114 , with the processor 110 , or both.
- the vehicle 202 may also include a microphone (not shown) to gather audio data. In some embodiments, one or more sensors 115 responsible for gathering data regarding auditory signals can take the place of the microphone.
- the distance detector system 101 can include, among others, the distance detector 105 and the directional controller 109 .
- the distance detector 105 can include a LIDAR sensor, such as a one-dimensional LIDAR sensor.
- a one-dimensional LIDAR having, for example, the range of about 50 m with the sampling frequency up to about 30 kHz can be used, and the directional controller 109 providing a two-dimensional gimbal whose rotational frequency about one axis can be at about 2-3 Hz for a vehicle 202 traveling at about 10 m/s or less.
- the vehicle 202 can gather distance data with an angular resolution of approximately 0.1°, for example.
- the directional controller 109 can be a motorized directional controller, which includes one or more brushless motors with a slip ring to implement a gimbal system. In other embodiments, the directional controller 109 be otherwise configured to move or rotate the distance detector 105 without embedded motors. In some embodiments, the directional controller 109 may further include drivers such as a controller device or circuit configured to interface between the processor 110 and the motors of the directional controller 109 for regulating and controlling speed, velocity, torque, or other operational parameters of their respective, coupled motors.
- drivers such as a controller device or circuit configured to interface between the processor 110 and the motors of the directional controller 109 for regulating and controlling speed, velocity, torque, or other operational parameters of their respective, coupled motors.
- the camera 111 can be configured to gather images and/or video.
- one or more of the sensors 115 and the distance detector 105 (in conjunction with other modules of the distance detector system 101 ) responsible for gathering data regarding visual fields can take the place of the camera 111 .
- the sensors 115 , the distance detector 105 (in conjunction with other modules of the distance detector system 101 ), and/or the camera 111 may be configured to gather parts of payload data, which includes data gathered by the vehicle 202 regarding its surroundings, such as images, video, and/or processed 3D mapping data, gathered for purposes of mission performance and/or delivered to the user for various purposes such as surveillance, observation, progress report, landscape analysis, etc.
- the sensors 115 may also gather what may be termed telemetry data, which is data regarding the status and activities of the vehicle 202 during the flight such as velocity, position, attitude, temperature, and rotor speeds. Such data may be collected to retain records or logs of flight activity and perform diagnostics.
- the sensors 115 , the distance detector 105 (in conjunction with other modules of the distance detector system 101 ), and/or the camera 111 may also be configured to gather data for purposes of aiding navigation and obstruction detection as disclosed herein.
- FIG. 5 is a flowchart for enhanced detection according to one embodiment.
- the illustrated process 500 can be performed in part by and/or in conjunction with one or more modules in the vehicle 202 ( FIG. 4 ), such as the processor(s) 110 ( FIG. 4 ), the distance detector system 101 ( FIG. 4 ), the distance detector 105 ( FIG. 4 ), the directional controller 109 ( FIG. 4 ), the sensors 115 ( FIG. 4 ), the camera 111 ( FIG. 4 ), the memory 124 ( FIG. 4 ), the GPS module 114 ( FIG. 4 ), the IMU 112 ( FIG. 4 ), and the motor controllers 120 ( FIG. 4 ).
- the processor(s) 110 FIG. 4
- the distance detector system 101 FIG. 4
- the distance detector 105 FIG. 4
- the directional controller 109 FIG. 4
- the sensors 115 FIG. 4
- the camera 111 FIG. 4
- the memory 124 FIG. 4
- the GPS module 114 FIG. 4
- IMU 112 FIG. 4
- steps 502 , 504 , 506 , 508 , 510 , and 512 may be concurrently, continuously, periodically, intermittently, repeatedly, or iteratively performed, and the illustrated process in FIG. 5 is only one example embodiment of inventive features disclosed herein.
- the process 500 may proceed to step 502 .
- the vehicle 202 initiates its mission.
- the mission can include high-level objectives, a predetermined travel path, and/or distance-specific instructions (e.g., follow a wall at a certain offset distance).
- distance-specific instructions e.g., follow a wall at a certain offset distance.
- current travel instructions can be continually determined and executed for a short window of time or distance ahead.
- one or more processors 110 of the vehicle 202 determines direction priorities based at least in part of the current travel instructions.
- the current travel instructions may encompass a short path along the planned travel path of the vehicle, and the distance detector system 101 of the vehicle 202 may, for example, determine whether, how, and to what extent the distance detector 105 gathers distance data regarding the vehicle's expected path forward and/or its surroundings.
- the direction priorities can be further based on other mission objectives (e.g., payload data gathering, locating the vehicle 202 relative to its surroundings) or unanticipated situations the vehicle 202 may encounter.
- the directional controller 109 of the vehicle 202 rotates distance detector 105 about first and second axes.
- the directional controller 109 may include motors to implement a motorized gimbal system that may extend the detection dimension of the distance detector 105 .
- the one or more processors 110 of the vehicle 202 acquire and/or process the distance data from the distance detector 105 based at least in part on the direction priorities determined in step 504 .
- the direction priorities may, for example, indicate that the vehicle 202 should confirm the travel path up to 5 seconds ahead should be clear of obstacles while the distance detector system 101 gathers intermittent distance data to the ground as the vehicle 202 travels.
- steps 504 , 506 , and 508 may be repeated throughout the mission.
- the vehicle 202 may optionally detect one or more obstacles with the distance detector 105 in a direction based in part on the direction priorities in step 504 .
- the distance detector 105 may gather data according to direction priorities to determine whether there are obstacles, and the processor 110 may determine that there is an obstacle on or near the vehicle's immediate travel path based on the distance data.
- the vehicle 202 may optionally adjust its travel path. For example, in case an obstruction is detected on or near the vehicle's planned travel path, the one or more processors 110 may determine that the obstruction would hinder the vehicle's travel as planned and determine that its travel path be adjusted. Upon the determination of path adjustment, the distance detector system 101 may further gather distance data of the vehicle's surroundings to find an alternative path to travel. In some embodiments, the vehicle 202 may constantly plan and/or adjust its path forward based on constant gathering of distance data of its surroundings. In other embodiments, the vehicle 202 may determine an alternative path forward, store the alternative path, and perform obstacle clearing steps similar to steps 504 , 506 , and 508 along the alternative path. The process 500 in FIG.
- step 5 only shows one iteration of the process that can be performed by the vehicle 202 , and in some embodiments, parts of the process 500 may repeat in an iterative loop. For example, after performing step 512 , the process may loop back to step 504 , and the vehicle 202 may determine direction priorities and proceed to steps 506 , 508 , 510 , and 512 . In such example, the process performed by the vehicle 202 may repeatedly loop back from step 512 to step 504 until, for instance, the vehicle 202 completes its mission.
- determining encompasses a wide variety of actions. For example, “determining” may include calculating, computing, processing, deriving, investigating, looking up (e.g., looking up in a table, a database or another data structure), ascertaining and the like. Also, “determining” may include receiving (e.g., receiving information), accessing (e.g., accessing data in a memory) and the like. Also, “determining” may include resolving, selecting, choosing, establishing and the like. Further, a “channel width” as used herein may encompass or may also be referred to as a bandwidth in certain aspects.
- any suitable means capable of performing the operations such as various hardware and/or software component(s), circuits, and/or module(s).
- any operations illustrated in the Figures may be performed by corresponding functional means capable of performing the operations.
- DSP digital signal processor
- ASIC application specific integrated circuit
- FPGA field programmable gate array signal
- PLD programmable logic device
- a general-purpose processor may be a microprocessor, but in the alternative, the processor may be any commercially available processor, controller, microcontroller or state machine.
- a processor may also be implemented as a combination of computing devices, e.g., a combination of a DSP and a microprocessor, a plurality of microprocessors, one or more microprocessors in conjunction with a DSP core, or any other such configuration.
- the methods disclosed herein comprise one or more steps or actions for achieving the described method.
- the method steps and/or actions may be interchanged with one another without departing from the scope of the claims.
- the order and/or use of specific steps and/or actions may be modified without departing from the scope of the claims.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Remote Sensing (AREA)
- Radar, Positioning & Navigation (AREA)
- General Physics & Mathematics (AREA)
- Computer Networks & Wireless Communication (AREA)
- Electromagnetism (AREA)
- Architecture (AREA)
- Aviation & Aerospace Engineering (AREA)
- Structural Engineering (AREA)
- Civil Engineering (AREA)
- Automation & Control Theory (AREA)
- Control Of Position, Course, Altitude, Or Attitude Of Moving Bodies (AREA)
- Traffic Control Systems (AREA)
Abstract
Description
- Field
- The described technology generally relates to unmanned aerial vehicles and, more specifically, to enhanced distance detection systems of unmanned aerial vehicles.
- Description of the Related Art
- An unmanned aerial vehicle, also commonly referred to as drone, can travel through a variety of environments, such as indoor, outdoor, and/or mixed indoor and outdoor environments. In some cases, an unmanned aerial vehicle can be configured to conduct surveillance, mapping, delivery, monitoring, or other tasks that can comprise combining movement and data collection. The unmanned aerial vehicle can travel over surfaces on which the unmanned aerial vehicle cannot safely land (e.g., water).
- The methods and devices of the described technology each have several aspects, no single one of which is solely responsible for its desirable attributes.
- In one embodiment, an unmanned aerial vehicle includes a distance detector, a directional controller coupled to the distance detector, and one or more processors. The one or more processors are configured to execute a mission of the unmanned aerial vehicle, determine direction priorities based at least in part on current travel instructions of the mission, and acquire and/or process distance data generated by the distance detector based at least in part on the determined direction priorities.
- In another embodiment, a method for targeted sensing for an unmanned aerial vehicle includes initiating a mission, determining direction priorities based at least in part on current travel instructions of the mission, controlling an orientation of a distance detector, and acquiring and/or processing distance data from the distance detector based at least in part on the determined direction priorities.
- These drawings and the associated description herein are provided to illustrate specific embodiments of the described technology and are not intended to be limiting.
-
FIG. 1 is a distance detector system according to one embodiment. -
FIG. 2 is an illustration of an unmanned aerial vehicle having a distance detector system according to one embodiment. -
FIG. 3 is an illustration of an example operation of an unmanned aerial vehicle having a distance detector system according to one embodiment. -
FIG. 4 is a block diagram of an unmanned aerial vehicle according to one embodiment. -
FIG. 5 is a flowchart for enhanced detection according to one embodiment. - Various aspects of the novel systems, apparatuses, and methods are described more fully hereinafter with reference to the accompanying drawings. Aspects of this disclosure may, however, be embodied in many different forms and should not be construed as limited to any specific structure or function presented throughout this disclosure. Rather, these aspects are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the disclosure to those skilled in the art. Based on the teachings herein, one skilled in the art should appreciate that the scope of the disclosure is intended to cover any aspect of the novel systems, apparatuses, and methods disclosed herein, whether implemented independently of or combined with any other aspect. For example, an apparatus may be implemented or a method may be practiced using any number of the aspects set forth herein. In addition, the scope is intended to encompass apparatus and/or methods which are practiced using structure and/or functionality in addition to or different than the various aspects specifically set forth herein. It should be understood that any aspect disclosed herein might be embodied by one or more elements of a claim.
- Although particular aspects are described herein, many variations and permutations of these aspects fall within the scope of the disclosure. Although some benefits and advantages of the preferred aspects are mentioned, the scope of the disclosure is not intended to be limited to particular benefits, uses, or objectives. Rather, aspects of the disclosure are intended to be broadly applicable to different wired and wireless technologies, system configurations, networks, including optical networks, hard disks, and transmission protocols, some of which are illustrated by way of example in the figures and in the following description of the preferred aspects. The detailed description and drawings are merely illustrative of the disclosure rather than limiting, the scope of the disclosure being defined by the appended claims and equivalents thereof.
- The term “autonomous vehicle” or “semi-autonomous vehicle,” as used herein, generally refers to a vehicle that is configured to operate without substantial or any involvement from an on-board operator (e.g., a driver or pilot). An “unmanned aerial vehicle,” or “UAV,” as used herein, can denote a type of autonomous or semi-autonomous vehicle whose physical operational capabilities include aerial travel or flight. An unmanned aerial vehicle can be an aircraft that is configured to take off and land on a surface. In some cases, an unmanned aerial vehicle can automatically travel from one location to another without any operator involvement. In some cases, an unmanned aerial vehicle can travel a far distance from a starting point. The distance can be far enough that the unmanned aerial vehicle cannot return to a starting point without refueling or recharging at an intermediate location. An unmanned aerial vehicle can be configured to land on a landing pad and/or charge at a charging station.
- An enhanced distance detection system for an autonomous or semi-autonomous vehicle is described herein. The distance detection system includes a distance detector, which may have a limited scope of distance detection, and a directional controller, which allows extending the dimension or scope of the distance detector as the vehicle travels and performs missions. The directional controller can change the detection direction of the distance detector with a motorized gimbal or functionally similar system, and the change in the detection direction can be integrated with the status of and/or instructions executed or to be executed by the vehicle.
-
FIG. 1 is a distance detector system according to one embodiment. The illustrateddistance detector system 101 includes adistance detector 105 and a motorizeddirectional controller 109. Also shown inFIG. 1 are reference Cartesian and spherical coordinates, which are shown only for ease of explanation and are not considered limiting the implementation of thedistance detector system 101. In some embodiments thedistance detector 105 can be implemented using at least one LIDAR, such as a one-, two-, or three-dimensional LIDAR, which has a field ofdetection 103. In other embodiments thedistance detector 105 can be implemented with technologies using other suitable form of signal or energy, such as radar, structured light, ultrasound, etc. The “field of detection” as used herein generally refers to one or more lines or paths of emission and detection of laser (or any other suitable form of signal or energy for distance detection) from and to the distance detector (e.g., 105) as defined by the orientation and internal configurations native or inherent in the design of the distance detector as one unit. For example, a one-dimensional LIDAR may have one emitter-detector pair side by side pointing at the same direction in parallel, and the field of detection of the one-dimensional LIDAR can be a unit-resolution line defined by a unit-resolution point in the space at which the emitter-detector pair is pointing. A higher dimensional LIDAR may have more than one emitter-detector pair and/or have internal mechanical and/or optical mechanisms to emit and detect radiation in more than one direction. In such a case, the field of detection can be an accumulation of the lines of emission and detection defined by multiple points in the space (e.g., a perimeter of a surface or a surface of a volume). - In some embodiments, the motorized
directional controller 109 can be implemented with a generally L-shaped arm gimbal system providing additional degrees of freedom (e.g., in θ and φ inFIG. 1 ) about 102, 104 to allow expanding upon or augmenting the field ofadditional axes detection 103 native to thedistance detector 105. In other embodiments, the gimbal system illustrated inFIG. 1 can be implemented with a different mechanical structure, such as a U-shaped arm, that can similarly provide additional degrees of freedom. The illustrateddirectional controller 109 includes two hubs allowing thedistance detector 105 to rotate about the 102 and 104. Rotations about theaxes 102 and 104 can be controlled by motors (not shown) and motor controllers (not shown). For example, a brushless motor with a slip ring can be implemented to allow rotations about theaxes axis 102, and another brushless motor can be implemented to allow rotations about theaxis 104. In some embodiments, thedirectional controller 109 can provide continual rotations about one or more of the 102, 104, and in other embodiments, theaxes directional controller 109 can provide limited or non-continual rotations about one or more of the 102, 104. For example, in certain implementations, theaxes directional controller 109 can allow unlimited rotation about the axis 102 (i.e., unlimited in φ inFIG. 1 ) so that thedistance detector 105 can rotate about theaxis 102 while rotation about theaxis 104 may have a range of 180° (e.g., 0° to 180° in θ inFIG. 1 ). The range of rotations about the 102, 104 can vary depending on, for example, the purpose, scope, and suitable level of redundancy in expanding the native field ofaxes detection 103. - The angle of orientation about the
102 and 104 at a given moment can each be measured at each hub by one or more sensors (e.g., optical encoder or angle sensors, such as Hall effect sensors or magneto-resistive sensors in combination with a dipole or ring magnet, etc.) and can be communicated to a processor 110 (axes FIG. 4 ). In some embodiments, the sensors, motors, and motor controllers of thedistance detector system 101 can be dedicated sensors (not shown), motors (not shown), and motor controllers (not shown) specific to thedistance detector system 101. In other embodiments, one or more of the sensors, motors, and motor controllers in thedistance detector system 101 can be at least partially integrated with sensors 115 (FIG. 4 ), motors 112 (FIG. 4 ), and motor controllers 120 (FIG. 4 ) configured to function with other parts of an autonomous vehicle 202 (FIG. 4 ). - In other embodiments, the
directional controller 109 can be motorless and implemented without dedicated motors controlling the added degrees of freedom discussed above. In some embodiments, thedirectional controller 109 can be implemented with other mechanical systems, such as one or more ball and socket joints, hinges, gears, chain, belts, pulleys, or other mechanical couplings, to otherwise add additional directional dimensions to thedistance detector 105, or to expand the scope of coverage of thedistance detector 105. Also, in some embodiments, the added additional degrees of freedom can be more or less depending on the native scope of the distance detector 105 (e.g., the field of detection 103). For example, thedistance detector 105 can be a two-dimensional LIDAR, and thedirectional controller 109 may only provide one additional rotational degree of freedom. In other embodiments, thedistance detector 105 can be a two- or three-dimensional LIDAR having a limited scope or breath of distance detection (e.g., incomplete circumference or sphere), and one or more degrees of freedom about the additional axes (e.g., 102, 104) can be added with thedirectional controller 109 to expand upon the scope of distance detection (e.g., the field of detection 103) native to thedistance detector 105. In other embodiments, thedirectional controller 109 can be implemented to provide additional one or more translational degrees of freedom (e.g., x, y, z, r, or any linear combination thereof) in addition to or instead of the rotational degrees of freedom, for example, to address mechanical blockage in the field ofdetection 103, or for ease of stowage of all or part of thedistance detector system 101. In some embodiments, the added degrees of freedom can be orthogonal, or fully independent, to each other and to the native field of detection (e.g., 103) for efficiency, but in other embodiments, the added degrees of freedom can be at least partially overlapping or redundant or not fully independent or orthogonal, depending on, for example, the physical constraints of theautonomous vehicle 202 and the purpose of the added degrees of freedom. Example embodiments of mechanical integration of thedistance detector system 101 and the autonomous vehicle 202 (FIG. 2 ) are further discussed in connection withFIG. 2 below. - In some instances, the
directional controller 109 can be configured to rotate thedistance detector 105 at a constant speed in at least one direction, and in other instances thedirectional controller 109 can be configured to rotate thedistance detector 105 in a directed, targeted, and non-constant manner that is integrated with execution of a mission or the status of the autonomous vehicle 202 (FIG. 2 ). For example, in some implementations, thedirectional controller 109 can be configured to rotate thedistance detector 105 about thefirst axis 102 constantly at about 3 Hz (i.e., three full-360° rotations in φ per second) for an unmanned aerial vehicle moving at about 10 m/s or less. Thedirectional controller 109 can be further configured to rotate thedistance detector 105 about thesecond axis 104 back and forth depending on the area of interest as the unmanned aerial vehicle operates (e.g., 90°±Δθ with Δθ being a small angle such as 5° if interested in detecting the surroundings at level with the vehicle altitude). Rotating about theaxis 102 at a constant speed while slightly varying the angle (e.g., θ) about theaxis 104 allows sensing and data collection of a three dimensional space of a partial sphere around the vehicle. For example, if thedirectional controller 109 rotates thedistance detector 105 at 3 Hz and varies θ to three values (e.g., 85°, 90°, and 95°), the data from the partial cone space defined by the location of thedistance detector 105 and the range of θ are gathered three times per second. The angle of interest in θ can be different depending on, for example, relative altitudes of thevehicle 202 at its current location and the direction to which it is traveling (e.g., if traveling to a higher altitude, θ can be less than 90°; if traveling to a lower altitude, θ can be greater than) 90°. Also, the variation of angle of rotation about the second axis 104 (e.g., ±Δθ or any other range of θ which is not necessarily symmetric about the angle based on the relative altitudes), and the number of datapoints to gather at a given φ and a range of θ can be determined based on the speed and/or status of thevehicle 202, the objective of the data collection, detection of obstacle(s), electromechanical constraints of thevehicle 202, or any other suitable parameters or conditions. - In other embodiments, the unmanned aerial vehicle may travel at a faster or slower speed than the example discussed above, and the
directional controller 109 can be configured to rotate or move thedistance detector 105 at a higher or lower speed as appropriate. The appropriate speed (or frequency) of the rotation can also depend on, for example, mechanical constraints within the motorizeddirectional controller 109, such as speed tolerance of the slip ring of the motor. In some embodiments, thedirectional controller 109 can be configured to enable and disable, start and stop, or accelerate and decelerate rotation or movements of thedistance detector 105 depending on, for example, mission objectives, mission stage (e.g., takeoff, travel, landing), emergency situations, environmental conditions, power level, duration of the mission, etc. Example operations of thedistance detector system 101 are further discussed in connection withFIG. 3 below. - It can be advantageous to implement the
distance detector system 101 to augment or expand upon the native scope (e.g., the line of detection 103) of the distance detector (e.g., 105) since the added degrees of freedom by thedirectional controller 109 may allow effective and efficient implementation of thedistance detector system 101 with reduced weight, cost, or complexity associated with a more comprehensive or exhaustive distance detector (e.g., three-dimensional LIDAR with full spherical scope of detection). Furthermore, it can be advantageous to implement thedistance detector system 101 as described herein since thesystem 101 can take advantage of high precision distance sensing of a limited scope distance detector, such as a one-dimensional LIDAR (as opposed to an ultrasound sensor), without suffering from the disadvantage of its limited scope. For instance, the internal mechanisms of a two-dimensional LIDAR may operate by spreading out transmission power over a larger area, which may render outdoor performance and range worse than a one-dimensional LIDAR at a given point in space, and a high-precision one-dimensional LIDAR rotated as needed can provide more advantages in performance over a standalone two-dimensional LIDAR. In some embodiments, thedistance detector system 101 as disclosed herein can achieve the resolution of, for example, 1 cm +/− in detecting the location of one or more objects in the relevant partial space surrounding thevehicle 202 during takeoff, travel, and landing with a cost-effective, lightweight one-dimensional LIDAR. As further described below in connection withFIG. 4 , in some embodiments, data collected from thedistance detector 105 can be stored in a data structure involving variable resolution (e.g., resolution as low as 1 m) regardless of the full resolution or data collection capabilities of thedistance detector 105 in order to improve efficiency in computation, memory, processing time, etc. As described herein, thedistance detector system 101 allows detection, collection, and sensing operations of thedistance detector 105 to be efficient and highly integrated with the status and operation of the vehicle 202 (FIGS. 2-4 ) in contrast to an unmanned aerial vehicle equipped with a more comprehensive distance detector and without a directional controller (e.g., 109) performing nondiscriminatory or rote data collection operations. Also, even for thedistance detector system 101 implemented with a more comprehensive distance detector, the added degrees of freedom using thedirectional controller 109 can provide better integration between thedistance detector system 101 and theautonomous vehicle 202, allow targeted or tailored scope of detection, and compensate for inherent limits of the distance detector or physical limits of the unmanned aerial vehicle. -
FIG. 2 is an illustration of an unmanned aerial vehicle having a distance detector system according to one embodiment.FIG. 2 shows the autonomous orsemi-autonomous vehicle 202 having thedistance detector system 101 mounted on the bottom of amain body 204 of thevehicle 202. In other embodiments, one or more of thedistance detector system 101 can be mounted on other parts of themain body 204, such as the top or the side(s), or thedistance detector system 101 can be otherwise attached to, coupled to, or incorporated into thevehicle 202 as part of thevehicle 202. In some embodiments, for example, thedistance detector system 101 can be at least partly internal to themain body 204 and/or configured to be at least partially stowed in and unstowed from themain body 204. - In some embodiments, the autonomous or
semi-autonomous vehicle 202 can be an unmanned aerial vehicle (UAV), and in other embodiments thevehicle 202 can be a terrestrial or part-terrestrial, part-aerial, and/or amphibious vehicle operating autonomously or semi-autonomously. Although the illustratedvehicle 202 inFIGS. 2-3 is a quadcopter having four arms with their respective rotors, the disclosure herein can be implemented in other types of autonomous or semi-autonomous aerial and/or terrestrial vehicles such as a multirotor helicopter having a different number of arms and/or rotors, an aerial vehicle other than a multirotor helicopter such as a fixed wing aerial vehicle, or other types of autonomous or semi-autonomous vehicles such as cars or robots requiring distance detection of the surroundings. -
FIG. 3 is an illustration of an example operation of an unmanned aerial vehicle having a distance detector system according to one embodiment.FIG. 3 shows thevehicle 202 with thedistance detector system 101 attached traveling along atravel path 302 based on travel instructions.FIG. 3 also shows anobstacle 306 in the vicinity of thevehicle 202 near or on thetravel path 302, and a zone or space ofinterest 304 for navigation by thevehicle 202. In some embodiments, thevehicle 202 can determine or be provided with thetravel path 302 based on, for example, the overall mission of the travel. As the vehicle travels along thetravel path 302, thevehicle 202 may execute current travel instructions, which can be understood as instructions for thevehicle 202 to travel from its current location (e.g., a point on the travel path 302) to immediately following series of locations (e.g., traveling through a series of points on the travel path 302) for a certain duration of time and/or distance in space. For example, one or more travel instructions to follow a part of thetravel path 302 from the current position of thevehicle 202 to a point on thetravel path 302 thevehicle 202 is instructed or expected to be in the next few seconds (e.g., 3 seconds) can constitute current travel instructions. In some instances, thetravel path 302 can be predetermined, and thevehicle 202 may be provided with an a priori knowledge of the full expected travel path, and in other instances, thetravel path 302 can be determined as the travel occurs on a case-by-case or ad hoc basis. In some instances, thetravel path 302 can be tentatively pre-computed based on the real time information gathered by thevehicle 202 or the status of thevehicle 202, and thevehicle 202 can be configured to obtain, determine, and/or adjust its travel path based on the combination of known or predetermined information and on-going gathering of data. - This short forward-looking time or distance can be understood as a clearance time or distance for the
vehicle 202 to detect obstructions, if any, along the short expected path of travel. For instance, the illustrated example inFIG. 3 shows that thevehicle 202 plans to turn around soon as it travels along thetravel path 302, which may have been preplanned before the start of the mission. To determine whether thevehicle 202 can travel as planned without encountering an obstacle, for example, thedistance detector system 101 can orient thedistance detector 105 by rotating thedistance detector 105 about the 102, 104 in a manner discussed above in connection withaxes FIG. 1 to observe the space ofinterest 304 around the plannedtravel path 302 before thevehicle 202 reaches the space ofinterest 304. In the illustrated example, using thedistance detector system 101, thevehicle 202 could identify that there is theobstacle 306 partially blocking the space ofinterest 304 and may determine whether it should revise or alter thetravel path 302. - In some embodiments, the
distance detector system 101 may be observing ahead of thevehicle 202 along thetravel path 302 by a few seconds or a few meters in accordance with current travel instructions. For example, before thevehicle 202 takes off, thedistance detector system 101 may have thedistance detector 105 pointed to the direction of the takeoff so that thevehicle 202 can determine that its takeoff space is clear of obstacles. As thevehicle 202 takes off thedistance detector system 101 can continue to observe its travel path ahead of thevehicle 202 throughout the travel, which can be completed with a landing. The duration and/or distance of advanced observation can be determined based, for example, on the speed of thevehicle 202, the unexpected nature of the environment, sensing of changes in the environment, detection of target object(s), emergency situation(s), or any combination thereof. For example, if thevehicle 202 is traveling along an irregular, less predictable, or unknown terrain, thevehicle 202 may travel at a slower speed and use a shorter clearance time than it would otherwise. Similarly, if the environment thevehicle 202 is traveling includes one or more animate objects (e.g., other vehicles, animals, etc.), thevehicle 202 may travel at a slower speed and use a shorter clearance time than it would otherwise as sudden change or movements by the animate objects can create unexpected obstructions to thevehicle 202. On the other hand, if thevehicle 202 is traveling in a known area or the environment with largely inanimate objects or objects with predictable or slow movements, thevehicle 202 may travel faster and the length of the clearance time or distance may not be as critical for its travel. - In some embodiments, the
vehicle 202 can also or alternatively look ahead based on detected objects or surroundings or the mission objectives. For example, in some instances, thevehicle 202 may be provided with a high-level mission without detailed specifications as to the travel path 302 (i.e., lacking a priori knowledge at least in part). In such instances, thevehicle 202 may constantly detect its surroundings a short time or distance ahead to ensure a clear path of travel as thevehicle 202 performs its mission. Even for travels involving a predetermined travel path, thevehicle 202 nonetheless may in part engage in on-the-fly determination of its own path in various situations, such as emergency, sudden encountering of unexpected obstructions, etc., until thevehicle 202 gets situated in a better known environment, for example. - In some instances, as the
vehicle 202 performs its mission, thevehicle 202 may follow along a wall at a certain distance, either for guidance (i.e., path planning) or for a specific objective (e.g., identify defects of the wall). As thevehicle 202 follows the wall in this example, thevehicle 202 may look ahead along the tangent lines of the surface of the wall to detect if there are any obstructions on or near the wall or any protrusions from the wall. Similarly, thevehicle 202 may detect other objects in the surroundings and look ahead based on detected edges, corners, surfaces or other features of the objects either to avoid, follow side-by-side, or perform other functions. Thevehicle 202 may perform tangent-line look ahead based on the physical or geometric features of the objects according to, for example, statistical estimation, modeling, extrapolation, or other similar mechanisms to determine in what direction thedistance detector system 101 should be configured to detect as thevehicle 202 travels along the surfaces. - In certain instances, the
vehicle 202 may store, compare, and/or otherwise process distance data that may or may not be on its path forward using thedistance detector system 101. For example, in some cases, thevehicle 202 may not only gather distance data in the direction of takeoff and the travel path, but also the area below thevehicle 202 as it takes off from ground if, for instance, thevehicle 202 plans to return to the same spot for landing. Thevehicle 202 may store the distance data around the planned landing or return area gathered during takeoff, and the stored distance data information can be compared with the forward looking (i.e., along the travel path) distance data as thevehicle 202 prepares to land upon completion of a mission, for example. Based on storage, processing, and comparison of the distance data, thevehicle 202 may determine the change of environment, appearance or movements of objects, and other relevant distance information along a timeline, in order to, for example, avoid obstacles, alter its course, abort or delay landing, and find a safe and flat landing spot. - In some embodiments, the
vehicle 202 may be configured to collect distance data as payload data for various purposes. For instance, thevehicle 202 may travel for the specific purpose of gathering distance information with respect to the terrains, environment, or a particular object. In some instances, thevehicle 202 can use distance data not only to clear or plan its travel path, but also to locate itself either relatively or absolutely based in part on known or previously detected distance information. For example, thevehicle 202 may travel a closed or narrow space, and thevehicle 202 may be able to locate itself relative to the walls or other boundaries of the closed or narrow space. In another example, thevehicle 202 may gather distance information in relation to objects or terrains of known absolute locations so that thevehicle 202 can determine its absolute location. In other instances, thevehicle 202 can use the distance information gathered using thedistance detector system 101 to aid other functionalities or performances, such as magnetometer or other sensor calibration, vehicle state estimation, or emergency landing, that can utilize and benefit from the distance data. - In serving the various exemplary purposes described above, one or more processors (e.g., 110 in
FIG. 4 ) may direct the directional controller 109 (FIGS. 1, 4 ) of thedistance detector system 101 according to direction priorities, such as clearing a travel path, detecting and avoiding obstructions, gathering payload data, spotting an object of interest, locating thevehicle 202 itself, performing targeted sensing objectives, aiding or configuring other vehicle functionalities, etc. for thedistance detector 105 as discussed herein. As described above, obstruction avoidance along the travel path can be based on current travel instructions, which may allow the one or more processors to determine and update the direction priorities. As the direction priorities for thedistance detector system 101 can be related to other aspects of the vehicle's travel and mission performance (e.g., the status of the vehicle 202), the instructions for thedistance detector system 101 can be highly integrated with other instructions. Also, in case thedistance detector system 101 serves multiple functionalities (e.g., obstruction avoidance and payload data gathering), thedistance detector system 101 in conjunction with the one or more processors can be configured to interleave or alternate between the multiple functionalities as thevehicle 202 travels. In some instances, same distance data can be stored, processed, and serve multiple purposes for a variety of unrelated functionalities. Based on the mission objectives, predetermined or known characteristics of the environment, previously gathered data, predictability of the course of travel, and/or encountering of unexpected or emergency situations, the direction priorities can be determined periodically or aperiodically. For instance, determination of direction priorities, and thus configuring of the directional controller 109 (FIGS. 1, 4 ) to move or rotate thedistance detector 105, can range from a predictable scenario of regular priority updates merely confirming no obstruction along a pre-planned travel path to an unpredictable scenario of interruptive priority updates (e.g., forgoing payload data gathering to plan an emergency travel path). - The disclosed technology provides advantages in utilizing the
distance detector 105 in an effective and intelligent manner in addition to the advantages in extending physical scope of detection capabilities described above in connection withFIG. 1 . For example, depending on the space of interest and direction priorities, thevehicle 202 may not acquire or process some distance data. The directional controller 109 (FIGS. 1, 4 ) allows the extended use of thedistance detector 105 to be integrated with the execution of various instructions by one or more processors (e.g., 110 inFIG. 4 ), which in turn allows high efficiency, agility, and responsiveness to high-level processes and determinations performed by thevehicle 202. As such, even avehicle 202 using a comprehensive distance detector (e.g., three-dimensional LIDAR) can benefit from the disclosure herein as it allows efficient use of the comprehensive distance detector avoiding rote data collection and handling or processing of unnecessary data. As described herein, the added intelligence from thevehicle 202 incorporated into thedistance detector system 101 may allow targeted sensing or gathering and efficient storing of the gathered data. -
FIG. 4 is a diagram showing components of an example unmanned aerial vehicle according to one embodiment. Thevehicle 202 illustrated inFIG. 2 includes one or more processor(s) 110 in communication with a state estimator which may be an inertial measurement unit (IMU) 112. Theprocessor 110 is in further communication with one ormore transceivers 108,sensors 115, acamera 111, a global positioning system (GPS)module 114,memory 124, andmotor controllers 120, which are in communication withmotors 122. Theprocessor 110 is in further communication with thedistance detector system 101, which may include the distance detector 105 (e.g., LIDAR), and thedirectional controller 109 as discussed above in connection withFIGS. 1-3 . Thevehicle 202 further includes apower supply 116 and abattery 118, which provides power to one or more modules of thevehicle 202, including theprocessor 110. Thetransceivers 108 and theGPS module 114 may be in further communication with their respective antennas (not shown). Thememory 124 may store one or more of mission instructions, travel instructions, pre-mission routines, payload data, flight data and/or telemetry, settings (e.g., calibration settings), parameters, or other similarly relevant information or data. Thevehicle 202 may also include arecovery system 106, which may be in communication with one or more modules in thevehicle 202, including theprocessor 110. In some embodiments, therecovery system 106 may include a dedicated recovery processor (not shown) in communication with a recovery state estimator (not shown), which may also be an additional IMU (not shown). Thevehicle 202 may include additional or intermediate modules, drivers, controllers, circuitries, lines of communication, and/or signals not illustrated inFIG. 4 . - The
vehicle 202 can perform its regular operation according to instructions executed by theprocessor 110 to, for example, take a course of action for a mission. Theprocessor 110 can be a microprocessor capable of communicating with various modules illustrated inFIG. 4 executing instructions pre-programmed and/or received during a mission, for example. Theprocessor 110 may receive settings, values, or parameters stored in thememory 124 and data from thesensors 115, thedistance detector system 101, thecamera 111, thetransceivers 108, theGPS module 114, theIMU 112, and themotor controllers 120 to evaluate the status of thevehicle 202 and determine a course of action. The status of thevehicle 202 can be determined based on data received through thesensors 115, thedistance detector system 101, and/or preloaded data stored in thememory 124 and accessed by theprocessor 110. For example, the altitude of thevehicle 202 above ground can be determined by theprocessor 108 based on a digital elevation model (DEM) of a world elevation map or with a LIDAR (e.g., from the distance detector system 101), a barometer, or ultrasound. In some embodiments, thevehicle 202 may include multiple processors of varying levels of computing power and reliability to execute low-level instructions or run high-level application code or a virtual machine. In such embodiments, one or more of the functionalities of the processor(s) 110 described herein may instead be performed by another processor in thevehicle 202. - The
transceivers 108 can be devices capable of transmitting and receiving data to and from a system, device, or module external to thevehicle 202. For example, thetransceivers 108 may include radio frequency (RF) transceivers capable of communicating data over a Wi-Fi network or any other suitable network in various frequency bands or channels, such as 900 MHz, 2.4 GHz, 5 GHz, etc. In some embodiments, thetransceivers 108 may be implemented with a combination of separate transmitters and receivers. Themotor controllers 120 may include a controller device or circuit configured to interface between theprocessor 110 and themotors 122 for regulating and controlling speed, velocity, torque, or other operational parameters of their respective, coupledmotors 122. In some embodiments, one or more motor control schemes, such as a feedback control loop, may be implemented with theprocessor 110 and/or themotor controllers 120. Themotors 122 may include electrical or any other suitable motors coupled to their respective rotors of thevehicle 202 to control their propellers, for example. - The
memory 124 can be a memory storage device (e.g., random-access memory, read-only memory, flash memory, or solid state drive (SSD) storage) to store data collected from thesensors 115, data processed in theprocessor 110, or preloaded data, parameters, or instructions. In some embodiments, thememory 124 may store data gathered from thedistance detector 105 using various computationally efficient data structures. For example, in some cases, the distance data from thedistance detector 105 can be stored using a three-dimensional occupancy grid mapping, with the gathered data grouped into cube-shaped bins of variable resolution in space. Depending on the need of distance data for the various processes or operations described herein using distance data, the resolution of the occupancy grid can be determined to indicate whether each variable resolution bin within the reach of the distance detector is free or occupied based on the gathered distance data. In some embodiments, the three-dimensional occupancy mapping values can be estimated using probabilistic approaches based on the gathered distance data. - The
IMU 112 may include a stand-alone IMU chip containing one or more magnetometers, gyroscopes, accelerometers, and/or barometers. In some embodiments, theIMU 112 may be implemented using a combination of multiple chips or modules configured to perform, for example, measuring of magnetic fields and vehicle orientation and acceleration and to generate related data for further processing with theprocessor 110. Regardless of integrated or multi-module implementation of theIMU 112, the term “magnetometer” as used herein, generally refers to the part(s) of theIMU 112 responsible for measuring the magnetic field at the location of thevehicle 202. Similarly, the term “accelerometer” as used herein, generally refers to the part(s) of theIMU 112 responsible for measuring acceleration of thevehicle 202, and the term “gyroscope” as used herein, generally refers to the part(s) of theIMU 112 responsible for measuring orientation of thevehicle 202. - The
recovery system 106 can be responsible for recovery operation of thevehicle 202 to, for example, safely deploy a parachute and land thevehicle 202. Therecovery system 106 may include a parachute (not shown) and an electromechanical deployment mechanism (not shown). Thepower supply 116 may include circuitry such as voltage regulators with outputs directly powering various modules of thevehicle 202 with Vcc_vehicle, and thebattery 118 can provide power to thepower supply 116. In some embodiments, thebattery 118 can be a multi-cell lithium battery or any other suitable battery capable of powering thevehicle 202. In some embodiments, thebattery 118 of thevehicle 202 can be removable for easy swapping and charging. - The
sensors 115 may include one or more proximity sensors using, for example, infrared, radar, sonar, ultrasound, LIDAR, barometer, and/or optical technology. Thesensors 115 may also include other types of sensors gathering data regarding visual fields, auditory signals, and/or environmental conditions (e.g., temperature, humidity, pressure, etc.). TheGPS module 114 may include a GPS transceiver and/or a GPS driver configured to receive raw and/or processed GPS data such as ephemerides for further processing within theGPS module 114, with theprocessor 110, or both. Thevehicle 202 may also include a microphone (not shown) to gather audio data. In some embodiments, one ormore sensors 115 responsible for gathering data regarding auditory signals can take the place of the microphone. - As described above in connection with
FIG. 1 above, thedistance detector system 101 can include, among others, thedistance detector 105 and thedirectional controller 109. In some embodiments, thedistance detector 105 can include a LIDAR sensor, such as a one-dimensional LIDAR sensor. In some implementations, a one-dimensional LIDAR having, for example, the range of about 50 m with the sampling frequency up to about 30 kHz can be used, and thedirectional controller 109 providing a two-dimensional gimbal whose rotational frequency about one axis can be at about 2-3 Hz for avehicle 202 traveling at about 10 m/s or less. In such implementations, thevehicle 202 can gather distance data with an angular resolution of approximately 0.1°, for example. Thedirectional controller 109 can be a motorized directional controller, which includes one or more brushless motors with a slip ring to implement a gimbal system. In other embodiments, thedirectional controller 109 be otherwise configured to move or rotate thedistance detector 105 without embedded motors. In some embodiments, thedirectional controller 109 may further include drivers such as a controller device or circuit configured to interface between theprocessor 110 and the motors of thedirectional controller 109 for regulating and controlling speed, velocity, torque, or other operational parameters of their respective, coupled motors. - The
camera 111 can be configured to gather images and/or video. In some embodiments, one or more of thesensors 115 and the distance detector 105 (in conjunction with other modules of the distance detector system 101) responsible for gathering data regarding visual fields can take the place of thecamera 111. In some embodiments, thesensors 115, the distance detector 105 (in conjunction with other modules of the distance detector system 101), and/or thecamera 111 may be configured to gather parts of payload data, which includes data gathered by thevehicle 202 regarding its surroundings, such as images, video, and/or processed 3D mapping data, gathered for purposes of mission performance and/or delivered to the user for various purposes such as surveillance, observation, progress report, landscape analysis, etc. Thesensors 115 may also gather what may be termed telemetry data, which is data regarding the status and activities of thevehicle 202 during the flight such as velocity, position, attitude, temperature, and rotor speeds. Such data may be collected to retain records or logs of flight activity and perform diagnostics. In some embodiments, thesensors 115, the distance detector 105 (in conjunction with other modules of the distance detector system 101), and/or thecamera 111 may also be configured to gather data for purposes of aiding navigation and obstruction detection as disclosed herein. -
FIG. 5 is a flowchart for enhanced detection according to one embodiment. The illustratedprocess 500 can be performed in part by and/or in conjunction with one or more modules in the vehicle 202 (FIG. 4 ), such as the processor(s) 110 (FIG. 4 ), the distance detector system 101 (FIG. 4 ), the distance detector 105 (FIG. 4 ), the directional controller 109 (FIG. 4 ), the sensors 115 (FIG. 4 ), the camera 111 (FIG. 4 ), the memory 124 (FIG. 4 ), the GPS module 114 (FIG. 4 ), the IMU 112 (FIG. 4 ), and the motor controllers 120 (FIG. 4 ). It is to be noted that all or parts of 502, 504, 506, 508, 510, and 512 may be concurrently, continuously, periodically, intermittently, repeatedly, or iteratively performed, and the illustrated process insteps FIG. 5 is only one example embodiment of inventive features disclosed herein. As thevehicle 202 is powered up, theprocess 500 may proceed to step 502. - In
step 502, thevehicle 202 initiates its mission. As discussed above, the mission can include high-level objectives, a predetermined travel path, and/or distance-specific instructions (e.g., follow a wall at a certain offset distance). Also as discussed above, as thevehicle 202 prepares to take off and continues to travel, current travel instructions can be continually determined and executed for a short window of time or distance ahead. - In
step 504, one ormore processors 110 of thevehicle 202 determines direction priorities based at least in part of the current travel instructions. The current travel instructions may encompass a short path along the planned travel path of the vehicle, and thedistance detector system 101 of thevehicle 202 may, for example, determine whether, how, and to what extent thedistance detector 105 gathers distance data regarding the vehicle's expected path forward and/or its surroundings. As discussed above, in certain instances, the direction priorities can be further based on other mission objectives (e.g., payload data gathering, locating thevehicle 202 relative to its surroundings) or unanticipated situations thevehicle 202 may encounter. - In
step 506, thedirectional controller 109 of thevehicle 202 rotatesdistance detector 105 about first and second axes. As discussed above, thedirectional controller 109 may include motors to implement a motorized gimbal system that may extend the detection dimension of thedistance detector 105. - In
step 508, the one ormore processors 110 of thevehicle 202 acquire and/or process the distance data from thedistance detector 105 based at least in part on the direction priorities determined instep 504. The direction priorities may, for example, indicate that thevehicle 202 should confirm the travel path up to 5 seconds ahead should be clear of obstacles while thedistance detector system 101 gathers intermittent distance data to the ground as thevehicle 202 travels. As the current travel instructions and direction priorities get updated as thevehicle 202 travels, steps 504, 506, and 508 may be repeated throughout the mission. - In
step 510, thevehicle 202 may optionally detect one or more obstacles with thedistance detector 105 in a direction based in part on the direction priorities instep 504. In some embodiments, thedistance detector 105 may gather data according to direction priorities to determine whether there are obstacles, and theprocessor 110 may determine that there is an obstacle on or near the vehicle's immediate travel path based on the distance data. - In
step 512, thevehicle 202 may optionally adjust its travel path. For example, in case an obstruction is detected on or near the vehicle's planned travel path, the one ormore processors 110 may determine that the obstruction would hinder the vehicle's travel as planned and determine that its travel path be adjusted. Upon the determination of path adjustment, thedistance detector system 101 may further gather distance data of the vehicle's surroundings to find an alternative path to travel. In some embodiments, thevehicle 202 may constantly plan and/or adjust its path forward based on constant gathering of distance data of its surroundings. In other embodiments, thevehicle 202 may determine an alternative path forward, store the alternative path, and perform obstacle clearing steps similar to 504, 506, and 508 along the alternative path. Thesteps process 500 inFIG. 5 only shows one iteration of the process that can be performed by thevehicle 202, and in some embodiments, parts of theprocess 500 may repeat in an iterative loop. For example, after performingstep 512, the process may loop back to step 504, and thevehicle 202 may determine direction priorities and proceed to 506, 508, 510, and 512. In such example, the process performed by thesteps vehicle 202 may repeatedly loop back fromstep 512 to step 504 until, for instance, thevehicle 202 completes its mission. - The foregoing description and claims may refer to elements or features as being “connected” or “coupled” together. As used herein, unless expressly stated otherwise, “connected” means that one element/feature is directly or indirectly connected to another element/feature, and not necessarily mechanically. Likewise, unless expressly stated otherwise, “coupled” means that one element/feature is directly or indirectly coupled to another element/feature, and not necessarily mechanically. Thus, although the various schematics shown in the Figures depict example arrangements of elements and components, additional intervening elements, devices, features, or components may be present in an actual embodiment (assuming that the functionality of the depicted circuits is not adversely affected).
- As used herein, the term “determining” encompasses a wide variety of actions. For example, “determining” may include calculating, computing, processing, deriving, investigating, looking up (e.g., looking up in a table, a database or another data structure), ascertaining and the like. Also, “determining” may include receiving (e.g., receiving information), accessing (e.g., accessing data in a memory) and the like. Also, “determining” may include resolving, selecting, choosing, establishing and the like. Further, a “channel width” as used herein may encompass or may also be referred to as a bandwidth in certain aspects.
- The various operations of methods described above may be performed by any suitable means capable of performing the operations, such as various hardware and/or software component(s), circuits, and/or module(s). Generally, any operations illustrated in the Figures may be performed by corresponding functional means capable of performing the operations.
- The various illustrative logical blocks, modules, and circuits described in connection with the present disclosure may be implemented or performed with a general purpose processor, a digital signal processor (DSP), an application specific integrated circuit (ASIC), a field programmable gate array signal (FPGA) or other programmable logic device (PLD), discrete gate or transistor logic, discrete hardware components or any combination thereof designed to perform the functions described herein. A general-purpose processor may be a microprocessor, but in the alternative, the processor may be any commercially available processor, controller, microcontroller or state machine. A processor may also be implemented as a combination of computing devices, e.g., a combination of a DSP and a microprocessor, a plurality of microprocessors, one or more microprocessors in conjunction with a DSP core, or any other such configuration.
- The methods disclosed herein comprise one or more steps or actions for achieving the described method. The method steps and/or actions may be interchanged with one another without departing from the scope of the claims. In other words, unless a specific order of steps or actions is specified, the order and/or use of specific steps and/or actions may be modified without departing from the scope of the claims.
- It is to be understood that the implementations are not limited to the precise configuration and components illustrated above. Various modifications, changes and variations may be made in the arrangement, operation and details of the methods and apparatus described above without departing from the scope of the implementations.
- Although this invention has been described in terms of certain embodiments, other embodiments that are apparent to those of ordinary skill in the art, including embodiments that do not provide all of the features and advantages set forth herein, are also within the scope of this invention. Moreover, the various embodiments described above can be combined to provide further embodiments. In addition, certain features shown in the context of one embodiment can be incorporated into other embodiments as well.
Claims (26)
Priority Applications (3)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US14/864,508 US9594381B1 (en) | 2015-09-24 | 2015-09-24 | Enhanced distance detection system |
| CA3000092A CA3000092A1 (en) | 2015-09-24 | 2016-09-21 | Enhanced distance detection system |
| PCT/US2016/052965 WO2017053485A1 (en) | 2015-09-24 | 2016-09-21 | Enhanced distance detection system |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US14/864,508 US9594381B1 (en) | 2015-09-24 | 2015-09-24 | Enhanced distance detection system |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US9594381B1 US9594381B1 (en) | 2017-03-14 |
| US20170090481A1 true US20170090481A1 (en) | 2017-03-30 |
Family
ID=57083374
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US14/864,508 Active US9594381B1 (en) | 2015-09-24 | 2015-09-24 | Enhanced distance detection system |
Country Status (3)
| Country | Link |
|---|---|
| US (1) | US9594381B1 (en) |
| CA (1) | CA3000092A1 (en) |
| WO (1) | WO2017053485A1 (en) |
Cited By (22)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US9886632B1 (en) | 2016-11-04 | 2018-02-06 | Loveland Innovations, LLC | Systems and methods for autonomous perpendicular imaging of test squares |
| US9996746B1 (en) | 2016-11-04 | 2018-06-12 | Loveland Innovations, LLC | Systems and methods for autonomous perpendicular imaging with a target field of view |
| US10012735B1 (en) | 2017-05-04 | 2018-07-03 | Loveland Innovations, LLC | GPS offset calibrations for UAVs |
| US10102428B2 (en) | 2017-02-27 | 2018-10-16 | Loveland Innovations, LLC | Systems and methods for surface and subsurface damage assessments, patch scans, and visualization |
| US20180322348A1 (en) * | 2017-05-02 | 2018-11-08 | Qualcomm Incorporated | Dynamic sensor operation and data processing based on motion information |
| US20190094345A1 (en) * | 2017-09-25 | 2019-03-28 | Leica Geosystems Ag | Laser scanner |
| US10678246B1 (en) | 2018-07-30 | 2020-06-09 | GM Global Technology Operations LLC | Occupancy grid movie system |
| WO2020139195A1 (en) * | 2018-12-27 | 2020-07-02 | Performance Rotors Pte. Ltd. | Drone for surface defects inspection |
| US10733443B2 (en) | 2018-08-24 | 2020-08-04 | Loveland Innovations, LLC | Image analysis and estimation of rooftop solar exposure |
| US10810426B2 (en) | 2016-11-04 | 2020-10-20 | Loveland Innovations, LLC | Systems and methods for autonomous perpendicular imaging of test squares |
| US10824156B1 (en) | 2018-07-30 | 2020-11-03 | GM Global Technology Operations LLC | Occupancy grid movie system |
| US10825346B2 (en) | 2016-11-04 | 2020-11-03 | Loveland Innovations, LLC | Systems and methods for adaptive property analysis via autonomous vehicles |
| US20200361452A1 (en) * | 2019-05-13 | 2020-11-19 | Toyota Research Institute, Inc. | Vehicles and methods for performing tasks based on confidence in accuracy of module output |
| WO2021025568A2 (en) | 2019-08-07 | 2021-02-11 | Ozzo Nebras | A lidar device, system and a control method of the same |
| US10984182B2 (en) | 2017-05-12 | 2021-04-20 | Loveland Innovations, LLC | Systems and methods for context-rich annotation and report generation for UAV microscan data |
| US11097841B2 (en) | 2017-10-24 | 2021-08-24 | Loveland Innovations, LLC | Crisscross boustrophedonic flight patterns for UAV scanning and imaging |
| US11205072B2 (en) | 2018-08-24 | 2021-12-21 | Loveland Innovations, LLC | Solar ray mapping via divergent beam modeling |
| US11205841B2 (en) * | 2017-04-21 | 2021-12-21 | SZ DJI Technology Co., Ltd. | Antenna assembly for communicating with unmanned aerial vehicle (UAV) and UAV system |
| US11210514B2 (en) | 2018-08-24 | 2021-12-28 | Loveland Innovations, LLC | Image analysis and estimation of rooftop solar exposure via solar ray mapping |
| US11279035B1 (en) * | 2019-05-23 | 2022-03-22 | Amazon Technologies, Inc. | LIDAR safely rings |
| US11402234B2 (en) * | 2019-09-27 | 2022-08-02 | Kitty Hawk Corporation | Distance sensor test system |
| US11532116B2 (en) | 2020-10-30 | 2022-12-20 | Loveland Innovations, Inc. | Graphical user interface for controlling a solar ray mapping |
Families Citing this family (24)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| USD809992S1 (en) * | 2016-07-05 | 2018-02-13 | Guangzhou Ehang Intelligent Technology Co., Ltd. | Multi-rotor aircraft |
| USD798961S1 (en) * | 2016-09-20 | 2017-10-03 | Shenzhen Hubsan Technology Co., Ltd. | Quadcopter model drone |
| USD801224S1 (en) * | 2016-09-20 | 2017-10-31 | Shenzhen Hubsan Technology Co., Ltd. | Quadcopter model drone |
| CN110192122B (en) * | 2017-01-24 | 2023-11-14 | 深圳市大疆创新科技有限公司 | Systems and methods for radar control on unmanned mobile platforms |
| CN108700888A (en) * | 2017-04-18 | 2018-10-23 | 深圳市大疆创新科技有限公司 | Control method, flight controller and the unmanned vehicle of unmanned vehicle |
| WO2019005823A1 (en) | 2017-06-26 | 2019-01-03 | The Trustees Of Columbia University In The City Of New York | Millimeter scale long grating coupler |
| CN109398731B (en) * | 2017-08-18 | 2020-09-08 | 深圳市道通智能航空技术有限公司 | Method and device for improving depth information of 3D image and unmanned aerial vehicle |
| USD862285S1 (en) * | 2017-08-25 | 2019-10-08 | MerchSource, LLC | Drone |
| CN107544073A (en) * | 2017-08-29 | 2018-01-05 | 北醒(北京)光子科技有限公司 | A kind of Air Vehicle Detection method and height control method |
| USD864083S1 (en) * | 2017-10-09 | 2019-10-22 | Guangdong Shiji Technology Co., Ltd | Quadcopter |
| USD864082S1 (en) * | 2017-10-09 | 2019-10-22 | Guangdong Shiji Technology Co., Ltd | Quadcopter |
| USD858353S1 (en) * | 2017-10-30 | 2019-09-03 | Shenzhen Valuelink E-Commerce Co., Ltd. | Drone |
| US11009359B2 (en) | 2018-01-05 | 2021-05-18 | Lacuna Technologies Inc. | Transportation systems and related methods |
| CN108408066A (en) * | 2018-02-01 | 2018-08-17 | 天津翰宇昊德科技发展有限公司 | Unmanned plane with height measurement function |
| USD864022S1 (en) * | 2018-03-30 | 2019-10-22 | Shenzhen Valuelink E-Commerce Co., Ltd. | Unmanned aerial vehicle |
| USD860047S1 (en) * | 2018-04-08 | 2019-09-17 | Shenzhen Valuelink E-Commerce Co., Ltd. | Unmanned aerial vehicle |
| USD873175S1 (en) * | 2018-05-23 | 2020-01-21 | Shenzhen Hubsan Technology Co., Ltd. | Drone |
| US11879958B2 (en) * | 2018-06-06 | 2024-01-23 | Honeywell International Inc. | System and method for using an industrial manipulator for atmospheric characterization lidar optics positioning |
| CN108657431A (en) * | 2018-07-19 | 2018-10-16 | 国网上海市电力公司 | A kind of flight instruments for detecting the vertical ground clearance of high-voltage AC transmission bank |
| WO2020062080A1 (en) * | 2018-09-28 | 2020-04-02 | 深圳市大疆创新科技有限公司 | Laser ranging apparatus and mobile device |
| CN112558631B (en) * | 2020-12-04 | 2021-10-29 | 北京理工大学 | Variable parameter guidance method with large falling angle constraint based on measurement information |
| DE102021203823A1 (en) * | 2021-04-19 | 2022-10-20 | Siemens Energy Global GmbH & Co. KG | Drone with direction-dependent sensor alignment for autonomous drone missions and collision avoidance processes |
| CN114608520B (en) * | 2021-04-29 | 2023-06-02 | 北京石头创新科技有限公司 | Ranging method, ranging device, robot and storage medium |
| CN113138397B (en) * | 2021-06-01 | 2023-12-26 | 中国计量大学 | Unmanned aerial vehicle keeps away barrier device and unmanned aerial vehicle |
Family Cites Families (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6556282B2 (en) * | 2001-09-04 | 2003-04-29 | Rosemount Aerospace, Inc. | Combined LOAS and LIDAR system |
| US20120170029A1 (en) | 2009-09-22 | 2012-07-05 | ISC8 Inc. | LIDAR System Comprising Large Area Micro-Channel Plate Focal Plane Array |
| US9681065B2 (en) | 2010-06-15 | 2017-06-13 | Flir Systems, Inc. | Gimbal positioning with target velocity compensation |
| US9070289B2 (en) | 2013-05-10 | 2015-06-30 | Palo Alto Research Incorporated | System and method for detecting, tracking and estimating the speed of vehicles from a mobile platform |
| CN103543752B (en) | 2013-10-09 | 2017-03-15 | 深圳市大疆创新科技有限公司 | A kind of remote control thereof and remote control systems |
| WO2015102731A2 (en) | 2013-10-18 | 2015-07-09 | Aerovironment, Inc. | Privacy shield for unmanned aerial systems |
-
2015
- 2015-09-24 US US14/864,508 patent/US9594381B1/en active Active
-
2016
- 2016-09-21 CA CA3000092A patent/CA3000092A1/en not_active Abandoned
- 2016-09-21 WO PCT/US2016/052965 patent/WO2017053485A1/en not_active Ceased
Cited By (36)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US9886632B1 (en) | 2016-11-04 | 2018-02-06 | Loveland Innovations, LLC | Systems and methods for autonomous perpendicular imaging of test squares |
| US9996746B1 (en) | 2016-11-04 | 2018-06-12 | Loveland Innovations, LLC | Systems and methods for autonomous perpendicular imaging with a target field of view |
| US11720104B2 (en) | 2016-11-04 | 2023-08-08 | Loveland Innovations, Inc. | Systems and methods for adaptive property analysis via autonomous vehicles |
| US10089530B2 (en) | 2016-11-04 | 2018-10-02 | Loveland Innovations, LLC | Systems and methods for autonomous perpendicular imaging of test squares |
| US10825346B2 (en) | 2016-11-04 | 2020-11-03 | Loveland Innovations, LLC | Systems and methods for adaptive property analysis via autonomous vehicles |
| US10810426B2 (en) | 2016-11-04 | 2020-10-20 | Loveland Innovations, LLC | Systems and methods for autonomous perpendicular imaging of test squares |
| US10102428B2 (en) | 2017-02-27 | 2018-10-16 | Loveland Innovations, LLC | Systems and methods for surface and subsurface damage assessments, patch scans, and visualization |
| US11205841B2 (en) * | 2017-04-21 | 2021-12-21 | SZ DJI Technology Co., Ltd. | Antenna assembly for communicating with unmanned aerial vehicle (UAV) and UAV system |
| US20180322348A1 (en) * | 2017-05-02 | 2018-11-08 | Qualcomm Incorporated | Dynamic sensor operation and data processing based on motion information |
| US11074463B2 (en) * | 2017-05-02 | 2021-07-27 | Qualcomm Incorporated | Dynamic sensor operation and data processing based on motion information |
| US12374124B2 (en) | 2017-05-02 | 2025-07-29 | Qualcomm Incorporated | Dynamic sensor operation and data processing based on motion information |
| US10012735B1 (en) | 2017-05-04 | 2018-07-03 | Loveland Innovations, LLC | GPS offset calibrations for UAVs |
| US10984182B2 (en) | 2017-05-12 | 2021-04-20 | Loveland Innovations, LLC | Systems and methods for context-rich annotation and report generation for UAV microscan data |
| US20190094345A1 (en) * | 2017-09-25 | 2019-03-28 | Leica Geosystems Ag | Laser scanner |
| US12123976B2 (en) * | 2017-09-25 | 2024-10-22 | Leica Geosystems Ag | Laser scanner |
| US11097841B2 (en) | 2017-10-24 | 2021-08-24 | Loveland Innovations, LLC | Crisscross boustrophedonic flight patterns for UAV scanning and imaging |
| US11731762B2 (en) | 2017-10-24 | 2023-08-22 | Loveland Innovations, Inc. | Crisscross boustrophedonic flight patterns for UAV scanning and imaging |
| US12181887B2 (en) | 2018-07-30 | 2024-12-31 | Gm Cruise Holdings Llc | Occupancy grid movie system |
| US10824156B1 (en) | 2018-07-30 | 2020-11-03 | GM Global Technology Operations LLC | Occupancy grid movie system |
| US12099356B2 (en) | 2018-07-30 | 2024-09-24 | Gm Cruise Holdings Llc | Occupancy grid movie system |
| US11703873B2 (en) | 2018-07-30 | 2023-07-18 | GM Global Technology Operations LLC | Occupancy grid movie system |
| US10678246B1 (en) | 2018-07-30 | 2020-06-09 | GM Global Technology Operations LLC | Occupancy grid movie system |
| US10733443B2 (en) | 2018-08-24 | 2020-08-04 | Loveland Innovations, LLC | Image analysis and estimation of rooftop solar exposure |
| US11188751B2 (en) | 2018-08-24 | 2021-11-30 | Loveland Innovations, LLC | Image analysis and estimation of rooftop solar exposure |
| US11878797B2 (en) | 2018-08-24 | 2024-01-23 | Loveland Innovations, Inc. | Image analysis and estimation of rooftop solar exposure |
| US11783544B2 (en) | 2018-08-24 | 2023-10-10 | Loveland Innovations, Inc. | Solar ray mapping via divergent beam modeling |
| US11205072B2 (en) | 2018-08-24 | 2021-12-21 | Loveland Innovations, LLC | Solar ray mapping via divergent beam modeling |
| US11210514B2 (en) | 2018-08-24 | 2021-12-28 | Loveland Innovations, LLC | Image analysis and estimation of rooftop solar exposure via solar ray mapping |
| WO2020139195A1 (en) * | 2018-12-27 | 2020-07-02 | Performance Rotors Pte. Ltd. | Drone for surface defects inspection |
| US20200361452A1 (en) * | 2019-05-13 | 2020-11-19 | Toyota Research Institute, Inc. | Vehicles and methods for performing tasks based on confidence in accuracy of module output |
| US11279035B1 (en) * | 2019-05-23 | 2022-03-22 | Amazon Technologies, Inc. | LIDAR safely rings |
| EP4010738A4 (en) * | 2019-08-07 | 2022-11-09 | Ozzo, Nebras | LIDAR DEVICE, SYSTEM AND METHOD FOR CONTROL THEREOF |
| WO2021025568A2 (en) | 2019-08-07 | 2021-02-11 | Ozzo Nebras | A lidar device, system and a control method of the same |
| US11402234B2 (en) * | 2019-09-27 | 2022-08-02 | Kitty Hawk Corporation | Distance sensor test system |
| US11699261B2 (en) | 2020-10-30 | 2023-07-11 | Loveland Innovations, Inc. | Graphical user interface for controlling a solar ray mapping |
| US11532116B2 (en) | 2020-10-30 | 2022-12-20 | Loveland Innovations, Inc. | Graphical user interface for controlling a solar ray mapping |
Also Published As
| Publication number | Publication date |
|---|---|
| CA3000092A1 (en) | 2017-03-30 |
| WO2017053485A1 (en) | 2017-03-30 |
| US9594381B1 (en) | 2017-03-14 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US9594381B1 (en) | Enhanced distance detection system | |
| US10060741B2 (en) | Topology-based data gathering | |
| CN111033561B (en) | Systems and methods for utilizing semantic information to navigate robotic devices | |
| US11015956B2 (en) | System and method for automatic sensor calibration | |
| US11247522B2 (en) | Vehicle capable of multiple varieties of locomotion | |
| US10534068B2 (en) | Localization system, vehicle control system, and methods thereof | |
| CN109476377B (en) | Universal Universal UAV Controller | |
| EP2177966B1 (en) | Systems and methods for unmanned aerial vehicle navigation | |
| EP2177965B1 (en) | High integrity coordination for multiple off-road machines | |
| US20180350086A1 (en) | System And Method Of Dynamically Filtering Depth Estimates To Generate A Volumetric Map Of A Three-Dimensional Environment Having An Adjustable Maximum Depth | |
| Moeller et al. | Autonomous navigation of an agricultural robot using RTK GPS and Pixhawk | |
| JP5688700B2 (en) | MOBILE BODY CONTROL DEVICE AND MOBILE BODY HAVING MOBILE BODY CONTROL DEVICE | |
| ES2883847T3 (en) | Vehicle collision prevention | |
| US11287261B2 (en) | Method and apparatus for controlling unmanned aerial vehicle | |
| CN111801717A (en) | Automatic exploration control of robotic vehicles | |
| US10386857B2 (en) | Sensor-centric path planning and control for robotic vehicles | |
| US11019270B2 (en) | Unmanned aerial vehicle | |
| JP7106417B2 (en) | Flight plan calculation device and program | |
| WO2019040179A1 (en) | Controlling landings of an aerial robotic vehicle using three-dimensional terrain maps generated using visual-inertial odometry | |
| JP6754518B2 (en) | Position estimation system | |
| US20200326706A1 (en) | Systems and methods for localizing aerial vehicle using unmanned vehicle | |
| WO2024142247A1 (en) | Unmanned aerial vehicle and unmanned aerial vehicle control system | |
| JP2020057311A (en) | Travel plan calculation device and program | |
| US20250321589A1 (en) | Unmanned aerial vehicle, unmanned aerial vehicle control system, and unmanned aerial vehicle control method |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| AS | Assignment |
Owner name: KESPRY, INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:CLARK, ROBERT PARKER;KIM, CHANG YOUNG;REEL/FRAME:037583/0813 Effective date: 20150929 |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
| MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2551); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY Year of fee payment: 4 |
|
| AS | Assignment |
Owner name: SILICON VALLEY BANK, CALIFORNIA Free format text: SECURITY INTEREST;ASSIGNOR:KESPRY, INC.;REEL/FRAME:057355/0980 Effective date: 20210831 |
|
| AS | Assignment |
Owner name: KESPRY, LLC, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:KESPRY, INC.;REEL/FRAME:058975/0676 Effective date: 20210929 Owner name: FIRMATEK SOFTWARE, LLC, TEXAS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:KESPRY, LLC;REEL/FRAME:058975/0798 Effective date: 20211022 |
|
| MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2552); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY Year of fee payment: 8 |