US20130131908A1 - Navigation and control system for autonomous vehicles - Google Patents
Navigation and control system for autonomous vehicles Download PDFInfo
- Publication number
- US20130131908A1 US20130131908A1 US13/715,586 US201213715586A US2013131908A1 US 20130131908 A1 US20130131908 A1 US 20130131908A1 US 201213715586 A US201213715586 A US 201213715586A US 2013131908 A1 US2013131908 A1 US 2013131908A1
- Authority
- US
- United States
- Prior art keywords
- vehicle
- obstacle
- present
- processor
- path
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000004091 panning Methods 0.000 claims abstract description 10
- 230000007246 mechanism Effects 0.000 claims abstract description 8
- 230000006735 deficit Effects 0.000 claims description 7
- 238000004458 analytical method Methods 0.000 claims description 3
- 238000005065 mining Methods 0.000 claims description 2
- 238000003384 imaging method Methods 0.000 description 67
- 238000000034 method Methods 0.000 description 38
- 238000004891 communication Methods 0.000 description 28
- 238000004422 calculation algorithm Methods 0.000 description 26
- 230000006870 function Effects 0.000 description 19
- 238000001514 detection method Methods 0.000 description 18
- 230000008569 process Effects 0.000 description 12
- 238000012545 processing Methods 0.000 description 12
- 230000003287 optical effect Effects 0.000 description 11
- 230000000875 corresponding effect Effects 0.000 description 10
- 238000005259 measurement Methods 0.000 description 10
- 239000002184 metal Substances 0.000 description 10
- 229910052751 metal Inorganic materials 0.000 description 10
- 230000009466 transformation Effects 0.000 description 9
- 238000013459 approach Methods 0.000 description 8
- 230000033001 locomotion Effects 0.000 description 8
- 238000012546 transfer Methods 0.000 description 8
- 230000010355 oscillation Effects 0.000 description 7
- 230000008901 benefit Effects 0.000 description 6
- 239000002131 composite material Substances 0.000 description 5
- 239000011159 matrix material Substances 0.000 description 5
- 238000003860 storage Methods 0.000 description 5
- 238000012360 testing method Methods 0.000 description 5
- 230000005540 biological transmission Effects 0.000 description 4
- 238000005516 engineering process Methods 0.000 description 4
- 238000013507 mapping Methods 0.000 description 4
- 230000004044 response Effects 0.000 description 4
- 230000001953 sensory effect Effects 0.000 description 4
- 238000012800 visualization Methods 0.000 description 4
- 230000001276 controlling effect Effects 0.000 description 3
- 230000007423 decrease Effects 0.000 description 3
- 238000001914 filtration Methods 0.000 description 3
- 231100001261 hazardous Toxicity 0.000 description 3
- 230000002829 reductive effect Effects 0.000 description 3
- 230000035945 sensitivity Effects 0.000 description 3
- 230000003068 static effect Effects 0.000 description 3
- 241000271566 Aves Species 0.000 description 2
- 230000001133 acceleration Effects 0.000 description 2
- 230000003044 adaptive effect Effects 0.000 description 2
- 230000008859 change Effects 0.000 description 2
- 230000000295 complement effect Effects 0.000 description 2
- 239000012141 concentrate Substances 0.000 description 2
- 230000002596 correlated effect Effects 0.000 description 2
- 230000008878 coupling Effects 0.000 description 2
- 238000010168 coupling process Methods 0.000 description 2
- 238000005859 coupling reaction Methods 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 239000000446 fuel Substances 0.000 description 2
- 230000001771 impaired effect Effects 0.000 description 2
- 230000000670 limiting effect Effects 0.000 description 2
- 230000004807 localization Effects 0.000 description 2
- 239000000463 material Substances 0.000 description 2
- 125000006850 spacer group Chemical group 0.000 description 2
- 241001517013 Calidris pugnax Species 0.000 description 1
- 244000025254 Cannabis sativa Species 0.000 description 1
- RYGMFSIKBFXOCR-UHFFFAOYSA-N Copper Chemical compound [Cu] RYGMFSIKBFXOCR-UHFFFAOYSA-N 0.000 description 1
- CWYNVVGOOAEACU-UHFFFAOYSA-N Fe2+ Chemical compound [Fe+2] CWYNVVGOOAEACU-UHFFFAOYSA-N 0.000 description 1
- 241001465754 Metazoa Species 0.000 description 1
- 101100521334 Mus musculus Prom1 gene Proteins 0.000 description 1
- 230000009471 action Effects 0.000 description 1
- 230000002411 adverse Effects 0.000 description 1
- 230000003466 anti-cipated effect Effects 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 238000013473 artificial intelligence Methods 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000015556 catabolic process Effects 0.000 description 1
- 230000001010 compromised effect Effects 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 238000005094 computer simulation Methods 0.000 description 1
- 238000012790 confirmation Methods 0.000 description 1
- 238000007405 data analysis Methods 0.000 description 1
- 230000007812 deficiency Effects 0.000 description 1
- 238000006731 degradation reaction Methods 0.000 description 1
- 230000001934 delay Effects 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000006073 displacement reaction Methods 0.000 description 1
- 239000000428 dust Substances 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000008030 elimination Effects 0.000 description 1
- 238000003379 elimination reaction Methods 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 244000144992 flock Species 0.000 description 1
- 238000009472 formulation Methods 0.000 description 1
- 238000010413 gardening Methods 0.000 description 1
- 230000004313 glare Effects 0.000 description 1
- 230000005484 gravity Effects 0.000 description 1
- 238000003306 harvesting Methods 0.000 description 1
- 238000007726 management method Methods 0.000 description 1
- 230000005055 memory storage Effects 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000003032 molecular docking Methods 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 208000010125 myocardial infarction Diseases 0.000 description 1
- 239000013307 optical fiber Substances 0.000 description 1
- 230000002085 persistent effect Effects 0.000 description 1
- 230000000644 propagated effect Effects 0.000 description 1
- 230000005855 radiation Effects 0.000 description 1
- 230000035484 reaction time Effects 0.000 description 1
- 230000003252 repetitive effect Effects 0.000 description 1
- 230000004043 responsiveness Effects 0.000 description 1
- 230000000717 retained effect Effects 0.000 description 1
- 239000011435 rock Substances 0.000 description 1
- 239000004576 sand Substances 0.000 description 1
- 238000013515 script Methods 0.000 description 1
- 238000000926 separation method Methods 0.000 description 1
- 230000035939 shock Effects 0.000 description 1
- 239000000779 smoke Substances 0.000 description 1
- 238000001228 spectrum Methods 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
- 230000008685 targeting Effects 0.000 description 1
- 231100000331 toxic Toxicity 0.000 description 1
- 230000002588 toxic effect Effects 0.000 description 1
- 238000000844 transformation Methods 0.000 description 1
- 238000013519 translation Methods 0.000 description 1
- XLYOFNOQVPJJNP-UHFFFAOYSA-N water Substances O XLYOFNOQVPJJNP-UHFFFAOYSA-N 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W10/00—Conjoint control of vehicle sub-units of different type or different function
- B60W10/04—Conjoint control of vehicle sub-units of different type or different function including control of propulsion units
- B60W10/06—Conjoint control of vehicle sub-units of different type or different function including control of propulsion units including control of combustion engines
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W10/00—Conjoint control of vehicle sub-units of different type or different function
- B60W10/20—Conjoint control of vehicle sub-units of different type or different function including control of steering systems
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W30/00—Purposes of road vehicle drive control systems not related to the control of a particular sub-unit, e.g. of systems using conjoint control of vehicle sub-units
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W30/00—Purposes of road vehicle drive control systems not related to the control of a particular sub-unit, e.g. of systems using conjoint control of vehicle sub-units
- B60W30/08—Active safety systems predicting or avoiding probable or impending collision or attempting to minimise its consequences
- B60W30/095—Predicting travel path or likelihood of collision
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/88—Lidar systems specially adapted for specific applications
- G01S17/93—Lidar systems specially adapted for specific applications for anti-collision purposes
- G01S17/931—Lidar systems specially adapted for specific applications for anti-collision purposes of land vehicles
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/48—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
- G01S7/481—Constructional features, e.g. arrangements of optical elements
- G01S7/4817—Constructional features, e.g. arrangements of optical elements relating to scanning
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0231—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means
- G05D1/0238—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means using obstacle or wall sensors
- G05D1/024—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means using obstacle or wall sensors in combination with a laser
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0268—Control of position or course in two dimensions specially adapted to land vehicles using internal positioning means
- G05D1/0274—Control of position or course in two dimensions specially adapted to land vehicles using internal positioning means using mapping information stored in a memory device
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0276—Control of position or course in two dimensions specially adapted to land vehicles using signals provided by a source external to the vehicle
- G05D1/0278—Control of position or course in two dimensions specially adapted to land vehicles using signals provided by a source external to the vehicle using satellite positioning signals, e.g. GPS
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2552/00—Input parameters relating to infrastructure
- B60W2552/35—Road bumpiness, e.g. potholes
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0231—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means
- G05D1/0238—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means using obstacle or wall sensors
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0231—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means
- G05D1/0242—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means using non-visible light signals, e.g. IR or UV signals
Definitions
- the invention relates to a three-dimensional (3D) imaging sensor and method for controlling an autonomous vehicle.
- the driver remains a critical component of the vehicle's control system as the driver makes numerous decisions directed to the safe operation of the vehicle including speed, steering, obstacle and hazard recognition, and avoidance thereof. Yet, the driver's ability to perform all of these functions can become compromised due to physical factors such as driver fatigue, driver impairment, driver inattention, or other factors such as visibility that reduce the reaction time needed by the driver to successfully avoid hazards.
- the driver may become disoriented or incapable of physically commanding the vehicle as would occur if the driver suffered a medical emergency or if for example the driver became disoriented under the driving conditions.
- a disorienting or incapacitating environment would be a car or ship being driven or steered under snow, fog, rain, and/or nighttime blackout conditions where the diver (or captain of the ship) is handicapped in his or her ability to perceive and react to hazards approaching or to which the ship is approaching.
- one object of the present invention accomplished in various of the embodiments is to provide a system and method for vehicle object identification in which a location of the object relative to the vehicle is identified.
- Another object of the present invention accomplished in various of the embodiments is to provide a system and method for vehicle object identification in which a path for the vehicle relative to the location of the object is identified.
- Yet another object of the present invention accomplished in various of the embodiments is to provide a system and method for identifying objects in a field of view of the vehicle as obstacles.
- a navigation and control system includes a sensor configured to locate objects in a predetermined field of view from a vehicle.
- the sensor has an emitter configured to repeatedly scan a beam into a two-dimensional sector of a plane defined with respect to a first predetermined axis of the vehicle, and a detector configured to detect a reflection of the emitted beam from one of the objects.
- the sensor includes a panning mechanism configured to pan the plane in which the beam is scanned about a second predetermined axis to produce a three dimensional field of view.
- the navigation and control system includes a processor configured to determine the existence and location of the objects in the three dimensional field of view based on a position of the vehicle and a time between an emittance of the beam and a reception of the reflection of the emitted beam from one of the objects.
- a method for navigation and control of a vehicle includes scanning a beam into a sector of a plane defined with respect to a first predetermined axis of the vehicle, detecting a reflection of the emitted beam from an object removed from the vehicle, panning the plane in which the beam is scanned about a second predetermined axis to produce a three dimensional field of view, and determining the existence and location of the object in the three dimensional field of view based on a position of the vehicle and a time between an emittance of the beam and a reception of the reflection of the emitted beam.
- FIG. 1A is a schematic illustration of an autonomous vehicle according to one embodiment of the present invention in which a two-dimensional (2D) scan is made in a sector of a plane normal to a predetermined axis of a vehicle;
- FIG. 1B is a schematic illustration of an autonomous vehicle according to one embodiment of the present invention in which a three-dimensional (3D) scan is made by displacing the scan out the plane normal to the predetermined axis of a vehicle;
- 3D three-dimensional
- FIG. 2 is a schematic illustration of an emitter and detector system according to one embodiment of the present invention.
- FIG. 3 is a detailed schematic illustration showing a perspective frontal view of a laser radar (LADAR) imaging sensor according to one embodiment of the present invention
- FIG. 4 is a detailed schematic illustration showing a perspective rear view of the laser radar (LADAR) imaging sensor depicted in FIG. 3 ;
- LADAR laser radar
- FIG. 5 is a contour map of a 3D scan obtained by methods of the present invention.
- FIG. 6A is a flow chart illustrating particular method steps of the present invention.
- FIG. 6B is a flow chart illustrating particular method steps of the present invention for obstacle identification and avoidance
- FIG. 7A is a schematic illustration of a vehicle according to the present invention utilizing multiple LADAR imaging sensors to scan separate fields of view;
- FIG. 7B is a schematic illustration of a vehicle according to the present invention utilizing multiple LADAR imaging sensors to scan the same or overlapping fields of view;
- FIG. 8 is a schematic illustration of an exemplary computer system of the present invention.
- FIG. 9 is a high level schematic of a computing network system of the present invention.
- FIG. 10 is a high level schematic of the vehicle control system of the present invention.
- FIGS. 11-13 are schematics of path determinations made by the present invention.
- FIG. 1A depicts an imaging sensor 8 mounted, in one embodiment, on top of a vehicle 10 in which a two-dimensional (2D) scan is made in a sector of a plane 11 normal to a predetermined axis of the vehicle 10 referred to here for illustration purposes as a “vertical” scanning plane.
- the imaging sensor 8 includes in one embodiment an emitter 12 (as shown in FIG. 2 ) that transmits laser pulses (or light) 14 from the imaging sensor 8 into the environment about the vehicle 10 . As shown in FIG. 1A , the laser (or light) pulses 14 are emitted into the vertical scanning plane 11 .
- the imaging sensor 8 is panned (or oscillated) in and out of plane 11 to create a 3D scanning volume 16 , as shown in FIG. 1B .
- the imaging sensor 8 detects objects 22 (as shown in FIG. 1B ) in the environment nearby the vehicle 10 by detecting light reflected from the objects 22 .
- the imaging sensor 8 includes a detector 18 for detecting return of an echoed signal 20 .
- the imaging sensor 8 utilizes a processor 24 for controlling the timing and emission of the laser pulses 14 and for correlating emission of the laser pulses 14 with reception of the echoed signal 20 .
- the processor 24 may be on-board the vehicle or a part of the imaging sensor 8 . Details of exemplary processors and their functions are provided later.
- laser pulses 14 from emitter 12 pass through a beam expander 13 a and a collimator 13 b .
- the laser pulses 14 are reflected at a stationary mirror 15 a to a rotating mirror 26 and then forwarded through lens 27 a and a telescope 27 b to form a beam for the laser pulses 14 with a diameter of 1-10 mm, providing a corresponding resolution for the synthesized three-dimensional field of view.
- the telescope 27 b serves to collect light reflected from objects 22 .
- the detector 18 is configured to detect light only of a wavelength of the emitted light in order to discriminate the laser light reflected from the object back to the detector from background light. Accordingly, the imaging sensor 8 operates, in one embodiment of the present invention, by sending out a laser pulse 14 that is reflected by an object 22 and measured by the detector 18 provided the object is within range of the sensitivity of the detector 18 . The elapsed time between emission and reception of the laser pulse permits the processor 24 is used to calculate the distance between the object 22 and the detector 18 .
- the optics i.e., 13 a , 13 b , 15 a , 26 , 27 a , and 27 b
- the detector 18 is a field-programmable gate array for reception of the received signals at predetermined angular positions corresponding to a respective angular direction ⁇ 1 shown in FIG. 1A .
- the rotating mirror 26 is rotated across an angular displacement ranging from 30 to 90 degrees, at angular speeds ranging from 100-10000 degrees per second.
- a 90 degree scanning range can be scanned 75 times per second or an 80 degree scanning range can be scanned between 5 and 100 times per second.
- the angular resolution can be dynamically adjusted (e.g., providing on command angular resolutions of 0.01, 0.5, 0.75, or 1 degrees for different commercially available sensors 8 .
- the emitter 12 and the detector 18 can be used for the emitter 12 and the detector 18 to provide ranging measurements.
- the emitter 12 , the detector 18 , and the associated optics constitute a laser radar (LADAR) system, but other systems capable of making precise distance measurements can be used in the present invention, such as for example a light detection and ranging (LIDAR) sensor, a radar, or a camera.
- LIDAR Light Detection and Ranging; or Laser Imaging Detection and Ranging
- LADAR Laser Detection and Ranging
- LADAR Laser Detection and Ranging
- the term laser radar is also in use, but with laser radar laser light (and not radio waves) are used.
- LIDAR Low-power laser
- a metallic object produces a significant reflection.
- non-metallic objects such as rain and rocks produce weaker reflections, and some materials may produce no detectable reflection at all, meaning some objects or features are effectively invisible at radar frequencies.
- Lasers provide one solution to these problems. The beam densities and coherency are excellent.
- the wavelengths are much smaller than can be achieved with radio systems, and range from about 10 micrometers to the UV (e.g., 250 nm). At these wavelengths, a LIDAR system can offer much higher resolution than radar.
- the imaging sensor 8 is panned (or oscillated) in and out the plane 11 to create a 3D scanning volume 16 , as shown in FIG. 1B .
- FIG. 1B defines the scanning volume 16 by the angle ⁇ (in the vertical scanning direction) and the angle ⁇ (in the horizontal scanning direction).
- the angle ⁇ ranges from 30 to 70 degrees, at angular speeds ranging from 100-1000 degrees per second.
- the angle ⁇ i.e., the panning angle
- the imaging sensor 8 typically can completely scan the 3D scanning volume 16 at more than two times a second.
- the direction that the imaging sensor 8 is pointed at the time of receiving light reflected from the objects 22 is needed (i.e., the angle of deflection from plane 28 is needed).
- geospatial positional data of the instantaneous vehicle position is utilized by processor 24 to calculate based on the distance of the object from the vehicle and its direction from the vehicle, the geospatial location of the objects in the field of view.
- the processor 24 includes a personal computer running on a Linux operating system, and the algorithms are programmed in Java programming language. Other computing systems and programming languages can be used in the present invention (as discussed in more detail below). As shown in FIG.
- processor 24 is in communication with a real time positioning device 25 , such as for example a global positioning system (GPS) and/or an inertial navigation system (INS), that transmits the location, heading, altitude, and speed of the vehicle multiple times per second to processor 24 .
- the real time positioning device 25 is typically mounted to the vehicle 10 and transmits data (such as location, heading, altitude, and speed of the vehicle) to all imaging sensors 8 (and all processors 24 ) on the vehicle 10 .
- processor 24 can determine a position of an object in the field of view to an accuracy of better than 10 cm.
- the processor 24 correlates GPS position, LADAR measurements, and angle of deflection data to produce a map of obstacles in a path of the vehicle.
- the accuracy of the map depends on the accuracy of the data from the positioning device 25 .
- a Kalman filter sorts through all data inputs to processor 24 .
- a Kalman filter is a known method of estimating the state of a system based upon recursive measurement of noisy data. In this instance, the Kalman filter is able to much more accurately estimate vehicle position by taking into account the type of noise inherent in each type of sensor and then constructing an optimal estimate of the actual position. Such filtering is described by A. Kelly, in “A 3d State Space Formulation of a Navigation Kalman Filter for Autonomous Vehicles,” CMU Robotics Institute, Tech. Rep., 1994, the entire contents of which are incorporated herein by reference.
- the Kalman filter is a set of mathematical equations that provides an efficient computational (recursive) means to estimate the state of a process, in a way that minimizes the mean of the squared error.
- the filter is very powerful in several aspects: it supports estimations of past, present, and even future states, and it can do so even when the precise nature of the modeled system is unknown.
- the positioning device 25 by including GPS and INS data, provide complementary data to the processor 24 .
- GPS and INS have reciprocal errors. That is GPS is noisy with finite drift, while INS is not noisy but has infinite drift.
- the processor 24 can be configured to accept additional inputs (discussed below) to reduce drift in its estimate of vehicle position when, for example the GPS data may not be available.
- FIG. 3 is a detailed schematic illustration of imaging sensor 8 of the present invention.
- FIG. 3 presents a frontal view of imaging sensor 8 .
- FIG. 4 presents a rear perspective view of imaging sensor 8 .
- FIG. 3 shows a motor 30 configured to oscillate the imaging sensor 8 in and out of a plane normal to a predetermined axis of the imaging sensor.
- a 12-volt DC motor operating at a speed of 120 RPM is used to oscillate the imaging sensor 8 in and out the plane.
- Other motors with reciprocating speeds different than 120 RPM can be used.
- an absolute rotary encoder 32 is placed on a shaft 34 that is oscillating.
- the encoder 32 provides an accurate reading of the angle at which the shaft 34 is instantaneously located.
- the encoder 32 an accurate measurement of the direction that the imaging sensor 8 is pointed, at the time of the scan, is known.
- the encoder 32 is an ethernet optical encoder (commercially available from Fraba Posital), placed on shaft 34 to provide both the angular position and angular velocity of the shaft.
- a separate 100 MBit ethernet connection with its own dedicated ethernet card connected the processor 24 with the encoder. This created communications delays between the encoder and the I/O computer that were consistent at approximately 0.5 ms. Testing revealed that an actual LADAR scan was taken approximately 12.5 ms before the data was available at the I/O computer. When this time was added to the 0.5 ms of delay from the encoder communications, a 13 ms delay from the actual scan to the actual reading of the encoder position and velocity was present.
- the velocity of the encoder is multiplied times the communications delay of 0.013 seconds to calculate the angular offset due to the delay.
- This angular offset (which was either negative or positive depending on the direction of oscillation) was then added to the encoder's position, giving the actual angle at the time when the scan occurred. This processing permits the orientation of the LADAR platform to be accurate within 0.05 degrees.
- the metal shaft 34 is attached to a detector bracket 36 which is supported by a metal casing 38 with bearing 40 .
- Bearing 40 is attached to metal casing 38 with a fastening mechanism such as bolts 42 and 44 .
- Detector bracket 36 is attached to metal shaft 34 .
- metal shaft 46 is attached to bearing 48 .
- Bearing 48 is attached to metal casing 38 with a fastening mechanism such as bolts 50 and 52 .
- Push rod 54 is attached to detector bracket 36 with ball joint 56 on slot 58 .
- Push rod 54 is attached to pivot spacer 60 with ball joint 62 .
- Pivot spacer 60 is attached to servo arm 64 on slot 66 .
- Servo arm 64 is attached to metal shaft 68 .
- Motor 30 is attached to servo arm 64 and is suspended from metal casing 38 by motor mounts 70 , 72 , and 74 .
- the imaging sensor 8 operates, in one embodiment, by oscillating a measurement sensor laterally about an axis of the vehicle 10 , as shown in FIG. 1A .
- the shaft 68 of motor 30 rotates at a constant speed, causing servo arm 64 to also spin at a constant speed.
- One end of push rod 54 moves with servo arm 64 , causing detector bracket 36 to oscillate back and forth.
- the degree of rotation can be adjusted by moving the mount point of ball joint 56 along slot 58 , and/or the mount point of ball joint 62 along slot 66 . Moving the mount point closer to shaft 46 increases the angle of rotation, while moving the mount point away from shaft 46 decreases the angle of rotation.
- a real time positioning device 76 such as a global positioning system (GPS) or an inertial navigation system (INS), transmits the location, heading, altitude, and speed of the vehicle multiple times per second to processor 24 .
- Software running on the processor 24 integrates the data, and, in one embodiment, uses matrix transformations to transform the YZ measurements from each 2D scan (as shown in FIG. 1 ) into a 3D view of the surrounding environment. Due to the use of the real time positioning device, in the present invention, a terrain map can be calculated even while the vehicle is moving at speeds in excess of 45 miles per hour.
- the imaging sensor 8 is configured, in one embodiment of the present invention, to receive data in the processor regarding echoed or returned signals, GPS position, LADAR measurements, and angle of deflection data to determine what sectors of the 3D scan contain obstacles, and therefore to determine which sectors are most likely impassable by the vehicle.
- data coming from detector 18 is formatted in a coordinate system relative to the detector 18 , as detailed in the exemplary illustrations below.
- the imaging sensor 8 Since the imaging sensor 8 is in a “vertical” configuration, the coordinate system in that configuration has Y-axis representing a distance from the sensor roughly parallel with the forward direction of the vehicle, with the X-axis lined up along the vertical direction.
- a search for obstacles is manifest in the slope and extent in the X-axis. To facilitate this search, all measurement points in the detector's scan are sorted by the Y-values (i.e., by a distance from the detector 18 ).
- FIG. 5 is a schematic representation of a contour map showing the reflected signatures from an object in a synthesized field of view.
- FIG. 5 depicts two common obstacles that the vehicle 10 needs to avoid.
- Obstacle 22 a is a depression in the highway (e.g., a pothole).
- Obstacle 22 b is an obstruction in the highway (e.g., a fallen boulder).
- lines of constant elevation have been added to the drawing in FIG. 5 for clarity.
- the imaging sensor 8 of the present invention as shown in FIG. 1A , produces a vertical scan, which in FIG. 5 is shown depicted on an X-Y axis.
- a vertical scan provides the advantage of directly supplying slope data (i.e., ⁇ X/ ⁇ Y data) along the road and obstacles shown in FIG. 5 without having to complete a full 3D scan, as shown in FIG. 1B , and then process the image map for slope data. Furthermore, in data from a vertical scan plane as in FIG. 1A , obstacles appear as either clusters of similar distance or gaps in distance, making obstacle identification more discernable in the data.
- any point P is classified as part of an obstacle if and only if there could be found a point Q such that Equations (1), (2), and (3) are satisfied (where h max is the maximum height or depth, m max is the maximum slope, is the unit vector in the X direction, and j is the unit vector in the Y direction).
- Equation (1) represents a condition in which there exists a first point Q in the field of view that is farther away from the vehicle than a second point P.
- Equation (2) represents a condition in which a slope between the first point and the second point is greater than a predetermined slope.
- Equation (3) represents a condition in which a height between the first point and the second point is greater than a predetermined height.
- T 1 ⁇ 2 is the transformation matrix for going from coordinate frame 1 to coordinate frame 2
- ⁇ 1 is the vector representing the position of the coordinate frame 1 with respect to coordinate frame 2
- P 1 and P 2 are the same point in coordinate frames 1 and 2 , respectively.
- the first step converts from the detector coordinate frame to the vehicle's coordinate frame.
- the transformation matrix, T s ⁇ v is defined as
- T s -> v [ cos ⁇ ⁇ ⁇ s - sin ⁇ ⁇ ⁇ s 0 sin ⁇ ⁇ ⁇ s cos ⁇ ⁇ ⁇ s 0 0 0 1 ] ⁇ [ cos ⁇ ⁇ ⁇ s 0 sin ⁇ ⁇ ⁇ s 0 1 0 - sin ⁇ ⁇ ⁇ s 0 cos ⁇ ⁇ ⁇ s ] [ 1 0 0 0 cos ⁇ ⁇ ⁇ s - sin ⁇ ⁇ ⁇ s 0 sin ⁇ ⁇ ⁇ s cos ⁇ ⁇ ⁇ s ] ( 5 )
- ⁇ s , ⁇ s and ⁇ s are the yaw (around the z-axis), pitch (around the y-axis), and roll (around the x-axis) of the detector coordinate frame relative to the vehicle's coordinate frame.
- This transformation takes into account deviations in yaw, pitch, or roll caused by the mounting of the sensor. For example, if the detector were mounted pointed slightly downward, it would have a negative pitch that would need to be countered by setting ⁇ s to its inverse (or positive) value.
- the angle of deflection caused by the oscillation is processed here by adding it to ⁇ s .
- the same basic transformation and translation is performed again in order to translate the obstacle's location from the vehicle's coordinate system to a global system.
- Yet another transformation matrix, T v ⁇ g is constructed for this purpose.
- T v -> g [ cos ⁇ ⁇ ⁇ v - sin ⁇ ⁇ ⁇ v 0 sin ⁇ ⁇ ⁇ v cos ⁇ ⁇ ⁇ v 0 0 0 1 ] ⁇ [ cos ⁇ ⁇ ⁇ v 0 sin ⁇ ⁇ ⁇ v 0 1 0 - sin ⁇ ⁇ ⁇ v 0 cos ⁇ ⁇ ⁇ v ] ⁇ [ 1 0 0 0 cos ⁇ ⁇ ⁇ v - sin ⁇ ⁇ ⁇ v 0 sin ⁇ ⁇ ⁇ v cos ⁇ ⁇ ⁇ v ] ( 6 )
- ⁇ v , ⁇ V and ⁇ v are the heading (around the z-axis), pitch (around the y-axis), and roll (around the x-axis) of the vehicle relative to the global coordinate frame.
- These heading, pitch, and roll values are generated by the GPS/INS navigation sensor that is mounted on the vehicle.
- ⁇ s is the vector representing the position of the sensor with respect to the center of the vehicle and ⁇ v is the vector representing the position of the center of the vehicle with respect to the center of the GPS/INS navigation sensor.
- FIG. 5 shows for the purposes of illustration a point R beyond the obstacles 22 a and 22 b .
- data regarding the positioning of objects (or other obstacles) beyond this distance can be discarded for example until the vehicle has safely navigated the obstacles 22 a and 22 b .
- the distance to the point R will also depend on the speed of the vehicle and the accuracy of the detector 18 to see objects beyond a sensitivity of the detector 18 . Discarding object data serves two purposes.
- the processor 24 in one embodiment of the present invention discards objects in the field of view of the detector 18 that are beyond a travel-point distance.
- the travel-point distance (also exemplified by point R on FIG. 5 ) is in one embodiment the calculated distance that is an arbitrary factor (e.g., factors of 5, 10, 100) multiplied times the distance that the vehicle will travel at the current speed before the next complete 3D scan of angular sector 16 is made.
- the travel-point distance is in one embodiment set to an arbitrary distance (e.g., 100 m).
- the travel-point distance in different embodiments of the present invention can be calculated based on linear, sigmoid, or exponential functions that are evaluated using values such as vehicle speed and time and path.
- every vertical scan produces from the slope and height data a set of obstacles that meet the vector criteria defined in eqns. (1)-(3).
- the location of the obstacle can be represented by a polygon represented in this illustrative example by a series of spatial coordinate points including a center point and four corner points: 0.2 m from the back of center, 0.2 m from front of center, 0.2 m from left of center, and 0.2 m from right of center.
- spatial indices such as a quadtree stores the polygon and efficiently subdivides space, so that object retrieval (such as identifying if an obstacle exists within a space) is faster than having to search a plot map containing obstacles on a point-by-point analysis.
- object retrieval such as identifying if an obstacle exists within a space
- Fitch et al. U.S. Pat. No. 6,212,392, the entire contents of which are incorporated herein by reference, describe the use of quadrees to determine whether or not a location associated with a wireless station is within an area of interest.
- a navigation and control system includes a sensor (e.g., imaging sensor 8 ) configured to locate objects in a field of view from a vehicle (e.g., vehicle 10 ) and a processor (e.g., processor 24 ) configured to determine the existence and location of the objects in the field of view based on a position of the vehicle and a time between an emittance of a beam from the sensor and a reception of a reflection of the emitted beam from one of the objects.
- the processor includes a memory including storage elements corresponding to subsectors in the field of view.
- the processor is configured 1) to perform an obstacle identification algorithm to determine if any object is an obstacle, 2) to determine in which of the field of view subsectors a determined obstacle exists, and 3) to store in each corresponding memory element an obstacle indicator indicative that an obstacle exists in the corresponding subsector.
- processor 24 is configured to determine if a projected path of the vehicle intersects any subsector storing an indicator.
- processor 24 is configured to execute an obstacle identification algorithm including: determining first and second points on a contour map of an object, determining if the first point is farther away from the vehicle than the second point, determining a slope between the first point and the second point, determining if the slope is greater than a predetermined slope, and determining if a height between the first point and the second point is greater than a predetermined height, and if so, determining the object is an obstacle.
- a path planning algorithm presents a “query” polygon for a sector in front of the vehicle and determines from the spatial index (e.g., a quadtree) if any of the stored obstacle polygons intersect the “query” polygon. If so, a path around the query polygon is chosen. If not, the path is clear of obstacles.
- the way that a quadtree (or most spatial indices) eliminates sectors of space is by keeping track of where it has inserted obstacles. For the quadtree techniques used in one embodiment of the present invention, the quadtree subdivides space as obstacles are inserted, thereby creating an indicator that this sector of space contains an obstacle. This leaves many sectors that are empty of obstacles, and the quadtree knows that these sectors are empty. Therefore, if the query is within one of these sectors, it cannot have an obstacle collision.
- One conventional way to determine the number of obstacles in the path of the vehicle is to mount stationary LADAR units pointed at the ground. With these LADAR units being pointed at the ground, the angle at which the LADAR beam hits the ground is calculated. Based on this angle, simple trigonometry can determine where the object is located relative to the car. In most conventional cases, multiple stationary LADAR units are used and each is placed at a different known angle.
- the conventional approach uses more computing resources than the present invention. For example, in the conventional approach, since the LADAR units are scanning horizontally, scan data must be placed into a large terrain elevation map. To detect obstacles, this terrain elevation map must be searched. This is a computationally expensive process.
- the imaging sensor 8 constantly oscillates to produce the three-dimensional scan. Accordingly, there is no need for vehicle motion to map data. Nor is there any need in the present invention to store massive amounts of map or sensor data in order to detect obstacles. Furthermore, the present invention in one embodiment only stores data of the geospatial locations of all the currently-identified obstacles, such as for example an obstacle that is within a certain distance of the vehicle and that has been seen within a certain amount of time. For example, in the present invention, an obstacle is considered pertinent if it was within 50 meters of the vehicle and if it had been seen within the last 5 seconds. Other definitions of pertinent obstacles based on their distance and historical observance in the field of view are possible.
- the imaging sensor 8 in one embodiment of the present invention rescans the same area multiple times per second, the chances of missing an obstacle are reduced as compared to a single scan over one area. As long as the area of oscillation covers a given area, any obstacle within that area should be seen.
- processor 24 is configured to perform the illustrative steps shown in FIG. 6A .
- a sensing beam is repeatedly scanned into a two-dimensional sector of a plane defined with respect to a first predetermined axis of the vehicle.
- a reflection of the emitted beam from objects removed from the vehicle is detected.
- the plane in which the beam is scanned is panned about a second predetermined axis to produce a three dimensional field of view.
- the existence and location of the objects in the three dimensional field of view based on a position of the vehicle and a time between an emittance of the beam and a reception of the reflection of the emitted beam is determined.
- the method of the present invention in one embodiment emits the beam into substantially a vertical sector. Accordingly, slope data related to the terrain and depressions or obstacle protruding from the terrain are accessible from the scan in the vertical sector without having to process the full three dimensional field of view.
- the processor operating with a vertically oriented sensor 8 does not need to process the full three dimensional field of view.
- the vertical scan lines (e.g., scans in ⁇ ) originate from the vehicle and have slope data encoded in data due to the vertical orientation. Each independent scan line provides data of relevant obstacles.
- the method of the present invention in one embodiment obtains a heading of the vehicle and a geospatial position of the vehicle, and processes data including the distance to the object, the direction to the object, the heading of the vehicle, and the geospatial position of the vehicle to determine whether the path for the vehicle intersects the object for example based on a predicted position of the vehicle.
- path planning and speed algorithms are used to direct the vehicle around the obstacles.
- the method of the present invention in one embodiment forms a three-dimensional field of view for the vehicle by scanning the beam in an angular sector of the plane while displacing the beam out of the plane, as shown in FIG. 1B .
- the method of the present invention in one embodiment determines an angular coordinate of the beam relative to the second predetermined axis for the direction to the object and offsets the angular coordinate by an amount indicative of a time delay in receiving a signal indicative of the angular coordinate at a processor (used to calculate the location of the object). This permits a more accurate determination of the location of the object.
- the method of the present invention in one embodiment directs the beam into respective angular sectors.
- the processor can determine the predetermined angular sectors based on a number of factors including but not limited to vehicle speed, an identified obstacle location, and a projected path of the vehicle.
- the sectors and/or the panning rate of each sector can be adjusted based upon some of the following non-limiting factors:
- the sectors can be adjusted to give that area of interest more coverage especially if the area of interest includes a projected path of an obstacle (i.e. projected path of a moving obstacle) or if the area of interest includes an obstacle with complexity (i.e., if a composite obstacle seems to have many facets and/or other features warranting more scans), or if the area of interest includes an obstacle density (i.e., if one area has more obstacles than another, the sectors and/or speed to cover that area are appropriately adjusted);
- the sectors could be adjusted to provide higher resolution coverage of the target area
- the sectors could be adjusted based upon whether the other vehicle is broadcasting a friend signal or a foe signal;
- the method of the present invention determines the angular sectors (and/or the panning rate to scan with) based on at least one of a vehicle speed, an identified obstacle location, a projected path of the vehicle, a resolution required to resolve a complex obstacle or a collection of obstacles to be resolved, other sensory input, an identified priority sector in which an obstacle has been identified (e.g., a region in which high threat obstacles exist), and auxiliary information indicating the presence of an obstacle, a moving obstacle, another vehicle, a landmark, or an area (or region) of interest.
- the imaging sensor 8 is configured in a vertical orientation, as shown in FIG. 1A , it may not even be necessary to pan the beam as shown in FIG. 1B , in which case the panning rate could be set to zero with the obstacles being identified on the basis of a single scan in which the angle ⁇ is varied, as discussed above.
- the method of the present invention in one embodiment stores in memory indicators of whether an obstacle exists in a subsector of the three dimensional field of view.
- the method of the present invention in one embodiment further determines if a projected path of the vehicle intersects one of the subsectors containing the indicators.
- the location of the object is determined in geospatial coordinates.
- the method determines angular coordinates (for the direction to the object) of the beam relative to the predetermined axis, obtains positioning data from a global positioning system device disposed on the vehicle, and/or obtains orientation data for the predetermined axis of the vehicle in a geospatial coordinate system.
- the method processes at least one of the angular coordinates, the positioning data, and the orientation data to determine the location of the object in geospatial coordinates.
- the data processed is limited to data for objects within a predetermined distance of the vehicle.
- the altitude of the vehicle is ascertained, and from that information the present invention determines an elevation of the object by correlating the location of the object relative to the vehicle with the geospatial position of the vehicle and the altitude of the vehicle, determines elevations for plural points on the object, and/or generates a contour map for the object.
- the present invention determines if the object is an obstacle. For instance, as detailed by the vector equations (1)-(3) listed above, one embodiment of the present invention determines if the object is an obstacle by assessing if there exists a first point on a contour map (or in the field of view) that is farther away from the vehicle than a second point P on the contour map, if a slope between the first point and the second point is greater than a predetermined slope, and if a height between the first point and the second point is greater than a predetermined height.
- the present invention determines if the obstacle is to be avoided, for example by determining if a geospatial position of the obstacle remains in the path of the vehicle over a time lapse period and/or by determining if a geospatial position of the obstacle is within a set distance from the vehicle.
- the obstacle (now identified) is mapped relative to the path of the vehicle.
- steering and speed control directions are provided to the vehicle in order to avoid the obstacle.
- the steering and speed control directions can take into consideration the destination of the vehicle, and thereby provide steering and speed control directions to the vehicle in order to avoid the obstacle and to return on course to the destination.
- the steering and speed control directions provide steering and speed control directions to the vehicle to guide the vehicle to the destination.
- FIG. 7A is a schematic illustration of a vehicle according to the present invention depicting one embodiment in which multiple LADAR imaging sensors are used.
- one or more of the imaging sensors is dedicated to scanning for the detection of objects nearby the vehicle (e.g., within 50 m) while another of the imaging sensors is dedicated to scanning for the detection of objects farther away from the vehicle (e.g., beyond 50 m).
- FIG. 7B is a schematic illustration of a vehicle according to the present invention depicting one embodiment in which multiple LADAR imaging sensors are used to scan the same or overlapping fields of view. This configuration provides redundant coverage in the center of the path so that, if one imaging sensor 8 fails, the other one can still sense obstacles most likely to be directly in the vehicle's path.
- the data from the imaging sensors 8 are correlated by placing all data onto the same elevation grid (as illustrated for example in FIG. 5 ).
- the imaging sensors 8 are configured to locate objects removed from a vehicle 10 ; and processor 24 is configured to direct one of the sensors to scan a first sector associated with a path of the vehicle, while directing another of the sensors to scan a second sector identified with an obstacle.
- the first and/or second sector determinations can be based on a number of factors including, but not limited to a vehicle speed, an identified obstacle location, a projected path of the vehicle, a resolution required to resolve a complex obstacle or a collection of obstacles to be resolved, sensory input other than from the sensors, an identified priority sector in which an obstacle has been identified, and auxiliary information indicating the presence of an obstacle, a moving obstacle, another vehicle, a landmark, or an area of interest.
- the processor 24 can direct one sensor to scan a first sector associated with a path of the vehicle, and in a programmed manner direct the same sensor (e.g., in a dynamic fashion) to scan a second sector identified with an obstacle.
- Factors which determine the programmed duty cycle by which one sensor scans the first sector and then a second sector include for example the speed of the vehicle, the proximity of the obstacle, any movement of the obstacle, an identified status of the obstacle (e.g., friend or foe), the proximity of the obstacle to the projected path of the vehicle, and the calculated clearance from the vehicle to the obstacle.
- one of the imaging sensors 8 is dedicated to scanning in a horizontal direction while another imaging sensor is directed to scan in the vertical direction. Scan information from this unit permits the processor 24 to better identify the general terrain and terrain curvature from which obstacles such as shown in FIG. 5 can be identified. Complementary data from both horizontal and vertical scans helps identity the edges of composite obstacles (groups of individual obstacles that should be treated as one obstacle) more accurately.
- One of the issues with handling moving obstacles is determining the full proportions of an obstacle. To calculate the full proportions of an obstacle, multiple “independent” obstacles are intelligently grouped to form one larger composite obstacle when for example the data points representing the independent obstacles are within a set distance of each other (e.g., within 100 cm).
- the grouping into composite obstacles is set by more than just a distance of separation between points normally qualifying as an obstacle point.
- Other factors that can be used in the determination include for example the number of times each point identified as an obstacle is seen, whether the obstacle point moves spatially in time, and whether (as discussed elsewhere) if there is confirmation of the obstacle by other image sensors or stereographic cameras.
- the obstacles being viewed from two separate dimensions (i.e., from top to bottom and from left to right). Since the beams tend to wrap around the curvature of an obstacle, this provides accurate estimations of the size and orientation of a composite obstacle. For instance, consider a spherical boulder. While the backside of the spherical boulder can not be seen, the sensing beam maps out a contour of the spherical boulder providing the aforementioned size and orientation, providing an estimate of the full size of the spherical boulder.
- stepper motors or in general any motor that can have its position dynamically adjusted oscillate the platform rather than a continuously revolving motor creates several distinct advantages: 1) the rate at which the platform oscillates can be changed dynamically, and 2) the size of the sector which is scanned can be dynamically changed.
- the present invention achieves several advantages:
- Path (or road) curvature can be tracked in the present invention.
- the imaging sensors 8 can change their sector to better match the path, thereby ensuring that the path always gets the best coverage. This also helps to ensure that hard 90 degree or more turns do not create blind spots on the path.
- An imaging sensor that cannot alter its sector would be fixated ahead of the vehicle, even though the path is turning towards the right or left, and therefore blindspots in the viewing field would develop that would not resolved until after the vehicle had started turning.
- the sensors 8 can preemptively adjust their orientation. For example, the imaging sensor 8 shown in FIG. 1B or the multiple imaging sensors 8 shown in FIGS.
- the angular sectors mapped by the multiple imaging sensors 8 can be dynamically adjusted using stepper motors to vary (in real time) the angular sector ⁇ scanned and the angular scan rate.
- Factors which influence selection of the angular sector to be scanned include vehicle speed, an identified obstacle location relative to a turn in a path of the vehicle, a projected path of the vehicle, a resolution required to resolve a complex obstacle or a collection of obstacles to be resolved, sensory input other than from the sensors, an identified priority sector in which an obstacle has been identified, and auxiliary information indicating the presence of an obstacle, a moving obstacle, another vehicle, a landmark, or an area of interest.
- the speed of oscillation of the platform can be adjusted based on the speed of the vehicle. If the vehicle is going slower, the speed of oscillation can be reduced, thereby causing a higher resolution 3D image of the environment. As the vehicle speeds up, the speed of oscillation can increase to give better overall coverage of obstacles further down the road. Accordingly, in one embodiment of the present invention, the angular scan rates for ⁇ and ⁇ (as shown in FIGS. 1A and 1B can be proportionate to the vehicle speed. Furthermore, in one embodiment of the present invention, as the vehicle moves faster, the scanned sectors could move farther out, or the speed of oscillation could change, or the angular sector could narrow (or all three could occur). The converse would apply when the vehicle slows down.
- Obstacles can be tracked in the present invention.
- Multiple imaging sensors can track obstacles and adjust their sectors to ensure that high threat obstacles are tracked (i.e., priority sectors or regions of interest can be established).
- Such capability permits the tracking of moving obstacles or other vehicles or for passing another vehicle at relatively high speed.
- respective ones of the imaging sensors 8 can be dedicated to imaging and obstacle identification and tracking in predetermined sectors.
- Zone coverage is another advantage that can be realized in the present invention.
- the processor 24 could specify several zones around the vehicle with different levels of priority and multiple imaging sensors could dynamically adjust their sectors to make sure that all zones were covered based upon priority. For example, if the path directly in front of the vehicle were the highest priority and the sides of the vehicle were lower priority, then the sensors could be dynamically adjusted to always ensure that the front of the vehicle was adequately covered, and if additional sensors were available, then the sides would be covered too. This approach allows for easy scalability and fault tolerance.
- the processor 24 in one embodiment of the present invention could realize that one of the sensors had failed, and then dynamically adjust for the failed sensor to ensure that the proper zones were always covered. Adding sensors increases coverage areas and redundancy or fault tolerance. Accordingly, in one embodiment of the present invention, respective ones of the imaging sensors 8 provide hardware redundancy and field of view redundancy.
- the processor 24 receives input from on-board vehicle sensors.
- One of these inputs is a wheel speed sensor which provides transistor-transistor logic (TTL) pulses based upon an encoder placed on a single wheel on the vehicle.
- TTL transistor-transistor logic
- the processor 24 uses GPS data from the positioning device 25 to initially calibrate the TTL pulses to actual vehicle movement. In the event that GPS data is not available, due to for example tunnels, canyons, or other obstructions, the processor 24 is able to minimize the positional drift by making use of the wheel speed sensor and its latest reliably known correspondence to the vehicle's movement.
- the wheel speed sensor included a digital sensor capable of detecting either ferrous metal or magnets that are in motion.
- the digital sensor was mounted in the wheel well adjacent to the stock Antilock Brake System (ABS) sensor, which allowed the wheel speed sensor to read the same magnets mounted on the wheel that the ABS sensor did. This level of accuracy permits the vehicle 10 to precisely know its location on the earth.
- ABS Stock Antilock Brake System
- Another input to processor 24 is input from a biometric sensor or input from command button in which a driver to the vehicle can relinquish or take control of the vehicle 10 .
- FIG. 8 illustrates one embodiment of a computer system 1201 in which the processor 24 (or any of the specific processors discussed below) of the present invention can be implemented.
- the computer system 1201 is programmed and/or configured to perform any or all of the functions described above. Further, respective functions can be divided among different computers on board the vehicle (as shown for example in the computing network system of FIG. 9 ). These computers may be in communication with each other via the communications network 1216 (discussed below).
- the computer system 1201 includes a bus 1202 or other communication mechanism for communicating information, and a internal processor 1203 coupled with the bus 1202 for processing the information.
- the computer system 1201 includes a memory 1204 , such as a random access memory (RAM) or other dynamic storage device (e.g., dynamic RAM (DRAM), static RAM (SRAM), and synchronous DRAM (SDRAM)), coupled to the bus 1202 for storing information and instructions to be executed by the internal processor 1203 .
- a memory 1204 such as a random access memory (RAM) or other dynamic storage device (e.g., dynamic RAM (DRAM), static RAM (SRAM), and synchronous DRAM (SDRAM)), coupled to the bus 1202 for storing information and instructions to be executed by the internal processor 1203 .
- the memory 1204 may be used for storing temporary variables or other intermediate information during the execution of instructions by the internal processor 1203 .
- the computer system 1201 preferably includes a non-volatile memory such as for example a read only memory (ROM) 1205 or other static storage device (e.g., programmable ROM (PROM), erasable PROM (EPROM), and electrically erasable PROM (EEPROM)) coupled to the bus 1202 for storing static information and instructions for the internal processor 1203 .
- ROM read only memory
- PROM programmable ROM
- EPROM erasable PROM
- EEPROM electrically erasable PROM
- the computer system 1201 may also include special purpose logic devices (e.g., application specific integrated circuits (ASICs)) or configurable logic devices (e.g., simple programmable logic devices (SPLDs), complex programmable logic devices (CPLDs), and field programmable gate arrays (FPGAs)).
- the computer system may also include one or more digital signal processors (DSPs) such as the TMS320 series of chips from Texas Instruments, the DSP56000, DSP56100, DSP56300, DSP56600, and DSP96000 series of chips from Motorola, the DSP1600 and DSP3200 series from Lucent Technologies or the ADSP2100 and ADSP21000 series from Analog Devices.
- DSPs digital signal processors
- Other processors especially designed to process analog signals that have been converted to the digital domain may also be used (as detailed in the working example below).
- the computer system 1201 performs a portion or all of the processing steps of the invention in response to the internal processor 1203 executing one or more sequences of one or more instructions contained in a memory, such as the main memory 1204 .
- a memory such as the main memory 1204 .
- Such instructions may be read into the main memory 1204 from another computer readable medium, such as a hard disk 1207 or a removable media drive 1208 .
- Such instructions may be read into the main memory 1204 from another computer readable medium, such as a USB flash drives or jump drives.
- Such drives are solid-state memory devices which have the ability to act as floppy disks or hard drives under most computer operating systems.
- One or more processors in a multi-processing arrangement may also be employed to execute the sequences of instructions contained in main memory 1204 .
- hard-wired circuitry may be used in place of or in combination with software instructions. Thus, embodiments are not limited to any specific combination of hardware circuitry and software.
- the computer system 1201 includes at least one computer readable medium or memory for holding instructions programmed according to the teachings of the invention and for containing data structures, tables, records, or other data described herein.
- Examples of computer readable media suitable for the present invention are compact discs, hard disks, floppy disks, tape, magneto-optical disks, PROMs (EPROM, EEPROM, flash EPROM), DRAM, SRAM, SDRAM, or any other magnetic medium, compact discs (e.g., CD-ROM), or any other optical medium, punch cards, paper tape, or other physical medium with patterns of holes, a carrier wave (described below), or any other medium from which a computer can read.
- the present invention includes software for controlling the computer system 1201 , for driving a device or devices for implementing the invention, and for enabling the computer system 1201 to interact with a human user (e.g., a driver).
- software may include, but is not limited to, device drivers, operating systems, development tools, and applications software.
- Such computer readable media further includes the computer program product of the present invention for performing all or a portion (if processing is distributed) of the processing performed in implementing the present invention.
- the computer code devices of the present invention may be any interpretable or executable code mechanism, including but not limited to scripts, interpretable programs, dynamic link libraries (DLLs), Java classes, and complete executable programs. Moreover, parts of the processing of the present invention may be distributed for better performance, reliability, and/or cost.
- Non-volatile media includes, for example, optical, magnetic disks, and magneto-optical disks, such as the hard disk 1207 or the removable media drive 1208 .
- Volatile media includes dynamic memory, such as the main memory 1204 .
- Transmission media includes coaxial cables, copper wire and fiber optics, including the wires that make up the bus 1202 . Transmission media also may also take the form of acoustic or light waves, such as those generated during radio wave and infrared data communications.
- Various forms of computer readable media may be involved in carrying out one or more sequences of one or more instructions to internal processor 1203 for execution.
- the instructions may initially be carried on a disk to a remote computer.
- An infrared detector coupled to the bus 1202 can receive the data carried in the infrared signal and place the data on the bus 1202 .
- the bus 1202 carries the data to the main memory 1204 , from which the internal processor 1203 retrieves and executes the instructions.
- the instructions received by the main memory 1204 may optionally be stored on storage device 1207 or 1208 either before or after execution by the internal processor 1203 .
- a computer readable medium contains program instructions for execution on a processor in a vehicle, which when executed by the processor, cause the processor to scan repeatedly a beam into a two-dimensional sector of a plane defined with respect to a first predetermined axis of the vehicle, to detect a reflection of the emitted beam from an object removed from the vehicle, to pan the plane in which the beam is scanned about a second predetermined axis to produce a three dimensional field of view, and to determine the existence and location of the object in the three dimensional field of view based on a position of the vehicle and a time between an emittance of the beam and a reception of the reflection of the emitted beam.
- a computer readable medium contains program instructions for execution on a processor in a vehicle, which when executed by the processor, cause the processor to scan objects in a field of view from the vehicle, to determine a location of the objects in the field of view, to determine first and second points on a contour map of an identified object, to determine if the first point is farther away from the vehicle than the second point, to determine a slope between the first point and the second point, to determine if the slope is greater than a predetermined slope, and to determine if a height between the first point and the second point is greater than a predetermined height, and if so, to determine that the object is an obstacle.
- a computer readable medium contains program instructions for execution on a processor in a vehicle, which when executed by the processor, cause the processor to scan objects in a field of view from the vehicle, to determine a location of the objects in the field of view, to perform an obstacle identification algorithm to determine if any object is an obstacle, to determine a field of view subsector in which a determined obstacle exists, and to store in a corresponding memory element an obstacle indicator indicative that an obstacle exists in a corresponding subsector.
- a computer readable medium contains program instructions for execution on a processor in a vehicle, which when executed by the processor, cause the processor to scan objects in a field of view from the vehicle, to determine a location of the objects in the field of view, to direct a scan in a first sector associated with a path of the vehicle, and to direct a scan in a second sector identified with an obstacle.
- a computer readable medium contains program instructions for execution on a processor in a vehicle, which when executed by the processor, cause the processor to scan objects in a field of view from the vehicle, to determine a location of the objects in the field of view, to direct a scan selectively in a sector associated with a projected turn in a path of the vehicle.
- the program instructions are configured to cause the processor to determine which of the sectors to scan based on at least one of a vehicle speed, an identified obstacle location, a projected path of the vehicle, a resolution required to resolve a complex obstacle or a collection of obstacles to be resolved, sensory input other than from the sensors, an identified priority sector in which an obstacle has been identified, and auxiliary information indicating the presence of an obstacle, a moving obstacle, another vehicle, a landmark, or an area of interest.
- the computer readable medium of the present invention can include program instructions detailing geographical information associated with a particular locale, path planning algorithms (as described below), navigational instructions, instructions particular to an installed image sensor on the vehicle, instructions for command and/or receipt of data from additional sensors such a stereoscopic cameras, or vehicle wheel speed sensors, or receipt of data from driver input control devices or other on-board devices (such as those described later), path planning algorithms, a particularized vehicle transfer function containing data regarding vehicle thrust and response to external forces for the autonomous vehicle in use, and steering control for the autonomous vehicle in use.
- path planning algorithms a particularized vehicle transfer function containing data regarding vehicle thrust and response to external forces for the autonomous vehicle in use, and steering control for the autonomous vehicle in use.
- the computer system 1201 also includes a communication interface 1213 coupled to the bus 1202 .
- the communication interface 1213 provides a two-way data communication coupling to a network link 1214 that is connected at least temporarily to, for example, a local area network (LAN) 1215 , or to another communications network 1216 such as the Internet during downloading of software to the processor 24 or an internal network between multiple computers on board the vehicle.
- LAN local area network
- the communication interface 1213 may be a network interface card to attach to any packet switched LAN.
- the communication interface 1213 may be an asymmetrical digital subscriber line (ADSL) card, an integrated services digital network (ISDN) card or a modem to provide a data communication connection to a corresponding type of communications line.
- ADSL asymmetrical digital subscriber line
- ISDN integrated services digital network
- Wireless links may also be implemented as part of the communication interface 1213 to provide data exchange with any of the on-board computers, image sensors, wheel speed sensors, biometric sensors, and/or driver command input devices.
- the communication interface 1213 sends and receives electrical, electromagnetic or optical signals that carry digital data streams representing various types of information.
- the network link 1214 typically provides data communication through one or more networks to other data devices to provide data exchange with any of the on-board computers, image sensors, wheel speed sensors, biometric sensors, and/or driver command input devices.
- the network link 1214 may provide a temporary connection to another computer through a local network 1215 (e.g., a LAN) or through equipment operated by a service provider, which provides communication services through a communications network 1216 .
- a local network 1215 e.g., a LAN
- a service provider which provides communication services through a communications network 1216 .
- the computing system 1201 can be in communication with an input device 1217 via the local network 1215 and the communications network 1216 which use, for example, electrical, electromagnetic, or optical signals that carry digital data streams, and the associated physical layer (e.g., CAT 5 cable, coaxial cable, optical fiber, etc).
- the associated physical layer e.g., CAT 5 cable, coaxial cable, optical fiber, etc.
- the signals through the various networks and the signals on the network link 1214 and through the communication interface 1213 , which carry the digital data to and from the computer system 1201 may be implemented in baseband signals, or carrier wave based signals.
- the baseband signals convey the digital data as unmodulated electrical pulses that are descriptive of a stream of digital data bits, where the term “bits” is to be construed broadly to mean symbol, where each symbol conveys at least one or more information bits.
- the digital data may also be used to modulate a carrier wave, such as with amplitude, phase and/or frequency shift keyed signals that are propagated over a conductive media, or transmitted as electromagnetic waves through a propagation medium.
- the digital data may be sent as unmodulated baseband data through a “wired” communication channel and/or sent within a predetermined frequency band, different than baseband, by modulating a carrier wave.
- the computer system 1201 can transmit and receive data, including program code, through the network(s) 1215 and 1216 , the network link 1214 , and the communication interface 1213 .
- the network link 1214 may provide a connection through a LAN 1215 to the various GPS and INS systems on board the vehicle.
- the input device 1217 in various embodiments of the present invention provides input to the processor 24 and represents schematically the image sensors, wheel speed sensors, biometric sensors, and/or driver command input devices discussed in the present invention.
- a 2005 Ford Escape HybridTM (hereinafter referred to as the working vehicle) was modified to include the imaging sensors 8 of the present invention.
- the working vehicle used a hybrid drive system in which an electric engine operates virtually all of the time and in which the gas engine starts and stops automatically to either provide extra horsepower or to recharge the electric engine's battery.
- the working vehicle's electrical system which was powered by a 330-volt battery, provides over 1300 watts of power to the equipment mounted in the working vehicle.
- the working vehicle utilized a commercially available Advanced Electronic Vehicle Interface Technology (AEVIT) “drive-by-wire” system from Electronic Mobility Controls (EMC) to physically control the car.
- AEVIT Advanced Electronic Vehicle Interface Technology
- EMC Electronic Mobility Controls
- the AEVIT system uses redundant servos and motors to turn the steering wheel, switch gears, apply throttle, and apply brake.
- This commercial system has a proven safety record in the automobile industry due to its use of redundant hardware.
- the working vehicle provided 110 amps of power at 12 volts.
- the imaging sensor 8 utilized in the working vehicle operated on 24 volts. Rather than provide this power from the working vehicle's 12 volt electrical system, an auxiliary 24 volt electrical system for the imaging sensor was provided.
- the auxiliary electrical system included two 12 volt batteries connected in series to provide 24 volts of power. Solar panels were installed on top of the vehicle to charge the batteries in the auxiliary system.
- a 2.8 Gigahertz Intel Pentium IIITM processor including 2 gigabytes of RAM was used as the processor 24 .
- the 2.8 Gigahertz Intel Pentium IIITM processor operated on 12 volt power supply, so an AC to DC power inverter was not needed.
- the 2.8 Gigahertz Intel Pentium IIITM processor hosted the principle computing functions, such as for example the sensor communication, vehicle controls, and artificial intelligence.
- the working vehicle included several 1.42 Gigahertz Apple Mac MiniTM computers to host the path-planning software.
- the 1.42 Gigahertz Apple Mac MiniTM computers (running on running Apple OS X 10.4 software) were programmed to perform all of the path calculations in a redundant cluster to ensure that the path planning software was not suspect to a single point of failure.
- the Mac MiniTM computers had low power consumption (less than 2.3 amps at peak) and vector math processing capabilities.
- the processors listed above can be shock mounted on the vehicle to prevent damage from sudden changes in the direction of the vehicle or the crossing of obstacles along the path.
- FIG. 9 shows a high level schematic of a computing network system of the present invention employing the computers and sensor data discussed above.
- the individual processors have similar capabilities and computer elements as described with respect to FIG. 8 .
- FIG. 9 shows that on-board electrical motor control (EMC) equipment is driven by instructions from the Navigational Computer.
- the Navigational Computer is connected by a switch (shown as a CISCO Switch) to a Sensor Processing Computer and a Path Planning Computer.
- These computers receive input from the RT3000TM GPS and INS positioning device (described below).
- the Path Planning Computer and the Navigational Computer utilize various path planning and speed control algorithms (described below) to provide steering instructions to the vehicle.
- the Sensor Processing Computer receives input from imaging sensor 8 (for example the LADAR devices shown) and encoders on the shaft of a LADAR device (for example the optical encoders shown).
- the LADAR devices provide distance data and the instantaneous angle ⁇ of the LADAR device in the predetermined angular scanning sector as shown in FIG. 1A .
- optical encoders are used as the rotary encoders 32 discussed above and provide the Sensor Processing Computer with the panning angle ⁇ relative to for example a vertical axis, as shown in FIG. 1B .
- Optical encoders are used in this embodiment of the present invention because of their high accuracy. Since optical encoders do not rely on physical contact with the shaft 34 , but rather visually determine movement, optical encoders are more accurate and less prone to failure.
- the encoders from Fraba Posital were specifically used in the working vehicle because of their Ethernet capability.
- the working vehicle utilized RT3000TM positioning devices from Oxford Technical Solutions to provide vehicle localization (i.e., positioning data).
- the RT3000TM uses OmnistarTM HP differential GPS signals to provide position accuracy of 10 centimeters or less.
- Integrated INS in the RT3000TM permits the RT3000TM to survive GPS outages of up to 30 seconds with virtually no performance degradation. Because the GPS and INS were integrated together, each can compensate for problems with the other. For example, if the INS started to drift laterally, the integrated GPS will automatically correct that drift.
- two Oxford RT3000TM GPS units were employed in a primary/secondary role.
- the working vehicle utilized commercially available laser modulation systems (LMS) for the imaging sensors 8 such as three SICKTM LMS 291 LADAR units and one RIEGLTM LMS-Q 120 LADAR on the front of the vehicle.
- LADAR units in the working vehicle were split into three different roles.
- the RIEGLTM LADAR was mounted vertically in the center of the vehicle and is used for long range (up to 80 meters) obstacle detection.
- the two SICKTM LMS 291 units were mounted vertically on either side of the RIEGLTM. These vertically mounted units were used for short range obstacle detection (up to 50 meters).
- the other SICKTM LMS 291 was mounted horizontally on the center of the vehicle and is used for analyzing terrain features.
- a sun/weather guard was constructed over the sensor mounts.
- This guard protects the LADAR units from being blinded by direct sun glare, a common problem with Sick LADAR units.
- the guard can constitute a simple optical shield such as for example a piece of sheet-metal that was attached to the top of the imaging sensor.
- the guard extended over the front of LADAR unit (or other imaging sensor) and had a slight bend to better protect the sensors from the sun. Since LADAR units produce their own light source, low light conditions have no effect on the sensors. In fact, testing showed that low light conditions actually produced better results than bright conditions, due to excessive reflections under bright sunlight.
- the present invention utilizes filtering algorithms to correct for anomalies before the anomalies can cause problems in the path planning software.
- the filtering algorithm in one embodiment associates isolated obstacle points with noise data and discards these data points, under the assumption that “real” obstacles have a physical size greater than the resolution of the imaging sensor.
- stereoscopic cameras capable of producing a three-dimensional point cloud were mounted on the front of the vehicle behind the windshield. These cameras operated in a secondary mode to the LADAR units. The cameras are used to help identify the road and to confirm the existence of obstacles. If a LADAR unit identifies an obstacle and the stereoscopic cameras also detect an obstacle in that area, the obstacle avoidance algorithms increase the confidence level associated with that obstacle. Confidence levels were used to help decide which obstacles must absolutely be avoided and which obstacles are possibly driven over without damage.
- FIG. 10 is a high level schematic of the vehicle control system of the present invention.
- FIG. 10 shows an iterative loop control system using inputs from the sensors and providing vehicle control.
- LADAR data stored in a memory unit for example of the Sensor Processing Computer in FIG. 9 is provided to obstacle detection unit 104 , speed planning unit 106 , and accelerator/brake control 108 .
- GPS/INS unit 110 also provided to the obstacle detection unit 104 .
- the obstacle detection unit 104 decides which objects in the field of view are obstacles (for example using the conditions specified in eqns. (1), (2), and (3) above, obstacle identification data is forwarded to the speed planning unit 106 and to path planning unit 112 .
- the path planning unit 112 receives input from a route definition data file (RDDF) unit 114 and provides the path data to steering control unit 116 .
- EMC Vehicle Control 118 takes input from the steering control unit 116 and the accelerator/brake control 108 to control the working vehicle.
- RDDF route definition data file
- the obstacle detection unit 104 utilizes in one embodiment of the present invention spatial indices to correlate the sensor data with localization data indicating in which sector an obstacle is located (for example a near sector requiring immediate consideration or a far sector not requiring immediate consideration).
- FIG. 5 shows a division of the field of view into sectors: the sector defined between rays A and B, the sector defined between rays B and C, and the sector defined beyond C.
- each segment of the path corridor has a corresponding spatial index in memory.
- Such division in one embodiment permits the present invention to call up from memory the sectors of importance for example when navigating obstacles 22 a and 22 b.
- the obstacle detection unit 104 in one embodiment of the present invention uses times-stamps taken with each LADAR scan and GPS reading. These timestamps are then correlated based upon position and heading to produce a geospatial location for obstacles, for example the obstacles themselves may be moving. Timestamping can counteract anomalies in the images. Highly reflective surfaces can cause the LADAR devices to register incorrect distances for obstacles. To counteract these anomalies, in one embodiment of the present invention, scans are only kept in memory for a designated amount of time. After that time has passed, if no more obstacles have been registered in that same area, the obstacle is removed from memory storage. This technique ensures that moving obstacles are handled correctly.
- the sensors would effectively register a sequence of obstacles in front of the autonomous vehicle. These obstacles would appear as a complete obstruction of the vehicle's path, forcing the vehicle to immediately stop. After enough time had passed, the obstacles would expire, and the autonomous vehicle would be able to start moving again. A persistent obstacle, on the other hand, would not expire.
- the LADAR devices will start to register parts of the boulder.
- the previous scans would begin to approach the obstacle expiration time. Since the sensors are still registering the boulder, the previous scans will not expire. Instead, the previous scans along with the current scans would all be retained, giving the boulder a higher count for total number of scans. This high count for the number of scans causes the boulder to have an extremely high probability of being an obstacle.
- the path-planning algorithms of the present invention avoid obstacles and routinely steer the vehicle within a planned corridor. In one embodiment, these algorithms without cause will not deviate the vehicle from the planned corridor. However, should the vehicle leave the route corridor for some reason, the navigation system will detect this and provide a safe route back into the corridor. If a waypoint is missed, the navigation system will simply continue to the next feasible waypoint on the path. If the path is obstructed by an obstacle, the path planning systems will determine a path around the obstacle.
- Path planning in one embodiment of the present invention is accomplished through the use of cubic b-splines designed to follow the center of the planned route while still ensuring that the path is not impossible for the vehicle to navigate. This assurance means that the curvature at any point along the path is below the maximum curvature that the vehicle can successfully follow. In addition, the curvature is kept continuous so that it is not necessary to stop the vehicle in order to turn the steering wheel to a new position before continuing. B-splines were chosen for use in the path planning algorithms primarily because of the ease in which the shape of their resulting curves can be controlled. After an initial path is created that follows the center of the corridor, the path is checked against the obstacle repository to determine if it is a safe path. If the path is not safe, a simple algorithm generates and adjusts control points on the problem spots of the curve until the spline avoids all known obstacles while still containing valid maximum curvature. At this point, the path is both safe and drivable.
- the path planning of the present invention can also uses a Level of Detail (LOD) based obstacle avoidance algorithm along with several planning algorithms to plan paths around obstacles.
- LOD analysis in one embodiment of the present invention permits running the same algorithm with different levels of detail. For example, running with less detail (e.g., to accommodate large safety margins), then iteratively increasing the detail (e.g., to accommodate smaller safety margins) until a valid path is found.
- the path planning algorithms run using several different parameters until a valid path is found.
- the initial parameters use safety margins (for example the clearance of the vehicle from an obstacle or between obstacles), while the final parameters use no safety margins around obstacles.
- FIGS. 11 , 12 , and 13 illustrate results of path planning.
- the obstacle detection unit identifies an obstacle 22 by the cluster of data points in the center of the road 80 .
- the path planning unit determines that there is adequate clearance to permit the vehicle 10 to deviate to the right as it advances to the obstacle 22 and then deviate left to return to the center of the road 80 .
- the projected path of the vehicle 10 is shown by path 82 .
- the path 82 of the vehicle 10 passes a center of the obstacle 22 to the left and then to the right as shown.
- the path planning unit has divided the path and the data map into sectors as shown. The first sector at the bottom contains no obstacles and requires no deviation along the planned path 82 .
- the second sector contains an obstacle 22 and requires the path planning unit to determine clearance and a path around the obstacle 22 . Further, deviation from the path will rely on the speed planning unit to control the vehicle so as to safely pass the vehicle at a speed suited for the radius of the turn.
- FIG. 13 shows a vehicle 10 that is about to navigate a series of obstacles in an immediate path of the vehicle.
- the present invention recognizes that smooth paths alone are not necessarily drivable by the vehicle.
- the car should be able to slow down proportionally from straight-aways down to hairpin turns.
- the next step in the planning process is to decide on the speed at which the vehicle should take each section of the path.
- the speed chosen in one embodiment of the present invention, is based on the curvature at that point in the path and upon the curvature further down the path.
- the speed is taken as the minimum of speed for the current curvature and the speed for future path curvature.
- the future path curvature is defined by a simple function that multiplies the curvature at a given future point by a fractional value that decreases towards zero linearly based upon the distance from the current path point to the given future path point.
- System identification is a method used by the present invention by which the parameters that define a system can be determined by relating input signal into a system with the system's response to develop a transfer function that behaves in the same way (or a similar way) as the actual vehicle system. For instance, when attempting to control the speed of a vehicle, the inputs are the brake and accelerator position and the output is the vehicle's speed. If it is assumed that the transfer function, H(s), is first-order, it can be written as
- H(s) is the transfer function of a system
- u(s) is the input to the system
- y(s) is the output from the system.
- System identification was applied to real world data from the vehicle propulsion system to arrive at a transfer function of the vehicle system for example empirically tested for the working vehicle until confidence in an accurate transfer function was obtained.
- speed control of the vehicle accommodated not only accelerator and brake functions but also accommodated many other factors in the physical engine system.
- the coupling of the two propulsion systems was controlled by an inaccessible factory-installed on-board computer tuned for fuel efficiency. Consequently, the mapping of the requested pedal position and the actual position achieved was not linear and had to be remapped in software by empirical determination.
- the speed of the vehicle is controlled by an integrated proportional-derivative (PD) controller. This controller bases its output on the previous output and on the current error and derivative of the error. In the time domain, the controller can be written as
- PD proportional-derivative
- u ( t 2) ( t 2 ⁇ t 1)( K p e ( t 2 )+ K d e ′( t 2 ))+ u ( t 1 ) (9)
- K p and K d are tunable coefficients
- u(t) is the output of the controller at time t
- e(t) is the error at time t.
- the error was defined as actual output subtracted from target output. Actual output was reported by the RT3000TM, and target speed was derived from the path planning algorithms.
- the integrated PD controller was designed and tuned against the derived transfer function detailed above. For instance, the weights (for the proportionate control in the PD controller) needed for optimal performance were derived initially against the computational model of the derived transfer function, and then tuned when operated on the vehicle.
- Accelerator and steering wheel control was achieved in the working vehicle using two separate processes, which were both independent of the path-planning systems. Once a path was decided on by the path-planning algorithms, acceleration and steering is used exclusively to remain on the chosen path. Since paths are checked for feasibility upon creation, it was assumed by the control systems that all paths given are possible for the vehicle to achieve. In this manner (although the present invention can use other starting assumptions), it becomes the burden of the control systems to decide how best to proceed in order to follow the selected path.
- the steering controller for the working vehicle was a lead-lag controller based on the classical single-track model or bicycle model described ed by Riekert and Schunck “Zur fahrmechanik des gummibereiften kraftchrists,” in Ingenieur Archiv, vol. 11, 1940, pp. 210-224, the entire contents of which are incorporated herein by reference.
- a lead compensator increases the responsiveness of a system; a lag compensator reduces (but does not eliminate) the steady state error.
- the lead-lag compensator was based on the frequency response of the system.
- the lead-lag compensator was similar to that described by D. Bernstein, A students guide to classical control, IEEE Control Systems Magazine, vol. 17, pp.
- Adaptive estimation uses a set of values (parameters) first gained from applying the theoretical functions and then iteratively tests and modifies the parameters in real-world scenarios (e.g., deep sand, rough terrain, and other terrain types) until the parameters are perfected.
- the discretized controller was implemented as shown in (12) and (13), where x is the state vector, ⁇ dot over (x) ⁇ is the derivative of the state vector with respect to time, u is the input vector and ⁇ f is the output steering angle as measured at the tire with respect to the centerline.
- the state vector x is defined as [y s ⁇ ] where y s refers to the distance from the virtual sensor to the reference path and ⁇ is the vehicle's yaw rate.
- the virtual sensor is a point projected a given distance ahead of the vehicle along the vehicle's centerline. This point is commonly referred to as the look-ahead point, and the distance from the look-ahead point to the RT3000 is referred to as the look-ahead distance.
- the input vector u to the controller is defined as [y s ].
- the output of the controller is the steering angle measured at the tire with respect to the centerline.
- the steering and control system in the working vehicle assumed that the relationship between the steering wheel angle and the resulting tire angle was linear and that the location of the vehicle's center of gravity was at the midway point between the front and rear axles.
- the magnitude of the y s signal was monitored to prevent the vehicle from becoming unstable. If y s were to cross a given threshold, meaning the vehicle is severely off path, the speed of the vehicle was reduced to 2 mph. This allowed the vehicle to return onto the desired path and prevented a possible rollover.
- the present invention has widespread application in both autonomously driven and human driven vehicles where the present invention functions respectively in a primary or secondary mode.
- the present invention with various of the attributes described above can be included on a variety of driveable units.
- Such units in one embodiment of the present invention are powered land or water based vehicles, which include for example automobiles, trucks, sport utility vehicles, armored vehicles, boats, ships, barges, tankers, and armored vessels.
- the present invention could be used not only when navigating in weather or nighttime conditions where visibility is limited and avoidance of other watercraft is desired, but also in docking and lock operations where control of the watercraft between dock and lock structures is important to minimize damage to the watercraft of the dock and lock structures.
- the present invention also has application to aircraft.
- the application to high speed aircraft will depend on detector sensitivity and accuracy to determine the existence of objects at a sufficient distance from the plane that the plane can take corrective action.
- airport approach and takeoff where the speeds are not high the present invention has utility.
- Birds are hard to see at distance for the human eye and hard to detect by aircraft radar.
- the present invention has application to helicopters, especially rescue helicopters that frequently land in make-shift areas with obstacles many times hidden from the view of the pilot.
- lighter-than-air vehicles e.g., autonomous weather balloons, autonomous border patrol system including for example remote control small aircraft
- amphibious vehicles such as for example land-sea assault vehicles including hovercrafts etc.
- the driveable unit can be an autonomous vehicle without driver-assisted control or a driver-controlled vehicle with computer-assisted control.
- the autonomous vehicles find application, according to one embodiment of the present invention, in the above noted environmentally dangerous surroundings where a driver would be at risk.
- the driver-controlled vehicles of the present invention find application in the above noted more conventional environments, where the driver may become disoriented or incapable of physically commanding the vehicle as would occur if the driver suffered a medical emergency or if for example the driver became disoriented under an adverse driving condition.
- processor 24 is configured to control the vehicle in an event of driver impairment or in an event of immediate path obstruction or in an event of driver request.
- the autonomous vehicle could recognize driver impairment through observing where the driver is driving in comparison to a predetermined path. If the present path of the vehicle and the predetermined path are not similar, the autonomous vehicle could then check, for example, to see if the driver has recently turned the steering wheel and/or pressed on the either the brakes or throttle. Both of the comparison and the steering and brake check can be included in the decision making process, because if the driver were driving on a long straight road with cruise control on, he may not be actively turning the steering wheel or applying brake or throttle. By the same logic, the driver could be driving a path that is not consistent with the autonomous vehicle's path as long as the driver is actively steering the car.
- the autonomous vehicle in one embodiment of the present invention first audibly and visibly warns the driver that it is taking over, then second takes over and steers the vehicle to a safe stopping position as smoothly and safely as possible. If the driver wanted to regain control, the autonomous vehicle of the present invention, in one embodiment, provides a press-button or other input device for the driver to resume control. In another embodiment of the present invention, the driver could press the button (or command input) again to relinquish control to the autonomous vehicle. Hence, the present invention provides in various embodiments a cooperative autonomous driving mode.
- a biometric sensor could represent another input device.
- the biometric sensor determines if the driver was actively driving through input to processor 24 from for example a heart rate monitor built into the steering wheel of the vehicle.
- a heart rate monitor that is suitable for the present invention is heart rate used in exercise equipment, which in one embodiment would be integrated into the steering wheel or alternatively could be worn by the driver and be in wireless communication to processor 24 . If the processor 24 detected either a complete loss of heart rate or an extremely low heart rate for an extended period of time (for example 5 seconds), the processor 24 is configured to take control of the vehicle.
- the processor by monitoring the normal heart rate of the driver when the car was under proper control would have a basis for determining for example if the driver was impaired due to haven fallen asleep at the wheel, having had a stroke, or having had a heart attack.
- This embodiment could be implemented also in the cooperative mode of operation (discussed above).
- an audible alarm is issued before taking over control of the vehicle and steering the vehicle to a safe stop. If the driver was in fact not impaired, the driver could simply press a button (or other input device) to take control back from the processor 24 .
- the autonomous vehicle can be operated repeatedly over a predetermined course. For instance, a human driver presses a button that turns the autonomous vehicle into record mode. The human drives the vehicle exactly like he would want the autonomous vehicle to drive the course. The human driver then presses the button again, and the autonomous vehicle drives the recorded course over and over again with a very high level of reliability and repeatability. (Repeatability is an issue for automobile testers). This capability is also useful for endurance testing vehicles, where the vehicle is driven offroad in hazardous conditions for days in a row in which currently many human drivers are used to perform this task due to the human body's relative weakness. This capability is also useful for driving a vehicle at consistent speeds for long distances. For example, this capability would be useful in testing a vehicle at highway speeds on a race track for fuel consumption performance.
- the driveable unit can be used, in conjunction with a mapping program (e.g., running on a laptop), in which the user could select a destination.
- a mapping program e.g., running on a laptop
- the autonomous navigation software would access the mapping software and generate a route (like the software normally does) except in GPS waypoints rather than in human directions like “turn left at main street.”
- the autonomous vehicle proceeds in normal operation to follow that route.
- the mapping software is customized to provide additional information to the autonomous navigation program such as the width of streets and speed limits.
- Other application areas for the navigation and control system of the present invention include, but are not limited to: 1) agricultural equipment (e.g., gardening equipment) that performs repetitive tasks or tasks on predicted paths such as for example the harvesting of crops from fields, plowing, grass cutting, etc., 2) mining equipment including for example powered carts that could transport for example goods or people through blackout or smoke filled passages that would normally prevent escape, 3) cave exploration equipment, 4) emergency or police vehicles such as for example fire fighting, ambulances, and rescue squads where visibility impairment need not prevent the emergency vehicles from proceeding forward (as discussed below) or vehicles operating in hazardous environmental conditions where manning of the vehicle places the driver at risk, 5) warehouse management equipment used to store/retrieve pallets, boxes, etc., and 6) toys.
- agricultural equipment e.g., gardening equipment
- mining equipment including for example powered carts that could transport for example goods or people through blackout or smoke filled passages that would normally prevent escape
- cave exploration equipment 4) emergency or police vehicles such as for example fire fighting, ambulances, and rescue squads where
- the autonomous vehicle of the present invention As an illustrative example of the application of the autonomous vehicle of the present invention to a police vehicle, on the Causeway bridge in New Louisiana (the longest bridge in the world at 24 miles long), there is a significant fog season. On mornings with a dense fog, traffic is convoyed across the bridge by a police car driving at 35 mph. The low speed is required due to the extremely short visibility. On mornings with intense fog, the bridge is closed as even convoying is not possible. If the lead police car in the convoy were an autonomous vehicle of the present invention operating in the above-noted cooperative mode, the police car could safely convoy in any type of fog, especially when the RIEGLTM imaging sensor is used. The same applies to driving at night. The autonomous vehicle of the present invention is not affected by darkness.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Automation & Control Theory (AREA)
- General Physics & Mathematics (AREA)
- Transportation (AREA)
- Mechanical Engineering (AREA)
- Chemical & Material Sciences (AREA)
- Combustion & Propulsion (AREA)
- Aviation & Aerospace Engineering (AREA)
- Electromagnetism (AREA)
- Computer Networks & Wireless Communication (AREA)
- Optics & Photonics (AREA)
- Traffic Control Systems (AREA)
- Navigation (AREA)
- Optical Radar Systems And Details Thereof (AREA)
Abstract
A navigation and control system including a sensor configured to locate objects in a predetermined field of view from a vehicle. The sensor has an emitter configured to repeatedly scan a beam into a two-dimensional sector of a plane defined with respect to a first predetermined axis of the vehicle, and a detector configured to detect a reflection of the emitted beam from one of the objects. The sensor includes a panning mechanism configured to pan the plane in which the beam is scanned about a second predetermined axis to produce a three dimensional field of view. The navigation and control system includes a processor configured to determine the existence and location of the objects in the three dimensional field of view based on a position of the vehicle and a time between an emittance of the beam and a reception of the reflection of the emitted beam from one of the objects.
Description
- This application is a continuation of Ser. No. 13/240,576 filed Sep. 22, 2011 which is a continuation of U.S. patent application Ser. No. 11/376,160, filed Mar. 16, 2006, the entire contents of each are incorporated herein by reference.
- 1. Field Of The Invention
- The invention relates to a three-dimensional (3D) imaging sensor and method for controlling an autonomous vehicle.
- 2. Background of the Invention
- In a modern vehicle, the driver remains a critical component of the vehicle's control system as the driver makes numerous decisions directed to the safe operation of the vehicle including speed, steering, obstacle and hazard recognition, and avoidance thereof. Yet, the driver's ability to perform all of these functions can become compromised due to physical factors such as driver fatigue, driver impairment, driver inattention, or other factors such as visibility that reduce the reaction time needed by the driver to successfully avoid hazards.
- Furthermore, in environmentally dangerous surroundings such as for example in warfare settings or in settings where toxic or nuclear radiation hazards are present, the driver is at risk. Indeed, roadside bombs in Iraq are just one contemporary example of the loss of human life which could in many situations be avoided if supply trucks bringing materials to the troops were unmanned.
- In other more conventional environments, the driver may become disoriented or incapable of physically commanding the vehicle as would occur if the driver suffered a medical emergency or if for example the driver became disoriented under the driving conditions. One example of such a disorienting or incapacitating environment would be a car or ship being driven or steered under snow, fog, rain, and/or nighttime blackout conditions where the diver (or captain of the ship) is handicapped in his or her ability to perceive and react to hazards approaching or to which the ship is approaching.
- Thus, whether addressing human deficiencies in the control of a vehicle or whether in environmentally hazardous conditions where human control is not preferred, there exists a need to have a system and method for vehicular identification of objects in the path of the vehicle.
- Numerous articles on the development of autonomously driven vehicles and laser detection and visualization systems have been reported such as the following reference articles all of which are incorporated herein by reference:
-
- 1) H. Wang, J. Kearney, J. Cremer, and P. Willemsen, “Steering Autonomous Driving Agents Through Intersections in Virtual Urban Environments,” 2004 International Conference on Modeling, Simulation, and Visualization Methods, (2004);
- 2) R. Frezza, G. Picci, and S. Soatto, “A Lagrangian Formulation of Nonholonomic Path Following,” The Confluence of Vision and Control, (A. S. Morse et al. (eds), Springer Verlag, 1998);
- 3) J. Shirazi, Java Performance Tuning, (OReilly & Associates, 2000);
- 4) J. Witt, C. Crane III, and D. Armstrong, “Autonomous Ground Vehicle Path Tracking,” Journal of Robotic Systems, (21(8), 2004);
- 5) C. Crane III, D. Armstrong Jr., M. Torrie, and S. Gray, “Autonomous Ground Vehicle Technologies Applied to the DARPA Grand Challenge,” International Conference on Control, Automation, and Systems, (2004);
- 6) T. Berglund, H. Jonsson, and I. Soderkvist, “An Obstacle-Avoiding Minimum Variation B-spline Problem,” International Conference on Geometric Modeling and Graphics, (July, 2003);
- 7) D. Coombs, B. Yoshimi, T. Tsai, and E. Kent, ‘Visualizing Terrain and Navigation Data,” NISTIR 6720, (Mar. 1, 2001);
- 8) U.S. Pat. No. 5,644,386 to Jenkins et al;
- 9) U.S. Pat. No. 5,870,181 to Andressen;
- 10) U.S. Pat. No. 5,200,606 to Krasutsky et al; and
- 11) U.S. Pat. No. 6,844,924 to Ruff et al;
Despite this work, realization of suitable visualization, obstacle identification, and obstacle avoidance systems and methods has not been without problems limiting the operation of vehicles.
- Accordingly, one object of the present invention accomplished in various of the embodiments is to provide a system and method for vehicle object identification in which a location of the object relative to the vehicle is identified.
- Another object of the present invention accomplished in various of the embodiments is to provide a system and method for vehicle object identification in which a path for the vehicle relative to the location of the object is identified.
- Yet another object of the present invention accomplished in various of the embodiments is to provide a system and method for identifying objects in a field of view of the vehicle as obstacles.
- Still another object of the present invention accomplished in various of the embodiments is to provide steering and vehicle control directions to avoid the obstacles
- Various of these and other objects are provided for in certain ones of the embodiments of the present invention.
- In one embodiment of the present invention, a navigation and control system includes a sensor configured to locate objects in a predetermined field of view from a vehicle. The sensor has an emitter configured to repeatedly scan a beam into a two-dimensional sector of a plane defined with respect to a first predetermined axis of the vehicle, and a detector configured to detect a reflection of the emitted beam from one of the objects. The sensor includes a panning mechanism configured to pan the plane in which the beam is scanned about a second predetermined axis to produce a three dimensional field of view. The navigation and control system includes a processor configured to determine the existence and location of the objects in the three dimensional field of view based on a position of the vehicle and a time between an emittance of the beam and a reception of the reflection of the emitted beam from one of the objects.
- In one embodiment of the present invention, a method for navigation and control of a vehicle includes scanning a beam into a sector of a plane defined with respect to a first predetermined axis of the vehicle, detecting a reflection of the emitted beam from an object removed from the vehicle, panning the plane in which the beam is scanned about a second predetermined axis to produce a three dimensional field of view, and determining the existence and location of the object in the three dimensional field of view based on a position of the vehicle and a time between an emittance of the beam and a reception of the reflection of the emitted beam.
- It is to be understood that both the foregoing general description of the invention and the following detailed description are exemplary, but are not restrictive of the invention.
- A more complete appreciation of the present invention and many attendant advantages thereof will be readily obtained as the same becomes better understood by reference to the following detailed description when considered in connection with the accompanying drawings, wherein:
-
FIG. 1A is a schematic illustration of an autonomous vehicle according to one embodiment of the present invention in which a two-dimensional (2D) scan is made in a sector of a plane normal to a predetermined axis of a vehicle; -
FIG. 1B is a schematic illustration of an autonomous vehicle according to one embodiment of the present invention in which a three-dimensional (3D) scan is made by displacing the scan out the plane normal to the predetermined axis of a vehicle; -
FIG. 2 is a schematic illustration of an emitter and detector system according to one embodiment of the present invention; -
FIG. 3 is a detailed schematic illustration showing a perspective frontal view of a laser radar (LADAR) imaging sensor according to one embodiment of the present invention; -
FIG. 4 is a detailed schematic illustration showing a perspective rear view of the laser radar (LADAR) imaging sensor depicted inFIG. 3 ; -
FIG. 5 is a contour map of a 3D scan obtained by methods of the present invention; -
FIG. 6A is a flow chart illustrating particular method steps of the present invention; -
FIG. 6B is a flow chart illustrating particular method steps of the present invention for obstacle identification and avoidance; -
FIG. 7A is a schematic illustration of a vehicle according to the present invention utilizing multiple LADAR imaging sensors to scan separate fields of view; -
FIG. 7B is a schematic illustration of a vehicle according to the present invention utilizing multiple LADAR imaging sensors to scan the same or overlapping fields of view; -
FIG. 8 is a schematic illustration of an exemplary computer system of the present invention; -
FIG. 9 is a high level schematic of a computing network system of the present invention; -
FIG. 10 is a high level schematic of the vehicle control system of the present invention; and -
FIGS. 11-13 are schematics of path determinations made by the present invention. - Referring now to the drawings, wherein like reference numerals designate identical, or corresponding parts throughout the several views, and more particularly to
FIG. 1A , which depicts animaging sensor 8 mounted, in one embodiment, on top of avehicle 10 in which a two-dimensional (2D) scan is made in a sector of aplane 11 normal to a predetermined axis of thevehicle 10 referred to here for illustration purposes as a “vertical” scanning plane. Theimaging sensor 8 includes in one embodiment an emitter 12 (as shown inFIG. 2 ) that transmits laser pulses (or light) 14 from theimaging sensor 8 into the environment about thevehicle 10. As shown inFIG. 1A , the laser (or light)pulses 14 are emitted into thevertical scanning plane 11. To produce a three-dimensional (3D) image, theimaging sensor 8 is panned (or oscillated) in and out ofplane 11 to create a3D scanning volume 16, as shown inFIG. 1B . Theimaging sensor 8 detects objects 22 (as shown inFIG. 1B ) in the environment nearby thevehicle 10 by detecting light reflected from theobjects 22. - As shown in
FIG. 2 , theimaging sensor 8 includes adetector 18 for detecting return of an echoed signal 20. Theimaging sensor 8 utilizes a processor 24 for controlling the timing and emission of thelaser pulses 14 and for correlating emission of thelaser pulses 14 with reception of the echoed signal 20. The processor 24 may be on-board the vehicle or a part of theimaging sensor 8. Details of exemplary processors and their functions are provided later. - In an exemplary example,
laser pulses 14 from emitter 12 pass through a beam expander 13 a and a collimator 13 b. Thelaser pulses 14 are reflected at a stationary mirror 15 a to a rotating mirror 26 and then forwarded through lens 27 a and a telescope 27 b to form a beam for thelaser pulses 14 with a diameter of 1-10 mm, providing a corresponding resolution for the synthesized three-dimensional field of view. The telescope 27 b serves to collect light reflected from objects 22. - In one embodiment of the present invention, the
detector 18 is configured to detect light only of a wavelength of the emitted light in order to discriminate the laser light reflected from the object back to the detector from background light. Accordingly, theimaging sensor 8 operates, in one embodiment of the present invention, by sending out alaser pulse 14 that is reflected by anobject 22 and measured by thedetector 18 provided the object is within range of the sensitivity of thedetector 18. The elapsed time between emission and reception of the laser pulse permits the processor 24 is used to calculate the distance between theobject 22 and thedetector 18. In one embodiment of the present invention, the optics (i.e., 13 a, 13 b, 15 a, 26, 27 a, and 27 b) are configured to direct the beam instantaneously into the sector shown inFIG. 1A , and thedetector 18 is a field-programmable gate array for reception of the received signals at predetermined angular positions corresponding to a respective angular direction α1 shown inFIG. 1A . - Via the rotating mirror 26,
laser pulses 14 are swept through a radial sector a withinplane 11, as shown illustratively inFIG. 1A . In one embodiment of the present invention, in order to accomplish mapping of objects in the field of view in front of theimaging sensor 8, the rotating mirror 26 is rotated across an angular displacement ranging from 30 to 90 degrees, at angular speeds ranging from 100-10000 degrees per second. For example, a 90 degree scanning range can be scanned 75 times per second or an 80 degree scanning range can be scanned between 5 and 100 times per second. Furthermore, the angular resolution can be dynamically adjusted (e.g., providing on command angular resolutions of 0.01, 0.5, 0.75, or 1 degrees for different commerciallyavailable sensors 8. - Commercially available components can be used for the emitter 12 and the
detector 18 to provide ranging measurements. In one embodiment, the emitter 12, thedetector 18, and the associated optics constitute a laser radar (LADAR) system, but other systems capable of making precise distance measurements can be used in the present invention, such as for example a light detection and ranging (LIDAR) sensor, a radar, or a camera. LIDAR (Light Detection and Ranging; or Laser Imaging Detection and Ranging) is a technology that determines distance to an object or surface using laser pulses. Like the similar radar technology, which uses radio waves instead of light, the range to an object is determined by measuring the time delay between transmission of a pulse and detection of the reflected signal. LADAR (Laser Detection and Ranging) refers to elastic backscatter LIDAR systems. The term laser radar is also in use, but with laser radar laser light (and not radio waves) are used. - The primary difference between LIDAR and radar is that with LIDAR, much shorter wavelengths of the electromagnetic spectrum are used, typically in the ultraviolet, visible, or near infrared. In general it is possible to image a feature or object only about the same size as the wavelength, or larger. Thus, LIDAR provides more accurate mapping than radar systems. Moreover, an object needs to produce a dielectric discontinuity in order to reflect the transmitted wave. At radar (microwave or radio) frequencies, a metallic object produces a significant reflection. However non-metallic objects, such as rain and rocks produce weaker reflections, and some materials may produce no detectable reflection at all, meaning some objects or features are effectively invisible at radar frequencies. Lasers provide one solution to these problems. The beam densities and coherency are excellent. Moreover the wavelengths are much smaller than can be achieved with radio systems, and range from about 10 micrometers to the UV (e.g., 250 nm). At these wavelengths, a LIDAR system can offer much higher resolution than radar.
- To produce a three-dimensional (3D) image, in one embodiment of the present invention, the
imaging sensor 8 is panned (or oscillated) in and out theplane 11 to create a3D scanning volume 16, as shown inFIG. 1B . For sake of illustration,FIG. 1B defines thescanning volume 16 by the angle α (in the vertical scanning direction) and the angle β (in the horizontal scanning direction). The angle α, as noted earlier, ranges from 30 to 70 degrees, at angular speeds ranging from 100-1000 degrees per second. The angle β (i.e., the panning angle) ranges from 1 to 270 degrees, at a panning rate ranging from 1-150 degrees per second. Combined theimaging sensor 8 typically can completely scan the3D scanning volume 16 at more than two times a second. - In order to accurately determine the distance to objects in the
3D scanning volume 16, the direction that theimaging sensor 8 is pointed at the time of receiving light reflected from theobjects 22 is needed (i.e., the angle of deflection from plane 28 is needed). Further, in one embodiment of the present invention, geospatial positional data of the instantaneous vehicle position is utilized by processor 24 to calculate based on the distance of the object from the vehicle and its direction from the vehicle, the geospatial location of the objects in the field of view. In one configuration of the present invention, the processor 24 includes a personal computer running on a Linux operating system, and the algorithms are programmed in Java programming language. Other computing systems and programming languages can be used in the present invention (as discussed in more detail below). As shown inFIG. 2 , processor 24 is in communication with a real time positioning device 25, such as for example a global positioning system (GPS) and/or an inertial navigation system (INS), that transmits the location, heading, altitude, and speed of the vehicle multiple times per second to processor 24. The real time positioning device 25 is typically mounted to thevehicle 10 and transmits data (such as location, heading, altitude, and speed of the vehicle) to all imaging sensors 8 (and all processors 24) on thevehicle 10. - With commercially available GPS and the INS units, processor 24 can determine a position of an object in the field of view to an accuracy of better than 10 cm. In one embodiment of the present invention, the processor 24 correlates GPS position, LADAR measurements, and angle of deflection data to produce a map of obstacles in a path of the vehicle. The accuracy of the map depends on the accuracy of the data from the positioning device 25. The following are illustrative examples of the accuracies of such data:
position 10 cm, forward velocity 0.07 km/hr, acceleration 0.01%, roll/pitch 0.03 degrees, heading 0.1 degrees, lateral velocity 0.2%. - In one embodiment of the present invention, a Kalman filter (commercially integrated) sorts through all data inputs to processor 24. A Kalman filter is a known method of estimating the state of a system based upon recursive measurement of noisy data. In this instance, the Kalman filter is able to much more accurately estimate vehicle position by taking into account the type of noise inherent in each type of sensor and then constructing an optimal estimate of the actual position. Such filtering is described by A. Kelly, in “A 3d State Space Formulation of a Navigation Kalman Filter for Autonomous Vehicles,” CMU Robotics Institute, Tech. Rep., 1994, the entire contents of which are incorporated herein by reference. The Kalman filter is a set of mathematical equations that provides an efficient computational (recursive) means to estimate the state of a process, in a way that minimizes the mean of the squared error. The filter is very powerful in several aspects: it supports estimations of past, present, and even future states, and it can do so even when the precise nature of the modeled system is unknown.
- The positioning device 25, by including GPS and INS data, provide complementary data to the processor 24. GPS and INS have reciprocal errors. That is GPS is noisy with finite drift, while INS is not noisy but has infinite drift. Further, the processor 24 can be configured to accept additional inputs (discussed below) to reduce drift in its estimate of vehicle position when, for example the GPS data may not be available.
-
FIG. 3 is a detailed schematic illustration ofimaging sensor 8 of the present invention.FIG. 3 presents a frontal view ofimaging sensor 8.FIG. 4 presents a rear perspective view ofimaging sensor 8.FIG. 3 shows amotor 30 configured to oscillate theimaging sensor 8 in and out of a plane normal to a predetermined axis of the imaging sensor. In one embodiment of the present invention, a 12-volt DC motor operating at a speed of 120 RPM is used to oscillate theimaging sensor 8 in and out the plane. Other motors with reciprocating speeds different than 120 RPM can be used. - As shown in
FIG. 3 , an absoluterotary encoder 32 is placed on ashaft 34 that is oscillating. Theencoder 32 provides an accurate reading of the angle at which theshaft 34 is instantaneously located. By theencoder 32, an accurate measurement of the direction that theimaging sensor 8 is pointed, at the time of the scan, is known. In one embodiment of the present invention, theencoder 32 is an ethernet optical encoder (commercially available from Fraba Posital), placed onshaft 34 to provide both the angular position and angular velocity of the shaft. - To decrease the delay between reading a value from the sensor and reading a value from the encoder, a separate 100 MBit ethernet connection with its own dedicated ethernet card connected the processor 24 with the encoder. This created communications delays between the encoder and the I/O computer that were consistent at approximately 0.5 ms. Testing revealed that an actual LADAR scan was taken approximately 12.5 ms before the data was available at the I/O computer. When this time was added to the 0.5 ms of delay from the encoder communications, a 13 ms delay from the actual scan to the actual reading of the encoder position and velocity was present. To counteract the angular offset this delay created, in one embodiment of the present invention, the velocity of the encoder is multiplied times the communications delay of 0.013 seconds to calculate the angular offset due to the delay. This angular offset (which was either negative or positive depending on the direction of oscillation) was then added to the encoder's position, giving the actual angle at the time when the scan occurred. This processing permits the orientation of the LADAR platform to be accurate within 0.05 degrees.
- Further, as shown in
FIGS. 3 and 4 , themetal shaft 34 is attached to a detector bracket 36 which is supported by ametal casing 38 withbearing 40.Bearing 40 is attached tometal casing 38 with a fastening mechanism such asbolts metal shaft 34. Further, as shown inFIGS. 3 and 4 ,metal shaft 46 is attached to bearing 48.Bearing 48 is attached tometal casing 38 with a fastening mechanism such asbolts rod 54 is attached to detector bracket 36 with ball joint 56 onslot 58. Pushrod 54 is attached to pivotspacer 60 with ball joint 62.Pivot spacer 60 is attached toservo arm 64 onslot 66.Servo arm 64 is attached tometal shaft 68.Motor 30 is attached toservo arm 64 and is suspended frommetal casing 38 by motor mounts 70, 72, and 74. - The
imaging sensor 8 operates, in one embodiment, by oscillating a measurement sensor laterally about an axis of thevehicle 10, as shown inFIG. 1A . In the one embodiment, theshaft 68 ofmotor 30 rotates at a constant speed, causingservo arm 64 to also spin at a constant speed. One end ofpush rod 54 moves withservo arm 64, causing detector bracket 36 to oscillate back and forth. The degree of rotation can be adjusted by moving the mount point of ball joint 56 alongslot 58, and/or the mount point of ball joint 62 alongslot 66. Moving the mount point closer toshaft 46 increases the angle of rotation, while moving the mount point away fromshaft 46 decreases the angle of rotation. - While
sensor 18 is oscillating, thesensor 18 is taking measurements of the surrounding environment along the vertical scanning plane, as shown inFIG. 1A . The absoluterotary encoder 32 operates as an angular position mechanism, and transmits the absolute angle of deflection of detector bracket 36 to processor 24. At the same time, a real time positioning device 76, such as a global positioning system (GPS) or an inertial navigation system (INS), transmits the location, heading, altitude, and speed of the vehicle multiple times per second to processor 24. Software running on the processor 24 integrates the data, and, in one embodiment, uses matrix transformations to transform the YZ measurements from each 2D scan (as shown inFIG. 1 ) into a 3D view of the surrounding environment. Due to the use of the real time positioning device, in the present invention, a terrain map can be calculated even while the vehicle is moving at speeds in excess of 45 miles per hour. - The
imaging sensor 8 is configured, in one embodiment of the present invention, to receive data in the processor regarding echoed or returned signals, GPS position, LADAR measurements, and angle of deflection data to determine what sectors of the 3D scan contain obstacles, and therefore to determine which sectors are most likely impassable by the vehicle. In one embodiment of the present invention, data coming fromdetector 18 is formatted in a coordinate system relative to thedetector 18, as detailed in the exemplary illustrations below. - Since the
imaging sensor 8 is in a “vertical” configuration, the coordinate system in that configuration has Y-axis representing a distance from the sensor roughly parallel with the forward direction of the vehicle, with the X-axis lined up along the vertical direction. Hence, in one embodiment of the present invention, a search for obstacles is manifest in the slope and extent in the X-axis. To facilitate this search, all measurement points in the detector's scan are sorted by the Y-values (i.e., by a distance from the detector 18). -
FIG. 5 is a schematic representation of a contour map showing the reflected signatures from an object in a synthesized field of view.FIG. 5 depicts two common obstacles that thevehicle 10 needs to avoid.Obstacle 22 a is a depression in the highway (e.g., a pothole).Obstacle 22 b is an obstruction in the highway (e.g., a fallen boulder). As shown inFIG. 5 , lines of constant elevation have been added to the drawing inFIG. 5 for clarity. Theimaging sensor 8 of the present invention, as shown inFIG. 1A , produces a vertical scan, which inFIG. 5 is shown depicted on an X-Y axis. A vertical scan provides the advantage of directly supplying slope data (i.e., ΔX/ΔY data) along the road and obstacles shown inFIG. 5 without having to complete a full 3D scan, as shown inFIG. 1B , and then process the image map for slope data. Furthermore, in data from a vertical scan plane as inFIG. 1A , obstacles appear as either clusters of similar distance or gaps in distance, making obstacle identification more discernable in the data. - According to one embodiment of the present invention, any point P is classified as part of an obstacle if and only if there could be found a point Q such that Equations (1), (2), and (3) are satisfied (where hmax is the maximum height or depth, mmax is the maximum slope, is the unit vector in the X direction, and j is the unit vector in the Y direction).
-
Q·ĵ>P·ĵ (1) -
|Q·î−P·î|>h max (3) - Equation (1) represents a condition in which there exists a first point Q in the field of view that is farther away from the vehicle than a second point P. Equation (2) represents a condition in which a slope between the first point and the second point is greater than a predetermined slope. Equation (3) represents a condition in which a height between the first point and the second point is greater than a predetermined height.
- Once an obstacle is found relative to the sensor, its location must be translated into the global reference frame (i.e., a geospatial reference frame). This can be accomplished in two similar steps. In each step, a transformation matrix is defined so that
-
P 2 =T 1→2 P 1+Δ1 (4) - where T1→2 is the transformation matrix for going from coordinate
frame 1 to coordinate frame 2, Δ1 is the vector representing the position of the coordinateframe 1 with respect to coordinate frame 2, and P1 and P2 are the same point in coordinateframes 1 and 2, respectively. - The first step converts from the detector coordinate frame to the vehicle's coordinate frame. The transformation matrix, Ts→v, is defined as
-
- where Ψs, θs and φs are the yaw (around the z-axis), pitch (around the y-axis), and roll (around the x-axis) of the detector coordinate frame relative to the vehicle's coordinate frame. This transformation takes into account deviations in yaw, pitch, or roll caused by the mounting of the sensor. For example, if the detector were mounted pointed slightly downward, it would have a negative pitch that would need to be countered by setting θs to its inverse (or positive) value. In addition, the angle of deflection caused by the oscillation is processed here by adding it to θs.
- In one embodiment of the present invention, the same basic transformation and translation is performed again in order to translate the obstacle's location from the vehicle's coordinate system to a global system. Yet another transformation matrix, Tv→g, is constructed for this purpose.
-
- where Ψv, θV and φv are the heading (around the z-axis), pitch (around the y-axis), and roll (around the x-axis) of the vehicle relative to the global coordinate frame. These heading, pitch, and roll values are generated by the GPS/INS navigation sensor that is mounted on the vehicle.
- After taking into account both of these transformation matrices, the full equation for transformation from detector frame to the global coordinate frame is:
-
P g =T v→g(T s→v P s+Δs)+Δv. (7) - where Δs is the vector representing the position of the sensor with respect to the center of the vehicle and Δv is the vector representing the position of the center of the vehicle with respect to the center of the GPS/INS navigation sensor.
- At this point, in one embodiment of the present invention, the geospatial locations of all obstacles in the LADAR sensor's current scan are known. Each of these obstacles is now checked to determine if it can be safely ignored. If an obstacle is far enough away, it can safely be ignored for the time being.
FIG. 5 shows for the purposes of illustration a point R beyond theobstacles obstacles detector 18 to see objects beyond a sensitivity of thedetector 18. Discarding object data serves two purposes. One, the further away an object is from the LADAR unit, the higher the chance that it is a false signature. False signatures could cause the autonomous vehicle to be unable to find a path, so they must be avoided at all costs. Two, by reducing the number of obstacles that must be processed at any given time, less computing resources are needed. The remaining obstacles are now processed to determine if the vehicle's current path will collide with any of these obstacles. If any of these obstacles will cause a collision, the path is shifted to avoid the obstacles (as will be discussed later). - Associated with the distance criteria, the processor 24 in one embodiment of the present invention discards objects in the field of view of the
detector 18 that are beyond a travel-point distance. The travel-point distance (also exemplified by point R onFIG. 5 ) is in one embodiment the calculated distance that is an arbitrary factor (e.g., factors of 5, 10, 100) multiplied times the distance that the vehicle will travel at the current speed before the next complete 3D scan ofangular sector 16 is made. The travel-point distance is in one embodiment set to an arbitrary distance (e.g., 100 m). The travel-point distance in different embodiments of the present invention can be calculated based on linear, sigmoid, or exponential functions that are evaluated using values such as vehicle speed and time and path. By discarding objects in the field of view beyond the travel-point distance, the processor will not be required to process data associated with objects so far removed from the vehicle that determination of whether or not the objects are obstacles to the path of the vehicle will make no difference in the steering of the vehicle. - Accordingly, in one embodiment of the present invention, every vertical scan produces from the slope and height data a set of obstacles that meet the vector criteria defined in eqns. (1)-(3). Since the vertical scan represents a very small slice of the 3D viewing field, the location of the obstacle can be represented by a polygon represented in this illustrative example by a series of spatial coordinate points including a center point and four corner points: 0.2 m from the back of center, 0.2 m from front of center, 0.2 m from left of center, and 0.2 m from right of center. In one embodiment of the present invention, spatial indices such as a quadtree stores the polygon and efficiently subdivides space, so that object retrieval (such as identifying if an obstacle exists within a space) is faster than having to search a plot map containing obstacles on a point-by-point analysis. Fitch et al., U.S. Pat. No. 6,212,392, the entire contents of which are incorporated herein by reference, describe the use of quadrees to determine whether or not a location associated with a wireless station is within an area of interest. Bao et al., “LOD-based Clustering Techniques for Efficient Large-scale Terrain Storage and Visualization,” in Proceedings SPIE Conference on Visualization and Data Analysis, pages 225-235 (2003), the entire contents of which are incorporated herein by reference, describe memory algorithms suitable for the present invention in which clustering algorithms and data structures including hierarchical quadree triangulation are explained in detail.
- In one embodiment of the present invention, a navigation and control system includes a sensor (e.g., imaging sensor 8) configured to locate objects in a field of view from a vehicle (e.g., vehicle 10) and a processor (e.g., processor 24) configured to determine the existence and location of the objects in the field of view based on a position of the vehicle and a time between an emittance of a beam from the sensor and a reception of a reflection of the emitted beam from one of the objects. In this embodiment, the processor includes a memory including storage elements corresponding to subsectors in the field of view. The processor is configured 1) to perform an obstacle identification algorithm to determine if any object is an obstacle, 2) to determine in which of the field of view subsectors a determined obstacle exists, and 3) to store in each corresponding memory element an obstacle indicator indicative that an obstacle exists in the corresponding subsector. In one embodiment of the present invention, processor 24 is configured to determine if a projected path of the vehicle intersects any subsector storing an indicator. In one embodiment of the present invention, processor 24 is configured to execute an obstacle identification algorithm including: determining first and second points on a contour map of an object, determining if the first point is farther away from the vehicle than the second point, determining a slope between the first point and the second point, determining if the slope is greater than a predetermined slope, and determining if a height between the first point and the second point is greater than a predetermined height, and if so, determining the object is an obstacle.
- A path planning algorithm (to be discussed in more detail later) according to one embodiment of the present invention presents a “query” polygon for a sector in front of the vehicle and determines from the spatial index (e.g., a quadtree) if any of the stored obstacle polygons intersect the “query” polygon. If so, a path around the query polygon is chosen. If not, the path is clear of obstacles. The way that a quadtree (or most spatial indices) eliminates sectors of space is by keeping track of where it has inserted obstacles. For the quadtree techniques used in one embodiment of the present invention, the quadtree subdivides space as obstacles are inserted, thereby creating an indicator that this sector of space contains an obstacle. This leaves many sectors that are empty of obstacles, and the quadtree knows that these sectors are empty. Therefore, if the query is within one of these sectors, it cannot have an obstacle collision.
- One conventional way to determine the number of obstacles in the path of the vehicle is to mount stationary LADAR units pointed at the ground. With these LADAR units being pointed at the ground, the angle at which the LADAR beam hits the ground is calculated. Based on this angle, simple trigonometry can determine where the object is located relative to the car. In most conventional cases, multiple stationary LADAR units are used and each is placed at a different known angle.
- There are several disadvantages to the conventional approach:
- 1) The conventional approach requires movement from the autonomous vehicle in order to map the environment. This means that while the vehicle is stationary, virtually no mapping is happening. The only data being mapped is from the same scans over and over again. Another problem with relying on vehicle movement to map data is that the integrity of the scans is dependent on the vehicle's speed. As the vehicle goes faster, the distance between scans increases. This could lead to obstacles being missed.
- 2) Moving obstacles cannot be detected reliably. This is due to the fact that the LADAR scans are fixed. If an obstacle were to be in between two LADAR scans and moving at a similar speed and direction as the vehicle (e.g. another vehicle), the obstacle could be missed in the conventional approach.
- 3) In general, the conventional approach uses more computing resources than the present invention. For example, in the conventional approach, since the LADAR units are scanning horizontally, scan data must be placed into a large terrain elevation map. To detect obstacles, this terrain elevation map must be searched. This is a computationally expensive process.
- In one embodiment of the present invention, the
imaging sensor 8 constantly oscillates to produce the three-dimensional scan. Accordingly, there is no need for vehicle motion to map data. Nor is there any need in the present invention to store massive amounts of map or sensor data in order to detect obstacles. Furthermore, the present invention in one embodiment only stores data of the geospatial locations of all the currently-identified obstacles, such as for example an obstacle that is within a certain distance of the vehicle and that has been seen within a certain amount of time. For example, in the present invention, an obstacle is considered pertinent if it was within 50 meters of the vehicle and if it had been seen within the last 5 seconds. Other definitions of pertinent obstacles based on their distance and historical observance in the field of view are possible. - Other advantages of the present invention include the detection of moving obstacles and the elimination of blind spot hidden obstacles. Since the
imaging sensor 8 in one embodiment of the present invention rescans the same area multiple times per second, the chances of missing an obstacle are reduced as compared to a single scan over one area. As long as the area of oscillation covers a given area, any obstacle within that area should be seen. - In one embodiment of the present invention, processor 24 is configured to perform the illustrative steps shown in
FIG. 6A . Atstep 602, a sensing beam is repeatedly scanned into a two-dimensional sector of a plane defined with respect to a first predetermined axis of the vehicle. Atstep 604, a reflection of the emitted beam from objects removed from the vehicle is detected. Atstep 606, the plane in which the beam is scanned is panned about a second predetermined axis to produce a three dimensional field of view. Atstep 608, the existence and location of the objects in the three dimensional field of view based on a position of the vehicle and a time between an emittance of the beam and a reception of the reflection of the emitted beam is determined. - The method of the present invention in one embodiment emits the beam into substantially a vertical sector. Accordingly, slope data related to the terrain and depressions or obstacle protruding from the terrain are accessible from the scan in the vertical sector without having to process the full three dimensional field of view. In one embodiment of the present invention, the processor operating with a vertically oriented
sensor 8 does not need to process the full three dimensional field of view. The vertical scan lines (e.g., scans in α) originate from the vehicle and have slope data encoded in data due to the vertical orientation. Each independent scan line provides data of relevant obstacles. - The method of the present invention in one embodiment obtains a heading of the vehicle and a geospatial position of the vehicle, and processes data including the distance to the object, the direction to the object, the heading of the vehicle, and the geospatial position of the vehicle to determine whether the path for the vehicle intersects the object for example based on a predicted position of the vehicle. As discussed below, path planning and speed algorithms are used to direct the vehicle around the obstacles. The method of the present invention in one embodiment forms a three-dimensional field of view for the vehicle by scanning the beam in an angular sector of the plane while displacing the beam out of the plane, as shown in
FIG. 1B . - The method of the present invention in one embodiment determines an angular coordinate of the beam relative to the second predetermined axis for the direction to the object and offsets the angular coordinate by an amount indicative of a time delay in receiving a signal indicative of the angular coordinate at a processor (used to calculate the location of the object). This permits a more accurate determination of the location of the object.
- The method of the present invention in one embodiment directs the beam into respective angular sectors. The processor can determine the predetermined angular sectors based on a number of factors including but not limited to vehicle speed, an identified obstacle location, and a projected path of the vehicle. In the present invention, the sectors and/or the panning rate of each sector can be adjusted based upon some of the following non-limiting factors:
- 1) if an area of interest does not have adequate scan data (i.e. it has not been scanned enough times to be confident in that area), the sectors can be adjusted to give that area of interest more coverage especially if the area of interest includes a projected path of an obstacle (i.e. projected path of a moving obstacle) or if the area of interest includes an obstacle with complexity (i.e., if a composite obstacle seems to have many facets and/or other features warranting more scans), or if the area of interest includes an obstacle density (i.e., if one area has more obstacles than another, the sectors and/or speed to cover that area are appropriately adjusted);
- 2) if other sensor input such as a stereoscopic camera or radar sees an anomaly in an area or if an infrared camera picks up a high heat signature, for example of a human being, animal, or engine block, the sectors can be adjusted to provide better coverage in that area;
- 3) if a priori information is known, such as for example from a topographical map indicating that a cliff should be on the left side of the road, then the sectors could be adjusted to scan where the cliff should be to ensure safety;
- 4) if the vehicle was being operated in a convoy mode (either following a lead vehicle or possibly a walking human), the sectors could be adjusted to keep the lead vehicle or human in view;
- 5) if the vehicle was being operated as part of weapons targeting, the sectors could be adjusted to provide higher resolution coverage of the target area;
- 6) if information from other vehicles informs the autonomous vehicle about an area of interest, then the sectors could be adjusted to concentrate on the area of interest;
- 7) if a driver through an input device such a keyboard or mouse indicates an area is of interest, the sectors could be adjusted to concentrate on the area of interest;
- 8) if the autonomous vehicle receives a friend or foe signal, the sectors could be adjusted based upon whether the other vehicle is broadcasting a friend signal or a foe signal; and
- 9) if the autonomous vehicle expects a landmark to appear in a certain area, the sectors could be adjusted to look for the landmark in that area (landmark recognition is one way to enhance navigational accuracy).
- Accordingly, the method of the present invention in one embodiment determines the angular sectors (and/or the panning rate to scan with) based on at least one of a vehicle speed, an identified obstacle location, a projected path of the vehicle, a resolution required to resolve a complex obstacle or a collection of obstacles to be resolved, other sensory input, an identified priority sector in which an obstacle has been identified (e.g., a region in which high threat obstacles exist), and auxiliary information indicating the presence of an obstacle, a moving obstacle, another vehicle, a landmark, or an area (or region) of interest. When the
imaging sensor 8 is configured in a vertical orientation, as shown inFIG. 1A , it may not even be necessary to pan the beam as shown inFIG. 1B , in which case the panning rate could be set to zero with the obstacles being identified on the basis of a single scan in which the angle α is varied, as discussed above. - The method of the present invention in one embodiment stores in memory indicators of whether an obstacle exists in a subsector of the three dimensional field of view. The method of the present invention in one embodiment further determines if a projected path of the vehicle intersects one of the subsectors containing the indicators.
- One method of the present invention for obstacle identification and avoidance is illustrated by the flowchart in
FIG. 6B . At 620, the location of the object is determined in geospatial coordinates. In this step, the method determines angular coordinates (for the direction to the object) of the beam relative to the predetermined axis, obtains positioning data from a global positioning system device disposed on the vehicle, and/or obtains orientation data for the predetermined axis of the vehicle in a geospatial coordinate system. In this step, the method processes at least one of the angular coordinates, the positioning data, and the orientation data to determine the location of the object in geospatial coordinates. In one embodiment of the present invention, the data processed is limited to data for objects within a predetermined distance of the vehicle. - At 622, the altitude of the vehicle is ascertained, and from that information the present invention determines an elevation of the object by correlating the location of the object relative to the vehicle with the geospatial position of the vehicle and the altitude of the vehicle, determines elevations for plural points on the object, and/or generates a contour map for the object.
- At 624, the present invention determines if the object is an obstacle. For instance, as detailed by the vector equations (1)-(3) listed above, one embodiment of the present invention determines if the object is an obstacle by assessing if there exists a first point on a contour map (or in the field of view) that is farther away from the vehicle than a second point P on the contour map, if a slope between the first point and the second point is greater than a predetermined slope, and if a height between the first point and the second point is greater than a predetermined height.
- At 626, if the object is an obstacle, the present invention determines if the obstacle is to be avoided, for example by determining if a geospatial position of the obstacle remains in the path of the vehicle over a time lapse period and/or by determining if a geospatial position of the obstacle is within a set distance from the vehicle. In one embodiment of the present invention, the obstacle (now identified) is mapped relative to the path of the vehicle.
- At 628, if the object is an obstacle, steering and speed control directions are provided to the vehicle in order to avoid the obstacle. The steering and speed control directions can take into consideration the destination of the vehicle, and thereby provide steering and speed control directions to the vehicle in order to avoid the obstacle and to return on course to the destination.
- Or, at 630, if no obstacles are present, then the steering and speed control directions provide steering and speed control directions to the vehicle to guide the vehicle to the destination.
-
FIG. 7A is a schematic illustration of a vehicle according to the present invention depicting one embodiment in which multiple LADAR imaging sensors are used. In this embodiment, one or more of the imaging sensors is dedicated to scanning for the detection of objects nearby the vehicle (e.g., within 50 m) while another of the imaging sensors is dedicated to scanning for the detection of objects farther away from the vehicle (e.g., beyond 50 m). - In another embodiment of the invention, multiple imaging sensors are used for redundancy and to provide different perspectives of the same object.
FIG. 7B is a schematic illustration of a vehicle according to the present invention depicting one embodiment in which multiple LADAR imaging sensors are used to scan the same or overlapping fields of view. This configuration provides redundant coverage in the center of the path so that, if oneimaging sensor 8 fails, the other one can still sense obstacles most likely to be directly in the vehicle's path. The data from theimaging sensors 8 are correlated by placing all data onto the same elevation grid (as illustrated for example inFIG. 5 ). - In another embodiment, the
imaging sensors 8 are configured to locate objects removed from avehicle 10; and processor 24 is configured to direct one of the sensors to scan a first sector associated with a path of the vehicle, while directing another of the sensors to scan a second sector identified with an obstacle. As such, the first and/or second sector determinations can be based on a number of factors including, but not limited to a vehicle speed, an identified obstacle location, a projected path of the vehicle, a resolution required to resolve a complex obstacle or a collection of obstacles to be resolved, sensory input other than from the sensors, an identified priority sector in which an obstacle has been identified, and auxiliary information indicating the presence of an obstacle, a moving obstacle, another vehicle, a landmark, or an area of interest. - In one variant of this embodiment, the processor 24 can direct one sensor to scan a first sector associated with a path of the vehicle, and in a programmed manner direct the same sensor (e.g., in a dynamic fashion) to scan a second sector identified with an obstacle. Factors which determine the programmed duty cycle by which one sensor scans the first sector and then a second sector include for example the speed of the vehicle, the proximity of the obstacle, any movement of the obstacle, an identified status of the obstacle (e.g., friend or foe), the proximity of the obstacle to the projected path of the vehicle, and the calculated clearance from the vehicle to the obstacle.
- Moreover, in one embodiment of the present invention, one of the
imaging sensors 8 is dedicated to scanning in a horizontal direction while another imaging sensor is directed to scan in the vertical direction. Scan information from this unit permits the processor 24 to better identify the general terrain and terrain curvature from which obstacles such as shown inFIG. 5 can be identified. Complementary data from both horizontal and vertical scans helps identity the edges of composite obstacles (groups of individual obstacles that should be treated as one obstacle) more accurately. One of the issues with handling moving obstacles is determining the full proportions of an obstacle. To calculate the full proportions of an obstacle, multiple “independent” obstacles are intelligently grouped to form one larger composite obstacle when for example the data points representing the independent obstacles are within a set distance of each other (e.g., within 100 cm). Moreover, in other embodiments of the present invention, the grouping into composite obstacles is set by more than just a distance of separation between points normally qualifying as an obstacle point. Other factors that can be used in the determination include for example the number of times each point identified as an obstacle is seen, whether the obstacle point moves spatially in time, and whether (as discussed elsewhere) if there is confirmation of the obstacle by other image sensors or stereographic cameras. - Having two completely different perspectives of the obstacles facilitates this task by the obstacles being viewed from two separate dimensions (i.e., from top to bottom and from left to right). Since the beams tend to wrap around the curvature of an obstacle, this provides accurate estimations of the size and orientation of a composite obstacle. For instance, consider a spherical boulder. While the backside of the spherical boulder can not be seen, the sensing beam maps out a contour of the spherical boulder providing the aforementioned size and orientation, providing an estimate of the full size of the spherical boulder.
- Furthermore, having stepper motors (or in general any motor that can have its position dynamically adjusted) oscillate the platform rather than a continuously revolving motor creates several distinct advantages: 1) the rate at which the platform oscillates can be changed dynamically, and 2) the size of the sector which is scanned can be dynamically changed.
- By having these dynamic capabilities, the present invention achieves several advantages:
- 1) Path (or road) curvature can be tracked in the present invention. As the path in front of the vehicle curves, the
imaging sensors 8 can change their sector to better match the path, thereby ensuring that the path always gets the best coverage. This also helps to ensure that hard 90 degree or more turns do not create blind spots on the path. An imaging sensor that cannot alter its sector would be fixated ahead of the vehicle, even though the path is turning towards the right or left, and therefore blindspots in the viewing field would develop that would not resolved until after the vehicle had started turning. Accordingly, in one embodiment of the present invention, thesensors 8 can preemptively adjust their orientation. For example, theimaging sensor 8 shown inFIG. 1B or themultiple imaging sensors 8 shown inFIGS. 7A and 7B can be directed into a three-dimensional field of view that follows the anticipated path of thevehicle 10 on for example a highway. Further, the angular sectors mapped by themultiple imaging sensors 8 can be dynamically adjusted using stepper motors to vary (in real time) the angular sector β scanned and the angular scan rate. Factors which influence selection of the angular sector to be scanned include vehicle speed, an identified obstacle location relative to a turn in a path of the vehicle, a projected path of the vehicle, a resolution required to resolve a complex obstacle or a collection of obstacles to be resolved, sensory input other than from the sensors, an identified priority sector in which an obstacle has been identified, and auxiliary information indicating the presence of an obstacle, a moving obstacle, another vehicle, a landmark, or an area of interest. - 2) The speed of oscillation of the platform can be adjusted based on the speed of the vehicle. If the vehicle is going slower, the speed of oscillation can be reduced, thereby causing a higher resolution 3D image of the environment. As the vehicle speeds up, the speed of oscillation can increase to give better overall coverage of obstacles further down the road. Accordingly, in one embodiment of the present invention, the angular scan rates for α and β (as shown in
FIGS. 1A and 1B can be proportionate to the vehicle speed. Furthermore, in one embodiment of the present invention, as the vehicle moves faster, the scanned sectors could move farther out, or the speed of oscillation could change, or the angular sector could narrow (or all three could occur). The converse would apply when the vehicle slows down. - 3) Obstacles can be tracked in the present invention. Multiple imaging sensors can track obstacles and adjust their sectors to ensure that high threat obstacles are tracked (i.e., priority sectors or regions of interest can be established). For example, in one embodiment of the present invention, there are three imaging sensors on the front of the vehicle. If an obstacle is detected, two imaging sensors can keep analyzing the path, while the other sensor changes its sector to ensure that the obstacle can be scanned. Such capability permits the tracking of moving obstacles or other vehicles or for passing another vehicle at relatively high speed. Accordingly, in one embodiment of the present invention, respective ones of the
imaging sensors 8 can be dedicated to imaging and obstacle identification and tracking in predetermined sectors. - 4) Zone coverage is another advantage that can be realized in the present invention. The processor 24 could specify several zones around the vehicle with different levels of priority and multiple imaging sensors could dynamically adjust their sectors to make sure that all zones were covered based upon priority. For example, if the path directly in front of the vehicle were the highest priority and the sides of the vehicle were lower priority, then the sensors could be dynamically adjusted to always ensure that the front of the vehicle was adequately covered, and if additional sensors were available, then the sides would be covered too. This approach allows for easy scalability and fault tolerance. The processor 24 in one embodiment of the present invention could realize that one of the sensors had failed, and then dynamically adjust for the failed sensor to ensure that the proper zones were always covered. Adding sensors increases coverage areas and redundancy or fault tolerance. Accordingly, in one embodiment of the present invention, respective ones of the
imaging sensors 8 provide hardware redundancy and field of view redundancy. - Besides input from redundant imaging sensors, the processor 24 in one embodiment of the present invention receives input from on-board vehicle sensors. One of these inputs is a wheel speed sensor which provides transistor-transistor logic (TTL) pulses based upon an encoder placed on a single wheel on the vehicle. When a wheel speed sensor is added, the processor 24 uses GPS data from the positioning device 25 to initially calibrate the TTL pulses to actual vehicle movement. In the event that GPS data is not available, due to for example tunnels, canyons, or other obstructions, the processor 24 is able to minimize the positional drift by making use of the wheel speed sensor and its latest reliably known correspondence to the vehicle's movement.
- The wheel speed sensor included a digital sensor capable of detecting either ferrous metal or magnets that are in motion. The digital sensor was mounted in the wheel well adjacent to the stock Antilock Brake System (ABS) sensor, which allowed the wheel speed sensor to read the same magnets mounted on the wheel that the ABS sensor did. This level of accuracy permits the
vehicle 10 to precisely know its location on the earth. - Another input to processor 24 (as described later) is input from a biometric sensor or input from command button in which a driver to the vehicle can relinquish or take control of the
vehicle 10. -
FIG. 8 illustrates one embodiment of acomputer system 1201 in which the processor 24 (or any of the specific processors discussed below) of the present invention can be implemented. Thecomputer system 1201 is programmed and/or configured to perform any or all of the functions described above. Further, respective functions can be divided among different computers on board the vehicle (as shown for example in the computing network system ofFIG. 9 ). These computers may be in communication with each other via the communications network 1216 (discussed below). Thecomputer system 1201 includes abus 1202 or other communication mechanism for communicating information, and ainternal processor 1203 coupled with thebus 1202 for processing the information. Thecomputer system 1201 includes amemory 1204, such as a random access memory (RAM) or other dynamic storage device (e.g., dynamic RAM (DRAM), static RAM (SRAM), and synchronous DRAM (SDRAM)), coupled to thebus 1202 for storing information and instructions to be executed by theinternal processor 1203. In addition, thememory 1204 may be used for storing temporary variables or other intermediate information during the execution of instructions by theinternal processor 1203. Thecomputer system 1201 preferably includes a non-volatile memory such as for example a read only memory (ROM) 1205 or other static storage device (e.g., programmable ROM (PROM), erasable PROM (EPROM), and electrically erasable PROM (EEPROM)) coupled to thebus 1202 for storing static information and instructions for theinternal processor 1203. - The
computer system 1201 may also include special purpose logic devices (e.g., application specific integrated circuits (ASICs)) or configurable logic devices (e.g., simple programmable logic devices (SPLDs), complex programmable logic devices (CPLDs), and field programmable gate arrays (FPGAs)). The computer system may also include one or more digital signal processors (DSPs) such as the TMS320 series of chips from Texas Instruments, the DSP56000, DSP56100, DSP56300, DSP56600, and DSP96000 series of chips from Motorola, the DSP1600 and DSP3200 series from Lucent Technologies or the ADSP2100 and ADSP21000 series from Analog Devices. Other processors especially designed to process analog signals that have been converted to the digital domain may also be used (as detailed in the working example below). - The
computer system 1201 performs a portion or all of the processing steps of the invention in response to theinternal processor 1203 executing one or more sequences of one or more instructions contained in a memory, such as themain memory 1204. Such instructions may be read into themain memory 1204 from another computer readable medium, such as ahard disk 1207 or aremovable media drive 1208. Such instructions may be read into themain memory 1204 from another computer readable medium, such as a USB flash drives or jump drives. Such drives are solid-state memory devices which have the ability to act as floppy disks or hard drives under most computer operating systems. One or more processors in a multi-processing arrangement may also be employed to execute the sequences of instructions contained inmain memory 1204. In alternative embodiments, hard-wired circuitry may be used in place of or in combination with software instructions. Thus, embodiments are not limited to any specific combination of hardware circuitry and software. - As stated above, the
computer system 1201 includes at least one computer readable medium or memory for holding instructions programmed according to the teachings of the invention and for containing data structures, tables, records, or other data described herein. Examples of computer readable media suitable for the present invention are compact discs, hard disks, floppy disks, tape, magneto-optical disks, PROMs (EPROM, EEPROM, flash EPROM), DRAM, SRAM, SDRAM, or any other magnetic medium, compact discs (e.g., CD-ROM), or any other optical medium, punch cards, paper tape, or other physical medium with patterns of holes, a carrier wave (described below), or any other medium from which a computer can read. - Stored on any one or on a combination of computer readable media, the present invention includes software for controlling the
computer system 1201, for driving a device or devices for implementing the invention, and for enabling thecomputer system 1201 to interact with a human user (e.g., a driver). Such software may include, but is not limited to, device drivers, operating systems, development tools, and applications software. Such computer readable media further includes the computer program product of the present invention for performing all or a portion (if processing is distributed) of the processing performed in implementing the present invention. The computer code devices of the present invention may be any interpretable or executable code mechanism, including but not limited to scripts, interpretable programs, dynamic link libraries (DLLs), Java classes, and complete executable programs. Moreover, parts of the processing of the present invention may be distributed for better performance, reliability, and/or cost. - The term “computer readable medium” as used herein refers to any medium that participates in providing instructions to the
internal processor 1203 for execution. A computer readable medium may take many forms, including but not limited to, non-volatile media, volatile media, and transmission media. Non-volatile media includes, for example, optical, magnetic disks, and magneto-optical disks, such as thehard disk 1207 or the removable media drive 1208. Volatile media includes dynamic memory, such as themain memory 1204. Transmission media includes coaxial cables, copper wire and fiber optics, including the wires that make up thebus 1202. Transmission media also may also take the form of acoustic or light waves, such as those generated during radio wave and infrared data communications. - Various forms of computer readable media may be involved in carrying out one or more sequences of one or more instructions to
internal processor 1203 for execution. For example, the instructions may initially be carried on a disk to a remote computer. An infrared detector coupled to thebus 1202 can receive the data carried in the infrared signal and place the data on thebus 1202. Thebus 1202 carries the data to themain memory 1204, from which theinternal processor 1203 retrieves and executes the instructions. The instructions received by themain memory 1204 may optionally be stored onstorage device internal processor 1203. - For instance, in one embodiment of the present invention, a computer readable medium contains program instructions for execution on a processor in a vehicle, which when executed by the processor, cause the processor to scan repeatedly a beam into a two-dimensional sector of a plane defined with respect to a first predetermined axis of the vehicle, to detect a reflection of the emitted beam from an object removed from the vehicle, to pan the plane in which the beam is scanned about a second predetermined axis to produce a three dimensional field of view, and to determine the existence and location of the object in the three dimensional field of view based on a position of the vehicle and a time between an emittance of the beam and a reception of the reflection of the emitted beam.
- In one embodiment of the present invention, a computer readable medium contains program instructions for execution on a processor in a vehicle, which when executed by the processor, cause the processor to scan objects in a field of view from the vehicle, to determine a location of the objects in the field of view, to determine first and second points on a contour map of an identified object, to determine if the first point is farther away from the vehicle than the second point, to determine a slope between the first point and the second point, to determine if the slope is greater than a predetermined slope, and to determine if a height between the first point and the second point is greater than a predetermined height, and if so, to determine that the object is an obstacle.
- In one embodiment of the present invention, a computer readable medium contains program instructions for execution on a processor in a vehicle, which when executed by the processor, cause the processor to scan objects in a field of view from the vehicle, to determine a location of the objects in the field of view, to perform an obstacle identification algorithm to determine if any object is an obstacle, to determine a field of view subsector in which a determined obstacle exists, and to store in a corresponding memory element an obstacle indicator indicative that an obstacle exists in a corresponding subsector.
- In one embodiment of the present invention, a computer readable medium contains program instructions for execution on a processor in a vehicle, which when executed by the processor, cause the processor to scan objects in a field of view from the vehicle, to determine a location of the objects in the field of view, to direct a scan in a first sector associated with a path of the vehicle, and to direct a scan in a second sector identified with an obstacle.
- In one embodiment of the present invention, a computer readable medium contains program instructions for execution on a processor in a vehicle, which when executed by the processor, cause the processor to scan objects in a field of view from the vehicle, to determine a location of the objects in the field of view, to direct a scan selectively in a sector associated with a projected turn in a path of the vehicle.
- The program instructions, in various embodiments of the present invention, are configured to cause the processor to determine which of the sectors to scan based on at least one of a vehicle speed, an identified obstacle location, a projected path of the vehicle, a resolution required to resolve a complex obstacle or a collection of obstacles to be resolved, sensory input other than from the sensors, an identified priority sector in which an obstacle has been identified, and auxiliary information indicating the presence of an obstacle, a moving obstacle, another vehicle, a landmark, or an area of interest.
- Further, the computer readable medium of the present invention can include program instructions detailing geographical information associated with a particular locale, path planning algorithms (as described below), navigational instructions, instructions particular to an installed image sensor on the vehicle, instructions for command and/or receipt of data from additional sensors such a stereoscopic cameras, or vehicle wheel speed sensors, or receipt of data from driver input control devices or other on-board devices (such as those described later), path planning algorithms, a particularized vehicle transfer function containing data regarding vehicle thrust and response to external forces for the autonomous vehicle in use, and steering control for the autonomous vehicle in use.
- The
computer system 1201 also includes acommunication interface 1213 coupled to thebus 1202. Thecommunication interface 1213 provides a two-way data communication coupling to anetwork link 1214 that is connected at least temporarily to, for example, a local area network (LAN) 1215, or to anothercommunications network 1216 such as the Internet during downloading of software to the processor 24 or an internal network between multiple computers on board the vehicle. For example, thecommunication interface 1213 may be a network interface card to attach to any packet switched LAN. As another example, thecommunication interface 1213 may be an asymmetrical digital subscriber line (ADSL) card, an integrated services digital network (ISDN) card or a modem to provide a data communication connection to a corresponding type of communications line. Wireless links may also be implemented as part of thecommunication interface 1213 to provide data exchange with any of the on-board computers, image sensors, wheel speed sensors, biometric sensors, and/or driver command input devices. In any such implementation, thecommunication interface 1213 sends and receives electrical, electromagnetic or optical signals that carry digital data streams representing various types of information. - The
network link 1214 typically provides data communication through one or more networks to other data devices to provide data exchange with any of the on-board computers, image sensors, wheel speed sensors, biometric sensors, and/or driver command input devices. For example, thenetwork link 1214 may provide a temporary connection to another computer through a local network 1215 (e.g., a LAN) or through equipment operated by a service provider, which provides communication services through acommunications network 1216. As shown inFIG. 8 , thecomputing system 1201 can be in communication with aninput device 1217 via thelocal network 1215 and thecommunications network 1216 which use, for example, electrical, electromagnetic, or optical signals that carry digital data streams, and the associated physical layer (e.g., CAT 5 cable, coaxial cable, optical fiber, etc). The signals through the various networks and the signals on thenetwork link 1214 and through thecommunication interface 1213, which carry the digital data to and from thecomputer system 1201 may be implemented in baseband signals, or carrier wave based signals. The baseband signals convey the digital data as unmodulated electrical pulses that are descriptive of a stream of digital data bits, where the term “bits” is to be construed broadly to mean symbol, where each symbol conveys at least one or more information bits. The digital data may also be used to modulate a carrier wave, such as with amplitude, phase and/or frequency shift keyed signals that are propagated over a conductive media, or transmitted as electromagnetic waves through a propagation medium. Thus, the digital data may be sent as unmodulated baseband data through a “wired” communication channel and/or sent within a predetermined frequency band, different than baseband, by modulating a carrier wave. Thecomputer system 1201 can transmit and receive data, including program code, through the network(s) 1215 and 1216, thenetwork link 1214, and thecommunication interface 1213. Moreover, thenetwork link 1214 may provide a connection through aLAN 1215 to the various GPS and INS systems on board the vehicle. Theinput device 1217 in various embodiments of the present invention provides input to the processor 24 and represents schematically the image sensors, wheel speed sensors, biometric sensors, and/or driver command input devices discussed in the present invention. - A 2005 Ford Escape Hybrid™ (hereinafter referred to as the working vehicle) was modified to include the
imaging sensors 8 of the present invention. The working vehicle used a hybrid drive system in which an electric engine operates virtually all of the time and in which the gas engine starts and stops automatically to either provide extra horsepower or to recharge the electric engine's battery. The working vehicle's electrical system, which was powered by a 330-volt battery, provides over 1300 watts of power to the equipment mounted in the working vehicle. - The working vehicle utilized a commercially available Advanced Electronic Vehicle Interface Technology (AEVIT) “drive-by-wire” system from Electronic Mobility Controls (EMC) to physically control the car. The AEVIT system uses redundant servos and motors to turn the steering wheel, switch gears, apply throttle, and apply brake. This commercial system has a proven safety record in the automobile industry due to its use of redundant hardware.
- The working vehicle provided 110 amps of power at 12 volts. The
imaging sensor 8 utilized in the working vehicle operated on 24 volts. Rather than provide this power from the working vehicle's 12 volt electrical system, an auxiliary 24 volt electrical system for the imaging sensor was provided. The auxiliary electrical system included two 12 volt batteries connected in series to provide 24 volts of power. Solar panels were installed on top of the vehicle to charge the batteries in the auxiliary system. - In the working vehicle, a 2.8 Gigahertz Intel Pentium III™ processor including 2 gigabytes of RAM was used as the processor 24. The 2.8 Gigahertz Intel Pentium III™ processor operated on 12 volt power supply, so an AC to DC power inverter was not needed. The 2.8 Gigahertz Intel Pentium III™ processor hosted the principle computing functions, such as for example the sensor communication, vehicle controls, and artificial intelligence.
- Additionally, the working vehicle included several 1.42 Gigahertz Apple Mac Mini™ computers to host the path-planning software. The 1.42 Gigahertz Apple Mac Mini™ computers (running on running Apple OS X 10.4 software) were programmed to perform all of the path calculations in a redundant cluster to ensure that the path planning software was not suspect to a single point of failure. The Mac Mini™ computers had low power consumption (less than 2.3 amps at peak) and vector math processing capabilities. The processors listed above can be shock mounted on the vehicle to prevent damage from sudden changes in the direction of the vehicle or the crossing of obstacles along the path.
-
FIG. 9 shows a high level schematic of a computing network system of the present invention employing the computers and sensor data discussed above. In this networked system, the individual processors have similar capabilities and computer elements as described with respect toFIG. 8 .FIG. 9 shows that on-board electrical motor control (EMC) equipment is driven by instructions from the Navigational Computer. The Navigational Computer is connected by a switch (shown as a CISCO Switch) to a Sensor Processing Computer and a Path Planning Computer. These computers receive input from the RT3000™ GPS and INS positioning device (described below). The Path Planning Computer and the Navigational Computer utilize various path planning and speed control algorithms (described below) to provide steering instructions to the vehicle. The Sensor Processing Computer as shown in one illustrative embodiment of the present invention receives input from imaging sensor 8 (for example the LADAR devices shown) and encoders on the shaft of a LADAR device (for example the optical encoders shown). The LADAR devices provide distance data and the instantaneous angle α of the LADAR device in the predetermined angular scanning sector as shown inFIG. 1A . In one embodiment, optical encoders are used as therotary encoders 32 discussed above and provide the Sensor Processing Computer with the panning angle β relative to for example a vertical axis, as shown inFIG. 1B . Optical encoders are used in this embodiment of the present invention because of their high accuracy. Since optical encoders do not rely on physical contact with theshaft 34, but rather visually determine movement, optical encoders are more accurate and less prone to failure. The encoders from Fraba Posital were specifically used in the working vehicle because of their Ethernet capability. - The working vehicle utilized RT3000™ positioning devices from Oxford Technical Solutions to provide vehicle localization (i.e., positioning data). The RT3000™ uses Omnistar™ HP differential GPS signals to provide position accuracy of 10 centimeters or less. Integrated INS in the RT3000™ permits the RT3000™ to survive GPS outages of up to 30 seconds with virtually no performance degradation. Because the GPS and INS were integrated together, each can compensate for problems with the other. For example, if the INS started to drift laterally, the integrated GPS will automatically correct that drift. In the working vehicle, two Oxford RT3000™ GPS units were employed in a primary/secondary role.
- The working vehicle utilized commercially available laser modulation systems (LMS) for the
imaging sensors 8 such as three SICK™ LMS 291 LADAR units and one RIEGL™ LMS-Q 120 LADAR on the front of the vehicle. The LADAR units in the working vehicle were split into three different roles. The RIEGL™ LADAR was mounted vertically in the center of the vehicle and is used for long range (up to 80 meters) obstacle detection. The two SICK™ LMS 291 units were mounted vertically on either side of the RIEGL™. These vertically mounted units were used for short range obstacle detection (up to 50 meters). The other SICK™ LMS 291 was mounted horizontally on the center of the vehicle and is used for analyzing terrain features. - In order to protect these sensors from the elements as much as possible, a sun/weather guard was constructed over the sensor mounts. This guard protects the LADAR units from being blinded by direct sun glare, a common problem with Sick LADAR units. The guard can constitute a simple optical shield such as for example a piece of sheet-metal that was attached to the top of the imaging sensor. The guard extended over the front of LADAR unit (or other imaging sensor) and had a slight bend to better protect the sensors from the sun. Since LADAR units produce their own light source, low light conditions have no effect on the sensors. In fact, testing showed that low light conditions actually produced better results than bright conditions, due to excessive reflections under bright sunlight. These reflections, along with dust or rain, can produce ghosting patterns or other anomalies in the sensor readings. To counteract these anomalies, the present invention utilizes filtering algorithms to correct for anomalies before the anomalies can cause problems in the path planning software. For example, the filtering algorithm in one embodiment associates isolated obstacle points with noise data and discards these data points, under the assumption that “real” obstacles have a physical size greater than the resolution of the imaging sensor.
- On the working vehicle, stereoscopic cameras capable of producing a three-dimensional point cloud were mounted on the front of the vehicle behind the windshield. These cameras operated in a secondary mode to the LADAR units. The cameras are used to help identify the road and to confirm the existence of obstacles. If a LADAR unit identifies an obstacle and the stereoscopic cameras also detect an obstacle in that area, the obstacle avoidance algorithms increase the confidence level associated with that obstacle. Confidence levels were used to help decide which obstacles must absolutely be avoided and which obstacles are possibly driven over without damage.
-
FIG. 10 is a high level schematic of the vehicle control system of the present invention.FIG. 10 shows an iterative loop control system using inputs from the sensors and providing vehicle control. For the sake of illustration, at 102, LADAR data stored in a memory unit for example of the Sensor Processing Computer inFIG. 9 is provided toobstacle detection unit 104,speed planning unit 106, and accelerator/brake control 108. GPS/INS unit 110 also provided to theobstacle detection unit 104. Once theobstacle detection unit 104 decides which objects in the field of view are obstacles (for example using the conditions specified in eqns. (1), (2), and (3) above, obstacle identification data is forwarded to thespeed planning unit 106 and topath planning unit 112. Thepath planning unit 112 receives input from a route definition data file (RDDF)unit 114 and provides the path data tosteering control unit 116.EMC Vehicle Control 118 takes input from thesteering control unit 116 and the accelerator/brake control 108 to control the working vehicle. - The
obstacle detection unit 104 utilizes in one embodiment of the present invention spatial indices to correlate the sensor data with localization data indicating in which sector an obstacle is located (for example a near sector requiring immediate consideration or a far sector not requiring immediate consideration).FIG. 5 shows a division of the field of view into sectors: the sector defined between rays A and B, the sector defined between rays B and C, and the sector defined beyond C. Instead of storing all obstacles in one large spatial index, each segment of the path corridor has a corresponding spatial index in memory. Such division in one embodiment permits the present invention to call up from memory the sectors of importance for example when navigatingobstacles - The
obstacle detection unit 104 in one embodiment of the present invention uses times-stamps taken with each LADAR scan and GPS reading. These timestamps are then correlated based upon position and heading to produce a geospatial location for obstacles, for example the obstacles themselves may be moving. Timestamping can counteract anomalies in the images. Highly reflective surfaces can cause the LADAR devices to register incorrect distances for obstacles. To counteract these anomalies, in one embodiment of the present invention, scans are only kept in memory for a designated amount of time. After that time has passed, if no more obstacles have been registered in that same area, the obstacle is removed from memory storage. This technique ensures that moving obstacles are handled correctly. If another vehicle or an obstacle(s) crosses the vehicle path in a perpendicular direction to the autonomous vehicle, the sensors would effectively register a sequence of obstacles in front of the autonomous vehicle. These obstacles would appear as a complete obstruction of the vehicle's path, forcing the vehicle to immediately stop. After enough time had passed, the obstacles would expire, and the autonomous vehicle would be able to start moving again. A persistent obstacle, on the other hand, would not expire. - Consider, for example, a large boulder that is located in the center of the path. At a distance of approximately forty to fifty meters, the LADAR devices will start to register parts of the boulder. As
vehicle 10 gets to within ten to twenty meters of the vehicle, the previous scans would begin to approach the obstacle expiration time. Since the sensors are still registering the boulder, the previous scans will not expire. Instead, the previous scans along with the current scans would all be retained, giving the boulder a higher count for total number of scans. This high count for the number of scans causes the boulder to have an extremely high probability of being an obstacle. - The path-planning algorithms of the present invention avoid obstacles and routinely steer the vehicle within a planned corridor. In one embodiment, these algorithms without cause will not deviate the vehicle from the planned corridor. However, should the vehicle leave the route corridor for some reason, the navigation system will detect this and provide a safe route back into the corridor. If a waypoint is missed, the navigation system will simply continue to the next feasible waypoint on the path. If the path is obstructed by an obstacle, the path planning systems will determine a path around the obstacle.
- Path planning in one embodiment of the present invention is accomplished through the use of cubic b-splines designed to follow the center of the planned route while still ensuring that the path is not impossible for the vehicle to navigate. This assurance means that the curvature at any point along the path is below the maximum curvature that the vehicle can successfully follow. In addition, the curvature is kept continuous so that it is not necessary to stop the vehicle in order to turn the steering wheel to a new position before continuing. B-splines were chosen for use in the path planning algorithms primarily because of the ease in which the shape of their resulting curves can be controlled. After an initial path is created that follows the center of the corridor, the path is checked against the obstacle repository to determine if it is a safe path. If the path is not safe, a simple algorithm generates and adjusts control points on the problem spots of the curve until the spline avoids all known obstacles while still containing valid maximum curvature. At this point, the path is both safe and drivable.
- The path planning of the present invention can also uses a Level of Detail (LOD) based obstacle avoidance algorithm along with several planning algorithms to plan paths around obstacles. LOD analysis in one embodiment of the present invention permits running the same algorithm with different levels of detail. For example, running with less detail (e.g., to accommodate large safety margins), then iteratively increasing the detail (e.g., to accommodate smaller safety margins) until a valid path is found. The path planning algorithms run using several different parameters until a valid path is found. The initial parameters use safety margins (for example the clearance of the vehicle from an obstacle or between obstacles), while the final parameters use no safety margins around obstacles. This ensures that if a path is available that will avoid an obstacle with a large margin of error (e.g., vehicle lateral clearance) the path planning software selects that path. Otherwise, the planning algorithm will keep reducing the safety margin around obstacles until a valid path is determined.
-
FIGS. 11 , 12, and 13 illustrate results of path planning. InFIG. 11 , the obstacle detection unit identifies anobstacle 22 by the cluster of data points in the center of theroad 80. The path planning unit determines that there is adequate clearance to permit thevehicle 10 to deviate to the right as it advances to theobstacle 22 and then deviate left to return to the center of theroad 80. The projected path of thevehicle 10 is shown bypath 82. InFIG. 12 , thepath 82 of thevehicle 10 passes a center of theobstacle 22 to the left and then to the right as shown. The path planning unit has divided the path and the data map into sectors as shown. The first sector at the bottom contains no obstacles and requires no deviation along theplanned path 82. The second sector contains anobstacle 22 and requires the path planning unit to determine clearance and a path around theobstacle 22. Further, deviation from the path will rely on the speed planning unit to control the vehicle so as to safely pass the vehicle at a speed suited for the radius of the turn.FIG. 13 shows avehicle 10 that is about to navigate a series of obstacles in an immediate path of the vehicle. - The present invention recognizes that smooth paths alone are not necessarily drivable by the vehicle. The car should be able to slow down proportionally from straight-aways down to hairpin turns. Once a safe path is designed that avoids obstacles and yet remains drivable, the next step in the planning process is to decide on the speed at which the vehicle should take each section of the path. The speed chosen, in one embodiment of the present invention, is based on the curvature at that point in the path and upon the curvature further down the path. The speed is taken as the minimum of speed for the current curvature and the speed for future path curvature. The future path curvature is defined by a simple function that multiplies the curvature at a given future point by a fractional value that decreases towards zero linearly based upon the distance from the current path point to the given future path point.
- The present invention accommodates factors such as vehicle thrust and external forces on the vehicle. System identification is a method used by the present invention by which the parameters that define a system can be determined by relating input signal into a system with the system's response to develop a transfer function that behaves in the same way (or a similar way) as the actual vehicle system. For instance, when attempting to control the speed of a vehicle, the inputs are the brake and accelerator position and the output is the vehicle's speed. If it is assumed that the transfer function, H(s), is first-order, it can be written as
-
y(s)=H(s)u(s) (8) - where H(s) is the transfer function of a system, u(s) is the input to the system, and y(s) is the output from the system. System identification was applied to real world data from the vehicle propulsion system to arrive at a transfer function of the vehicle system for example empirically tested for the working vehicle until confidence in an accurate transfer function was obtained.
- Accordingly, speed control of the vehicle, according to the present invention, accommodated not only accelerator and brake functions but also accommodated many other factors in the physical engine system. For instance, since the working vehicle had a gas-electric hybrid engine, the coupling of the two propulsion systems was controlled by an inaccessible factory-installed on-board computer tuned for fuel efficiency. Consequently, the mapping of the requested pedal position and the actual position achieved was not linear and had to be remapped in software by empirical determination. In one embodiment of the present invention, the speed of the vehicle is controlled by an integrated proportional-derivative (PD) controller. This controller bases its output on the previous output and on the current error and derivative of the error. In the time domain, the controller can be written as
-
u(t2)=(t2−t1)(K p e(t 2)+K d e′(t 2))+u(t 1) (9) - where Kp and Kd are tunable coefficients, u(t) is the output of the controller at time t, and e(t) is the error at time t. The error was defined as actual output subtracted from target output. Actual output was reported by the RT3000™, and target speed was derived from the path planning algorithms.
- The integrated PD controller was designed and tuned against the derived transfer function detailed above. For instance, the weights (for the proportionate control in the PD controller) needed for optimal performance were derived initially against the computational model of the derived transfer function, and then tuned when operated on the vehicle.
- Accelerator and steering wheel control was achieved in the working vehicle using two separate processes, which were both independent of the path-planning systems. Once a path was decided on by the path-planning algorithms, acceleration and steering is used exclusively to remain on the chosen path. Since paths are checked for feasibility upon creation, it was assumed by the control systems that all paths given are possible for the vehicle to achieve. In this manner (although the present invention can use other starting assumptions), it becomes the burden of the control systems to decide how best to proceed in order to follow the selected path.
- The steering controller for the working vehicle was a lead-lag controller based on the classical single-track model or bicycle model described ed by Riekert and Schunck “Zur fahrmechanik des gummibereiften kraftfahrzeugs,” in Ingenieur Archiv, vol. 11, 1940, pp. 210-224, the entire contents of which are incorporated herein by reference. A lead compensator increases the responsiveness of a system; a lag compensator reduces (but does not eliminate) the steady state error. The lead-lag compensator was based on the frequency response of the system. The lead-lag compensator was similar to that described by D. Bernstein, A students guide to classical control, IEEE Control Systems Magazine, vol. 17, pp. 96-100 (1997), the entire contents of which are incorporated herein by reference. The resulting controller in the working vehicle was a convolution of the two lead and lag functions multiplied by the low frequency gain, which was 0.045. Adaptive estimation parameters were used. Adaptive estimation uses a set of values (parameters) first gained from applying the theoretical functions and then iteratively tests and modifies the parameters in real-world scenarios (e.g., deep sand, rough terrain, and other terrain types) until the parameters are perfected.
-
- The discretized controller was implemented as shown in (12) and (13), where x is the state vector, {dot over (x)} is the derivative of the state vector with respect to time, u is the input vector and δf is the output steering angle as measured at the tire with respect to the centerline. The state vector x is defined as [ys Ψ] where ys refers to the distance from the virtual sensor to the reference path and Ψ is the vehicle's yaw rate. The virtual sensor is a point projected a given distance ahead of the vehicle along the vehicle's centerline. This point is commonly referred to as the look-ahead point, and the distance from the look-ahead point to the RT3000 is referred to as the look-ahead distance.
-
δf=[0.02150 0.00005]x+[0.00005]u (13) - The input vector u to the controller is defined as [ys]. The output of the controller is the steering angle measured at the tire with respect to the centerline.
- The steering and control system in the working vehicle assumed that the relationship between the steering wheel angle and the resulting tire angle was linear and that the location of the vehicle's center of gravity was at the midway point between the front and rear axles. As a measure of safety the magnitude of the ys signal was monitored to prevent the vehicle from becoming unstable. If ys were to cross a given threshold, meaning the vehicle is severely off path, the speed of the vehicle was reduced to 2 mph. This allowed the vehicle to return onto the desired path and prevented a possible rollover.
- The present invention has widespread application in both autonomously driven and human driven vehicles where the present invention functions respectively in a primary or secondary mode.
- In general, the present invention with various of the attributes described above can be included on a variety of driveable units. Such units in one embodiment of the present invention are powered land or water based vehicles, which include for example automobiles, trucks, sport utility vehicles, armored vehicles, boats, ships, barges, tankers, and armored vessels. For watercraft, the present invention could be used not only when navigating in weather or nighttime conditions where visibility is limited and avoidance of other watercraft is desired, but also in docking and lock operations where control of the watercraft between dock and lock structures is important to minimize damage to the watercraft of the dock and lock structures.
- The present invention also has application to aircraft. In particular, the application to high speed aircraft will depend on detector sensitivity and accuracy to determine the existence of objects at a sufficient distance from the plane that the plane can take corrective action. However, airport approach and takeoff where the speeds are not high the present invention has utility. For example, on take off and landing, one concern has been whether flocks of birds frequently at the end of the run way area pose risk to engines. Birds are hard to see at distance for the human eye and hard to detect by aircraft radar. Moreover, the present invention has application to helicopters, especially rescue helicopters that frequently land in make-shift areas with obstacles many times hidden from the view of the pilot.
- Other application areas include lighter-than-air vehicles (e.g., autonomous weather balloons, autonomous border patrol system including for example remote control small aircraft), other small reconnaissance aircraft, and amphibious vehicles (such as for example land-sea assault vehicles including hovercrafts etc.).
- In one embodiment of the present invention, the driveable unit can be an autonomous vehicle without driver-assisted control or a driver-controlled vehicle with computer-assisted control. The autonomous vehicles find application, according to one embodiment of the present invention, in the above noted environmentally dangerous surroundings where a driver would be at risk. The driver-controlled vehicles of the present invention find application in the above noted more conventional environments, where the driver may become disoriented or incapable of physically commanding the vehicle as would occur if the driver suffered a medical emergency or if for example the driver became disoriented under an adverse driving condition. Accordingly, in one embodiment of the present invention, processor 24 is configured to control the vehicle in an event of driver impairment or in an event of immediate path obstruction or in an event of driver request.
- As an illustrative example of this embodiment of the present invention, the autonomous vehicle could recognize driver impairment through observing where the driver is driving in comparison to a predetermined path. If the present path of the vehicle and the predetermined path are not similar, the autonomous vehicle could then check, for example, to see if the driver has recently turned the steering wheel and/or pressed on the either the brakes or throttle. Both of the comparison and the steering and brake check can be included in the decision making process, because if the driver were driving on a long straight road with cruise control on, he may not be actively turning the steering wheel or applying brake or throttle. By the same logic, the driver could be driving a path that is not consistent with the autonomous vehicle's path as long as the driver is actively steering the car. In the event that the autonomous vehicle needs to take over, then the autonomous vehicle in one embodiment of the present invention first audibly and visibly warns the driver that it is taking over, then second takes over and steers the vehicle to a safe stopping position as smoothly and safely as possible. If the driver wanted to regain control, the autonomous vehicle of the present invention, in one embodiment, provides a press-button or other input device for the driver to resume control. In another embodiment of the present invention, the driver could press the button (or command input) again to relinquish control to the autonomous vehicle. Hence, the present invention provides in various embodiments a cooperative autonomous driving mode.
- In another embodiment of the present invention, a biometric sensor could represent another input device. In this embodiment, the biometric sensor determines if the driver was actively driving through input to processor 24 from for example a heart rate monitor built into the steering wheel of the vehicle. One example of a heart rate monitor that is suitable for the present invention is heart rate used in exercise equipment, which in one embodiment would be integrated into the steering wheel or alternatively could be worn by the driver and be in wireless communication to processor 24. If the processor 24 detected either a complete loss of heart rate or an extremely low heart rate for an extended period of time (for example 5 seconds), the processor 24 is configured to take control of the vehicle. The processor, by monitoring the normal heart rate of the driver when the car was under proper control would have a basis for determining for example if the driver was impaired due to haven fallen asleep at the wheel, having had a stroke, or having had a heart attack. This embodiment could be implemented also in the cooperative mode of operation (discussed above). As before, in one embodiment of the present invention, an audible alarm is issued before taking over control of the vehicle and steering the vehicle to a safe stop. If the driver was in fact not impaired, the driver could simply press a button (or other input device) to take control back from the processor 24.
- In another embodiment of the present invention, the autonomous vehicle can be operated repeatedly over a predetermined course. For instance, a human driver presses a button that turns the autonomous vehicle into record mode. The human drives the vehicle exactly like he would want the autonomous vehicle to drive the course. The human driver then presses the button again, and the autonomous vehicle drives the recorded course over and over again with a very high level of reliability and repeatability. (Repeatability is an issue for automobile testers). This capability is also useful for endurance testing vehicles, where the vehicle is driven offroad in hazardous conditions for days in a row in which currently many human drivers are used to perform this task due to the human body's relative weakness. This capability is also useful for driving a vehicle at consistent speeds for long distances. For example, this capability would be useful in testing a vehicle at highway speeds on a race track for fuel consumption performance.
- In another embodiment of the present invention, the driveable unit can be used, in conjunction with a mapping program (e.g., running on a laptop), in which the user could select a destination. At this point, the autonomous navigation software would access the mapping software and generate a route (like the software normally does) except in GPS waypoints rather than in human directions like “turn left at main street.” At this point, the autonomous vehicle proceeds in normal operation to follow that route. In one embodiment, the mapping software is customized to provide additional information to the autonomous navigation program such as the width of streets and speed limits.
- Other application areas for the navigation and control system of the present invention include, but are not limited to: 1) agricultural equipment (e.g., gardening equipment) that performs repetitive tasks or tasks on predicted paths such as for example the harvesting of crops from fields, plowing, grass cutting, etc., 2) mining equipment including for example powered carts that could transport for example goods or people through blackout or smoke filled passages that would normally prevent escape, 3) cave exploration equipment, 4) emergency or police vehicles such as for example fire fighting, ambulances, and rescue squads where visibility impairment need not prevent the emergency vehicles from proceeding forward (as discussed below) or vehicles operating in hazardous environmental conditions where manning of the vehicle places the driver at risk, 5) warehouse management equipment used to store/retrieve pallets, boxes, etc., and 6) toys.
- As an illustrative example of the application of the autonomous vehicle of the present invention to a police vehicle, on the Causeway bridge in New Orleans (the longest bridge in the world at 24 miles long), there is a significant fog season. On mornings with a dense fog, traffic is convoyed across the bridge by a police car driving at 35 mph. The low speed is required due to the extremely short visibility. On mornings with intense fog, the bridge is closed as even convoying is not possible. If the lead police car in the convoy were an autonomous vehicle of the present invention operating in the above-noted cooperative mode, the police car could safely convoy in any type of fog, especially when the RIEGL™ imaging sensor is used. The same applies to driving at night. The autonomous vehicle of the present invention is not affected by darkness.
- Numerous modifications and variations on the present invention are possible in light of the above teachings. It is, therefore, to be understood that within the scope of the accompanying claims, the invention may be practiced otherwise than as specifically described herein.
Claims (8)
1. A land-based drivable unit comprising:
a vehicle for ground transportation including,
a sensor configured to locate objects in a predetermined field of view from a vehicle, including,
an emitter configured to scan a beam into a sector of a plane defined with respect to a first predetermined axis of the vehicle,
a detector configured to detect a reflection of the emitted beam from one of the objects, and
a panning mechanism configured to pan the plane in which the beam is scanned about a second predetermined axis to produce a three dimensional field of view, and
a processor configured to determine the existence and location of the objects in the three dimensional field of view based on the three dimensional field of view.
2. The unit of claim 1 , wherein the processor determines the existence and location of the objects based on a position of the vehicle and a time between an emittance of the beam and a reception of the reflection of the emitted beam.
3. The unit of claim 1 , wherein the vehicle comprises at least one of an automobile, a truck, a sport utility vehicle, rescue vehicle, an agricultural vehicle, a mining vehicle, an escort vehicle, a toy vehicle, a reconnaissance vehicle, a test-track vehicle, and an armored vehicle.
4. The unit of claim 1 , wherein the vehicle comprises an autonomous vehicle without driver-assisted control.
5. The unit of claim 1 , wherein the vehicle comprises a driver-controlled vehicle with computer-assisted control.
6. The unit of claim 5 , wherein the processor is configured to recognize driver impairment.
7. The unit of claim 6 , wherein the processor is configured to recognize driver impairment from a biometric sensor or from an analysis of driver-control of the vehicle.
8. The unit of claim 5 , wherein the processor is configured to control the vehicle in an event of an input command indicating driver impairment.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/715,586 US20130131908A1 (en) | 2006-03-16 | 2012-12-14 | Navigation and control system for autonomous vehicles |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/376,160 US8050863B2 (en) | 2006-03-16 | 2006-03-16 | Navigation and control system for autonomous vehicles |
US13/240,576 US8346480B2 (en) | 2006-03-16 | 2011-09-22 | Navigation and control system for autonomous vehicles |
US13/715,586 US20130131908A1 (en) | 2006-03-16 | 2012-12-14 | Navigation and control system for autonomous vehicles |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/240,576 Continuation US8346480B2 (en) | 2006-03-16 | 2011-09-22 | Navigation and control system for autonomous vehicles |
Publications (1)
Publication Number | Publication Date |
---|---|
US20130131908A1 true US20130131908A1 (en) | 2013-05-23 |
Family
ID=38518970
Family Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/376,160 Active 2027-05-09 US8050863B2 (en) | 2006-03-16 | 2006-03-16 | Navigation and control system for autonomous vehicles |
US13/240,576 Active US8346480B2 (en) | 2006-03-16 | 2011-09-22 | Navigation and control system for autonomous vehicles |
US13/715,586 Abandoned US20130131908A1 (en) | 2006-03-16 | 2012-12-14 | Navigation and control system for autonomous vehicles |
Family Applications Before (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/376,160 Active 2027-05-09 US8050863B2 (en) | 2006-03-16 | 2006-03-16 | Navigation and control system for autonomous vehicles |
US13/240,576 Active US8346480B2 (en) | 2006-03-16 | 2011-09-22 | Navigation and control system for autonomous vehicles |
Country Status (2)
Country | Link |
---|---|
US (3) | US8050863B2 (en) |
WO (1) | WO2008048707A2 (en) |
Cited By (39)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110093177A1 (en) * | 2009-10-21 | 2011-04-21 | Horn Berthold K P | Method and Apparatus for Reducing Motor Vehicle Traffic Flow Instabilities and Increasing Vehicle Throughput |
US20120265380A1 (en) * | 2011-04-13 | 2012-10-18 | California Institute Of Technology | Target Trailing with Safe Navigation with colregs for Maritime Autonomous Surface Vehicles |
US8706394B2 (en) | 2008-10-24 | 2014-04-22 | Gray & Company, Inc. | Control and systems for autonomously driven vehicles |
US20140236470A1 (en) * | 2012-04-06 | 2014-08-21 | Thales | Method for determining an instantaneous or anticipated probable zone of occupancy of an aircraft in an airport navigation zone |
US20150088373A1 (en) * | 2013-09-23 | 2015-03-26 | The Boeing Company | Optical communications and obstacle sensing for autonomous vehicles |
US9043072B1 (en) * | 2013-04-04 | 2015-05-26 | Google Inc. | Methods and systems for correcting an estimated heading using a map |
US9043069B1 (en) * | 2012-11-07 | 2015-05-26 | Google Inc. | Methods and systems for scan matching approaches for vehicle heading estimation |
US9259838B1 (en) * | 2014-07-24 | 2016-02-16 | Google Inc. | Systems and methods for ground plane estimation |
WO2016126316A1 (en) * | 2015-02-06 | 2016-08-11 | Delphi Technologies, Inc. | Autonomous guidance system |
US20170110022A1 (en) * | 2015-10-14 | 2017-04-20 | Toyota Motor Engineering & Manufacturing North America, Inc. | Assessing driver readiness for transition between operational modes of an autonomous vehicle |
CN106774312A (en) * | 2016-12-05 | 2017-05-31 | 遨博(北京)智能科技有限公司 | A kind of method and apparatus that motion track is determined for mobile robot |
WO2017095493A3 (en) * | 2015-09-11 | 2017-07-20 | The Trustees Of The University Of Pennsylvania | Systems and methods for generating safe trajectories for multi-vehicle teams |
US9720415B2 (en) * | 2015-11-04 | 2017-08-01 | Zoox, Inc. | Sensor-based object-detection optimization for autonomous vehicles |
US9784835B1 (en) * | 2013-09-27 | 2017-10-10 | Waymo Llc | Laser diode timing feedback using trace loop |
CN107578427A (en) * | 2017-07-31 | 2018-01-12 | 深圳市易成自动驾驶技术有限公司 | Detection method, device and the computer-readable recording medium of dynamic barrier |
WO2018134296A1 (en) * | 2017-01-19 | 2018-07-26 | Robert Bosch Gmbh | System and method of using crowd-sourced driving path data in an autonmous or semi-autonomous driving system |
US10037028B2 (en) | 2015-07-24 | 2018-07-31 | The Trustees Of The University Of Pennsylvania | Systems, devices, and methods for on-board sensing and control of micro aerial vehicles |
WO2018224874A1 (en) * | 2017-06-06 | 2018-12-13 | PlusAI Corp | Method and system for distributed learning and adaptation in autonomous driving vehicles |
US20190041859A1 (en) * | 2017-08-04 | 2019-02-07 | Aptiv Technologies Limited | Sensor failure compensation system for an automated vehicle |
US20190072957A1 (en) * | 2016-05-19 | 2019-03-07 | Denso Corporation | Automated driving system and automated driving switch determination program product |
WO2019084130A1 (en) * | 2017-10-25 | 2019-05-02 | Bae Systems Information And Electronic Systems Integration Inc. | Method and system of digital light processing and light detection and ranging for guided autonomous vehicles |
US10338594B2 (en) | 2017-03-13 | 2019-07-02 | Nio Usa, Inc. | Navigation of autonomous vehicles to enhance safety under one or more fault conditions |
US10369974B2 (en) | 2017-07-14 | 2019-08-06 | Nio Usa, Inc. | Control and coordination of driverless fuel replenishment for autonomous vehicles |
US10379540B2 (en) | 2016-10-17 | 2019-08-13 | Waymo Llc | Light detection and ranging (LIDAR) device having multiple receivers |
US10395115B2 (en) | 2015-01-27 | 2019-08-27 | The Trustees Of The University Of Pennsylvania | Systems, devices, and methods for robotic remote sensing for precision agriculture |
US10423162B2 (en) | 2017-05-08 | 2019-09-24 | Nio Usa, Inc. | Autonomous vehicle logic to identify permissioned parking relative to multiple classes of restricted parking |
CN110320919A (en) * | 2019-07-31 | 2019-10-11 | 河海大学常州校区 | A kind of circulating robot method for optimizing route in unknown geographical environment |
US20190353783A1 (en) * | 2018-05-18 | 2019-11-21 | Koito Manufacturing Co., Ltd. | Sensor system |
US10496890B2 (en) | 2016-10-28 | 2019-12-03 | International Business Machines Corporation | Vehicular collaboration for vehicular blind spot detection |
US20200208993A1 (en) * | 2017-07-31 | 2020-07-02 | Israel Aerospace Industries Ltd. | Path planning within a traversed area |
US10710633B2 (en) | 2017-07-14 | 2020-07-14 | Nio Usa, Inc. | Control of complex parking maneuvers and autonomous fuel replenishment of driverless vehicles |
US10732647B2 (en) | 2013-11-27 | 2020-08-04 | The Trustees Of The University Of Pennsylvania | Multi-sensor fusion for robust autonomous flight in indoor and outdoor environments with a rotorcraft micro-aerial vehicle (MAV) |
US10948924B2 (en) | 2015-02-06 | 2021-03-16 | Aptiv Technologies Limited | Method and apparatus for controlling an autonomous vehicle |
US10991247B2 (en) | 2015-02-06 | 2021-04-27 | Aptiv Technologies Limited | Method of automatically controlling an autonomous vehicle based on electronic messages from roadside infrastructure or other vehicles |
US11022971B2 (en) | 2018-01-16 | 2021-06-01 | Nio Usa, Inc. | Event data recordation to identify and resolve anomalies associated with control of driverless vehicles |
US11042155B2 (en) | 2017-06-06 | 2021-06-22 | Plusai Limited | Method and system for closed loop perception in autonomous driving vehicles |
US11392133B2 (en) | 2017-06-06 | 2022-07-19 | Plusai, Inc. | Method and system for object centric stereo in autonomous driving vehicles |
US20240061119A1 (en) * | 2022-08-16 | 2024-02-22 | Qualcomm Incorporated | Cellular radio frequency (rf) sensing in automobile navigation |
EP4300237A4 (en) * | 2021-02-23 | 2024-12-18 | Jingdong Technology Information Technology Co., Ltd. | Robot traveling method and apparatus, and electronic device, storage medium and program product |
Families Citing this family (660)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7983802B2 (en) * | 1997-10-22 | 2011-07-19 | Intelligent Technologies International, Inc. | Vehicular environment scanning techniques |
GB0520829D0 (en) * | 2005-10-13 | 2005-11-23 | Univ Cambridge Tech | Image processing methods and apparatus |
EP1949188B1 (en) | 2005-10-28 | 2015-04-22 | Mojix, Inc. | Rfid receiver |
US8381982B2 (en) * | 2005-12-03 | 2013-02-26 | Sky-Trax, Inc. | Method and apparatus for managing and controlling manned and automated utility vehicles |
US10878646B2 (en) * | 2005-12-08 | 2020-12-29 | Smartdrive Systems, Inc. | Vehicle event recorder systems |
US20070150138A1 (en) | 2005-12-08 | 2007-06-28 | James Plante | Memory management in event recording systems |
US8996240B2 (en) * | 2006-03-16 | 2015-03-31 | Smartdrive Systems, Inc. | Vehicle event recorders with integrated web server |
US9201842B2 (en) | 2006-03-16 | 2015-12-01 | Smartdrive Systems, Inc. | Vehicle event recorder systems and networks having integrated cellular wireless communications systems |
US8868256B2 (en) * | 2006-05-15 | 2014-10-21 | Honeywell International Inc. | Relative navigation for aerial refueling of an unmanned aerial vehicle |
US20070288156A1 (en) * | 2006-05-17 | 2007-12-13 | The Boeing Company | Route search planner |
DE102006031580A1 (en) | 2006-07-03 | 2008-01-17 | Faro Technologies, Inc., Lake Mary | Method and device for the three-dimensional detection of a spatial area |
CN101688774A (en) * | 2006-07-13 | 2010-03-31 | 威力登音响公司 | High definition lidar system |
USRE46672E1 (en) | 2006-07-13 | 2018-01-16 | Velodyne Lidar, Inc. | High definition LiDAR system |
US20080033645A1 (en) * | 2006-08-03 | 2008-02-07 | Jesse Sol Levinson | Pobabilistic methods for mapping and localization in arbitrary outdoor environments |
DE602006005079D1 (en) * | 2006-08-24 | 2009-03-19 | Harman Becker Automotive Sys | A method of mapping the environment of a vehicle and system therefor |
EP2084555A2 (en) * | 2006-10-31 | 2009-08-05 | Valeo Raytheon Systems, Inc. | System and method for generating an alert signal in a detection system |
DE102006051961A1 (en) * | 2006-11-03 | 2008-05-15 | Robert Bosch Gmbh | Device for adjusting the distance of a motor vehicle |
US8989959B2 (en) | 2006-11-07 | 2015-03-24 | Smartdrive Systems, Inc. | Vehicle operator performance history recording, scoring and reporting systems |
US8649933B2 (en) | 2006-11-07 | 2014-02-11 | Smartdrive Systems Inc. | Power management systems for automotive video event recorders |
US8868288B2 (en) | 2006-11-09 | 2014-10-21 | Smartdrive Systems, Inc. | Vehicle exception event management systems |
JP4858197B2 (en) * | 2007-01-31 | 2012-01-18 | ソニー株式会社 | Information processing apparatus, image display apparatus, information processing system, information processing method, and program |
US7739006B2 (en) * | 2007-02-07 | 2010-06-15 | Disney Enterprises, Inc. | System and method for autonomous navigation in a ride vehicle |
US8554478B2 (en) * | 2007-02-23 | 2013-10-08 | Honeywell International Inc. | Correlation position determination |
JP5134020B2 (en) | 2007-03-23 | 2013-01-30 | モジクス, インコーポレイティッド | RFID system using distributed exciter network |
US7392696B1 (en) * | 2007-04-10 | 2008-07-01 | Toyota Motor Engineering & Manufacturing North America, Inc. | Sensorless tire pressure estimating method for vehicle with ABS brake control |
US8239092B2 (en) | 2007-05-08 | 2012-08-07 | Smartdrive Systems Inc. | Distributed vehicle event recorder systems having a portable memory data transfer system |
US20090018770A1 (en) * | 2007-07-12 | 2009-01-15 | Yahoo! Inc. | Mobile notification system |
JP2009025898A (en) * | 2007-07-17 | 2009-02-05 | Toyota Motor Corp | Route planning apparatus, route planning method, and moving body |
US7852463B2 (en) * | 2007-08-13 | 2010-12-14 | Honeywell International Inc. | Range measurement device |
US20090082879A1 (en) * | 2007-09-20 | 2009-03-26 | Evolution Robotics | Transferable intelligent control device |
US8165806B2 (en) * | 2007-09-28 | 2012-04-24 | General Motors Llc | Vehicle compass using telematics unit and vehicle sensor information |
US8174562B2 (en) * | 2007-11-09 | 2012-05-08 | Honeywell International Inc. | Stereo camera having 360 degree field of view |
US8210435B2 (en) * | 2008-01-14 | 2012-07-03 | Sky-Trax, Inc. | Optical position marker apparatus |
US8565913B2 (en) | 2008-02-01 | 2013-10-22 | Sky-Trax, Inc. | Apparatus and method for asset tracking |
US8915692B2 (en) | 2008-02-21 | 2014-12-23 | Harvest Automation, Inc. | Adaptable container handling system |
CN101532844A (en) * | 2008-03-14 | 2009-09-16 | 鸿富锦精密工业(深圳)有限公司 | Mobile communicator and method for guiding stroke of mobile communicator |
JP4930446B2 (en) * | 2008-04-14 | 2012-05-16 | トヨタ自動車株式会社 | Vehicle travel control device |
US8072311B2 (en) | 2008-04-14 | 2011-12-06 | Mojix, Inc. | Radio frequency identification tag location estimation and tracking system and method |
US8060280B2 (en) * | 2008-04-15 | 2011-11-15 | Autoliv Asp, Inc. | Vision system for deploying safety systems |
US8744741B2 (en) * | 2008-04-18 | 2014-06-03 | Bae Systems Plc | Lidars |
EP2439493A1 (en) * | 2008-05-09 | 2012-04-11 | Research in Motion Limited | Predictive downloading of map data |
US8346468B2 (en) * | 2008-07-08 | 2013-01-01 | Sky-Trax Incorporated | Method and apparatus for collision avoidance |
JP5200732B2 (en) * | 2008-07-29 | 2013-06-05 | 日産自動車株式会社 | Travel control device and travel control method |
US8755997B2 (en) * | 2008-07-30 | 2014-06-17 | Honeywell International Inc. | Laser ranging process for road and obstacle detection in navigating an autonomous vehicle |
JP5345350B2 (en) * | 2008-07-30 | 2013-11-20 | 富士重工業株式会社 | Vehicle driving support device |
US8930058B1 (en) * | 2008-10-20 | 2015-01-06 | The United States Of America As Represented By The Secretary Of The Navy | System and method for controlling a vehicle traveling along a path |
US8364334B2 (en) * | 2008-10-30 | 2013-01-29 | Honeywell International Inc. | System and method for navigating an autonomous vehicle using laser detection and ranging |
US9522817B2 (en) * | 2008-12-04 | 2016-12-20 | Crown Equipment Corporation | Sensor configuration for a materials handling vehicle |
WO2010123529A2 (en) * | 2008-12-19 | 2010-10-28 | Xollai, Llc | System and method for autonomous vehicle control |
TW201025217A (en) * | 2008-12-30 | 2010-07-01 | Ind Tech Res Inst | System and method for estimating state of carrier |
WO2010085940A1 (en) * | 2009-01-27 | 2010-08-05 | Micro-Epsilon Optronic Gmbh | Device and method for non-contacting measurement of a distance and/or profile |
DE102009010465B3 (en) | 2009-02-13 | 2010-05-27 | Faro Technologies, Inc., Lake Mary | laser scanner |
CA2753042C (en) * | 2009-02-23 | 2017-10-24 | 2G Robotics Inc. | Laser scanner assembly |
JP4614005B2 (en) * | 2009-02-27 | 2011-01-19 | トヨタ自動車株式会社 | Moving locus generator |
US9551575B2 (en) | 2009-03-25 | 2017-01-24 | Faro Technologies, Inc. | Laser scanner having a multi-color light source and real-time color receiver |
DE102009015920B4 (en) | 2009-03-25 | 2014-11-20 | Faro Technologies, Inc. | Device for optically scanning and measuring an environment |
TWI383127B (en) * | 2009-06-12 | 2013-01-21 | Inventec Appliances Corp | Method for velocity estimation in gps dead reckoning |
DE102009035337A1 (en) | 2009-07-22 | 2011-01-27 | Faro Technologies, Inc., Lake Mary | Method for optically scanning and measuring an object |
US8731777B2 (en) | 2009-08-18 | 2014-05-20 | Crown Equipment Corporation | Object tracking and steer maneuvers for materials handling vehicles |
KR101314588B1 (en) * | 2009-10-26 | 2013-10-07 | 한국전자통신연구원 | Method and apparatus for producing map of artificial mark, method and apparatus for measuring position of mobile object by using same |
DE102009055989B4 (en) | 2009-11-20 | 2017-02-16 | Faro Technologies, Inc. | Device for optically scanning and measuring an environment |
DE102009057101A1 (en) | 2009-11-20 | 2011-05-26 | Faro Technologies, Inc., Lake Mary | Device for optically scanning and measuring an environment |
DE102009055988B3 (en) | 2009-11-20 | 2011-03-17 | Faro Technologies, Inc., Lake Mary | Device, particularly laser scanner, for optical scanning and measuring surrounding area, has light transmitter that transmits transmission light ray by rotor mirror |
US9210288B2 (en) | 2009-11-20 | 2015-12-08 | Faro Technologies, Inc. | Three-dimensional scanner with dichroic beam splitters to capture a variety of signals |
US9529083B2 (en) | 2009-11-20 | 2016-12-27 | Faro Technologies, Inc. | Three-dimensional scanner with enhanced spectroscopic energy detector |
US9113023B2 (en) | 2009-11-20 | 2015-08-18 | Faro Technologies, Inc. | Three-dimensional scanner with spectroscopic energy detector |
US9848114B2 (en) * | 2009-12-07 | 2017-12-19 | Cobra Electronics Corporation | Vehicle camera system |
US8949069B2 (en) * | 2009-12-16 | 2015-02-03 | Intel Corporation | Position determination based on propagation delay differences of multiple signals received at multiple sensors |
US20110153169A1 (en) * | 2009-12-18 | 2011-06-23 | Agco Corporation | Sensor-Based Implement Motion Interlock System |
JP5421467B2 (en) | 2010-01-20 | 2014-02-19 | ファロ テクノロジーズ インコーポレーテッド | Mounting device for coordinate measuring machine |
US9163922B2 (en) | 2010-01-20 | 2015-10-20 | Faro Technologies, Inc. | Coordinate measurement machine with distance meter and camera to determine dimensions within camera images |
US9879976B2 (en) | 2010-01-20 | 2018-01-30 | Faro Technologies, Inc. | Articulated arm coordinate measurement machine that uses a 2D camera to determine 3D coordinates of smoothly continuous edge features |
US9607239B2 (en) | 2010-01-20 | 2017-03-28 | Faro Technologies, Inc. | Articulated arm coordinate measurement machine having a 2D camera and method of obtaining 3D representations |
US9628775B2 (en) | 2010-01-20 | 2017-04-18 | Faro Technologies, Inc. | Articulated arm coordinate measurement machine having a 2D camera and method of obtaining 3D representations |
US8508590B2 (en) * | 2010-03-02 | 2013-08-13 | Crown Equipment Limited | Method and apparatus for simulating a physical environment to facilitate vehicle operation and task completion |
US8538577B2 (en) * | 2010-03-05 | 2013-09-17 | Crown Equipment Limited | Method and apparatus for sensing object load engagement, transportation and disengagement by automated vehicles |
DE102010010875A1 (en) * | 2010-03-10 | 2011-09-15 | Siemens Aktiengesellschaft | Method for monitoring the spatial environment of a movable device, in particular a medical device |
CN102822881B (en) * | 2010-04-05 | 2015-04-29 | 丰田自动车株式会社 | Vehicle collision judgment device |
US20120232717A1 (en) * | 2010-04-06 | 2012-09-13 | Honeywell International Inc. | Remote coordinate identifier system and method for aircraft |
WO2011130582A1 (en) | 2010-04-14 | 2011-10-20 | Mojix, Inc. | Systems and methods for detecting patterns in spatio-temporal data collected using an rfid system |
US8346426B1 (en) | 2010-04-28 | 2013-01-01 | Google Inc. | User interface for displaying internal state of autonomous driving system |
US8260482B1 (en) | 2010-04-28 | 2012-09-04 | Google Inc. | User interface for displaying internal state of autonomous driving system |
DE102010020925B4 (en) | 2010-05-10 | 2014-02-27 | Faro Technologies, Inc. | Method for optically scanning and measuring an environment |
US9639688B2 (en) | 2010-05-27 | 2017-05-02 | Ford Global Technologies, Llc | Methods and systems for implementing and enforcing security and resource policies for a vehicle |
JP5560978B2 (en) * | 2010-07-13 | 2014-07-30 | 村田機械株式会社 | Autonomous mobile |
EP2410294A1 (en) * | 2010-07-21 | 2012-01-25 | Harman Becker Automotive Systems GmbH | Method and device for providing cost information associated with junctions and method of determining a route |
DE102010032725B4 (en) | 2010-07-26 | 2012-04-26 | Faro Technologies, Inc. | Device for optically scanning and measuring an environment |
DE102010032726B3 (en) | 2010-07-26 | 2011-11-24 | Faro Technologies, Inc. | Device for optically scanning and measuring an environment |
DE102010032723B3 (en) | 2010-07-26 | 2011-11-24 | Faro Technologies, Inc. | Device for optically scanning and measuring an environment |
DE102010046915A1 (en) * | 2010-09-29 | 2012-03-29 | Gm Global Technology Operations Llc (N.D.Ges.D. Staates Delaware) | Motor vehicle with warning system |
US8509982B2 (en) | 2010-10-05 | 2013-08-13 | Google Inc. | Zone driving |
JP2012083267A (en) * | 2010-10-13 | 2012-04-26 | Japan Aerospace Exploration Agency | Multi-lidar system |
US9168654B2 (en) | 2010-11-16 | 2015-10-27 | Faro Technologies, Inc. | Coordinate measuring machines with dual layer arm |
US11480965B2 (en) | 2010-11-19 | 2022-10-25 | Maid Ip Holdings Pty/Ltd | Automatic location placement system |
US9778657B2 (en) | 2010-11-19 | 2017-10-03 | Bradley Tyers | Automatic location placement system |
TWI430902B (en) * | 2010-12-15 | 2014-03-21 | Wistron Neweb Corp | Wireless signal transceiver and blind spot detection system |
US9199616B2 (en) | 2010-12-20 | 2015-12-01 | Caterpillar Inc. | System and method for determining a ground speed of a machine |
DE102011010334B4 (en) * | 2011-02-04 | 2014-08-28 | Eads Deutschland Gmbh | Camera system and method for observing objects at a great distance, in particular for monitoring target objects at night, mist, dust or rain |
US9452735B2 (en) | 2011-02-10 | 2016-09-27 | Ford Global Technologies, Llc | System and method for controlling a restricted mode in a vehicle |
US9602316B2 (en) | 2011-03-07 | 2017-03-21 | Mojix, Inc. | Multiple symbol noncoherent soft output detector |
KR101061066B1 (en) | 2011-03-24 | 2011-08-31 | 국방과학연구소 | Terrain sensor assembly and autonomous mobile vehicle |
US20120253583A1 (en) * | 2011-04-01 | 2012-10-04 | David Kevin Herdle | Imaging-based proximity detection systems for mining machines |
US9650893B2 (en) | 2011-04-01 | 2017-05-16 | Joy Mm Delaware, Inc. | Imaging-based interface sensor and control device for mining machines |
US8522320B2 (en) | 2011-04-01 | 2013-08-27 | Ford Global Technologies, Llc | Methods and systems for authenticating one or more users of a vehicle communications and information system |
EP3435189B1 (en) | 2011-04-11 | 2022-02-09 | Crown Equipment Corporation | Apparatus for efficient scheduling for multiple automated non-holonomic vehicles using a coordinated path planner |
US20120268260A1 (en) * | 2011-04-21 | 2012-10-25 | Ford Global Technologies, Llc | Method and apparatus for dynamically providing space management alerts for a vehicle |
TW201248347A (en) * | 2011-05-18 | 2012-12-01 | Hon Hai Prec Ind Co Ltd | System and method for controlling unmanned aerial vehicle |
WO2012160373A2 (en) | 2011-05-24 | 2012-11-29 | Bae Systems Plc | Vehicle navigation |
US8655588B2 (en) | 2011-05-26 | 2014-02-18 | Crown Equipment Limited | Method and apparatus for providing accurate localization for an industrial vehicle |
DE112011105204T5 (en) * | 2011-05-31 | 2014-07-03 | Hitachi, Ltd. | Autonomous movement system |
US8565958B1 (en) * | 2011-06-02 | 2013-10-22 | Google Inc. | Removing extraneous objects from maps |
US8548671B2 (en) | 2011-06-06 | 2013-10-01 | Crown Equipment Limited | Method and apparatus for automatically calibrating vehicle parameters |
US8788113B2 (en) | 2011-06-13 | 2014-07-22 | Ford Global Technologies, Llc | Vehicle driver advisory system and method |
US8589012B2 (en) | 2011-06-14 | 2013-11-19 | Crown Equipment Limited | Method and apparatus for facilitating map data processing for industrial vehicle navigation |
US8594923B2 (en) | 2011-06-14 | 2013-11-26 | Crown Equipment Limited | Method and apparatus for sharing map data associated with automated industrial vehicles |
KR20120140104A (en) * | 2011-06-20 | 2012-12-28 | 현대자동차주식회사 | Data storing method of memory |
US8825391B1 (en) | 2011-08-04 | 2014-09-02 | Google Inc. | Building elevation maps from laser data |
US8849519B2 (en) | 2011-08-09 | 2014-09-30 | Ford Global Technologies, Llc | Method and apparatus for vehicle hardware theft prevention |
US9151613B2 (en) * | 2011-08-12 | 2015-10-06 | Qualcomm Incorporated | Methods and apparatus for detecting, measuring, and mitigating effects of moving an inertial navigation device's cradle |
US20140058634A1 (en) | 2012-08-24 | 2014-02-27 | Crown Equipment Limited | Method and apparatus for using unique landmarks to locate industrial vehicles at start-up |
US10061189B2 (en) | 2011-09-06 | 2018-08-28 | King Abdullah University of Science and Technology (KAUST) | Stereo-panoramic data |
US9056754B2 (en) | 2011-09-07 | 2015-06-16 | Crown Equipment Limited | Method and apparatus for using pre-positioned objects to localize an industrial vehicle |
US8510029B2 (en) | 2011-10-07 | 2013-08-13 | Southwest Research Institute | Waypoint splining for autonomous vehicle following |
US8630806B1 (en) | 2011-10-20 | 2014-01-14 | Google Inc. | Image processing for vehicle control |
US9147173B2 (en) | 2011-10-31 | 2015-09-29 | Harvest Automation, Inc. | Methods and systems for automated transportation of items between variable endpoints |
US9163948B2 (en) * | 2011-11-17 | 2015-10-20 | Speedgauge, Inc. | Position accuracy testing system |
SE536333C2 (en) * | 2011-11-18 | 2013-08-27 | Atlas Copco Rock Drills Ab | Method and apparatus for operating a mining and / or construction machine |
US9346428B2 (en) * | 2011-11-22 | 2016-05-24 | Autoliv Asp, Inc. | System and method for determining when to deploy a vehicle safety system |
KR101338075B1 (en) * | 2011-12-14 | 2013-12-06 | 현대자동차주식회사 | Method for warning pedestrian using laser beam |
US9187118B2 (en) * | 2011-12-30 | 2015-11-17 | C & P Technologies, Inc. | Method and apparatus for automobile accident reduction using localized dynamic swarming |
US8937410B2 (en) | 2012-01-17 | 2015-01-20 | Harvest Automation, Inc. | Emergency stop method and system for autonomous mobile robots |
DE102012100609A1 (en) | 2012-01-25 | 2013-07-25 | Faro Technologies, Inc. | Device for optically scanning and measuring an environment |
US8930128B2 (en) * | 2012-02-27 | 2015-01-06 | Lit Motors Corporation | Vehicle collision mitigation system |
US9760092B2 (en) * | 2012-03-16 | 2017-09-12 | Waymo Llc | Actively modifying a field of view of an autonomous vehicle in view of constraints |
US8718861B1 (en) | 2012-04-11 | 2014-05-06 | Google Inc. | Determining when to drive autonomously |
US8676428B2 (en) * | 2012-04-17 | 2014-03-18 | Lytx, Inc. | Server request for downloaded information from a vehicle-based monitor |
US9240079B2 (en) | 2012-04-17 | 2016-01-19 | Lytx, Inc. | Triggering a specialized data collection mode |
WO2013162559A1 (en) * | 2012-04-26 | 2013-10-31 | Intel Corporation | Determining relative positioning information |
DE102012103669A1 (en) * | 2012-04-26 | 2013-10-31 | Continental Teves Ag & Co. Ohg | Method for displaying a vehicle environment |
US9569403B2 (en) | 2012-05-03 | 2017-02-14 | Ford Global Technologies, Llc | Methods and systems for authenticating one or more users of a vehicle communications and information system |
US8595037B1 (en) * | 2012-05-08 | 2013-11-26 | Elwha Llc | Systems and methods for insurance based on monitored characteristics of an autonomous drive mode selection system |
US9384668B2 (en) | 2012-05-09 | 2016-07-05 | Singularity University | Transportation using network of unmanned aerial vehicles |
US9632168B2 (en) | 2012-06-19 | 2017-04-25 | Lockheed Martin Corporation | Visual disruption system, method, and computer program product |
US9714815B2 (en) | 2012-06-19 | 2017-07-25 | Lockheed Martin Corporation | Visual disruption network and system, method, and computer program product thereof |
US9180882B1 (en) | 2012-06-20 | 2015-11-10 | Google Inc. | Avoiding blind spots of other vehicles |
US9558667B2 (en) | 2012-07-09 | 2017-01-31 | Elwha Llc | Systems and methods for cooperative collision detection |
US9000903B2 (en) | 2012-07-09 | 2015-04-07 | Elwha Llc | Systems and methods for vehicle monitoring |
US9165469B2 (en) | 2012-07-09 | 2015-10-20 | Elwha Llc | Systems and methods for coordinating sensor operation for collision detection |
US8997362B2 (en) | 2012-07-17 | 2015-04-07 | Faro Technologies, Inc. | Portable articulated arm coordinate measuring machine with optical communications bus |
GB201213171D0 (en) * | 2012-07-24 | 2012-09-05 | Sentinel Systems Ltd | Braking systems |
DE102012106932A1 (en) | 2012-07-30 | 2014-05-15 | Continental Teves Ag & Co. Ohg | Method for displaying a vehicle environment with position points |
US9728228B2 (en) | 2012-08-10 | 2017-08-08 | Smartdrive Systems, Inc. | Vehicle event playback apparatus and methods |
DE102012107544B3 (en) | 2012-08-17 | 2013-05-23 | Faro Technologies, Inc. | Optical scanning device i.e. laser scanner, for evaluating environment, has planetary gears driven by motor over vertical motor shaft and rotating measuring head relative to foot, where motor shaft is arranged coaxial to vertical axle |
US9195236B1 (en) | 2012-09-05 | 2015-11-24 | Google Inc. | Road flare detection |
US9383753B1 (en) | 2012-09-26 | 2016-07-05 | Google Inc. | Wide-view LIDAR with areas of special attention |
US9633564B2 (en) | 2012-09-27 | 2017-04-25 | Google Inc. | Determining changes in a driving environment based on vehicle behavior |
US9234618B1 (en) | 2012-09-27 | 2016-01-12 | Google Inc. | Characterizing optically reflective features via hyper-spectral sensor |
US8949016B1 (en) | 2012-09-28 | 2015-02-03 | Google Inc. | Systems and methods for determining whether a driving environment has changed |
US9513107B2 (en) | 2012-10-05 | 2016-12-06 | Faro Technologies, Inc. | Registration calculation between three-dimensional (3D) scans based on two-dimensional (2D) scan data from a 3D scanner |
US10067231B2 (en) | 2012-10-05 | 2018-09-04 | Faro Technologies, Inc. | Registration calculation of three-dimensional scanner data performed between scans based on measurements by two-dimensional scanner |
DE102012109481A1 (en) | 2012-10-05 | 2014-04-10 | Faro Technologies, Inc. | Device for optically scanning and measuring an environment |
US8965691B1 (en) * | 2012-10-05 | 2015-02-24 | Google Inc. | Position and direction determination using multiple single-channel encoders |
US8849557B1 (en) | 2012-11-15 | 2014-09-30 | Google Inc. | Leveraging of behavior of vehicles to detect likely presence of an emergency vehicle |
US8838321B1 (en) | 2012-11-15 | 2014-09-16 | Google Inc. | Modifying a vehicle state based on the presence of a special-purpose vehicle |
CA2833985C (en) * | 2012-11-19 | 2020-07-07 | Rosemount Aerospace, Inc. | Collision avoidance system for aircraft ground operations |
US8825258B2 (en) | 2012-11-30 | 2014-09-02 | Google Inc. | Engaging and disengaging for autonomous driving |
USD735214S1 (en) | 2012-11-30 | 2015-07-28 | Google Inc. | Display screen or portion thereof with graphical user interface |
US20140158037A1 (en) * | 2012-12-10 | 2014-06-12 | Ronald A. George | Trimaran having outriggers with propulsors |
DE102013200387A1 (en) * | 2013-01-14 | 2014-07-17 | Robert Bosch Gmbh | Creation of a obstacle map |
JP6128584B2 (en) * | 2013-01-16 | 2017-05-17 | 株式会社Soken | Travel route generator |
US9086481B1 (en) | 2013-01-18 | 2015-07-21 | Google Inc. | Methods and systems for estimating vehicle speed |
US9267792B2 (en) * | 2013-01-21 | 2016-02-23 | Systèmes Pavemetrics Inc. | Method and apparatus for compensating lateral displacements and low speed variations in the measure of a longitudinal profile of a surface |
US9285477B1 (en) | 2013-01-25 | 2016-03-15 | Apple Inc. | 3D depth point cloud from timing flight of 2D scanned light beam pulses |
US9688246B2 (en) | 2013-02-25 | 2017-06-27 | Ford Global Technologies, Llc | Method and apparatus for in-vehicle alarm activation and response handling |
US8947221B2 (en) | 2013-02-26 | 2015-02-03 | Ford Global Technologies, Llc | Method and apparatus for tracking device connection and state change |
US9274222B1 (en) * | 2013-03-04 | 2016-03-01 | Toyota Motor Engineering & Manufacturing North America, Inc. | Dynamic allocation of radar beams in automotive environments with phased array radar |
US8676431B1 (en) | 2013-03-12 | 2014-03-18 | Google Inc. | User interface for displaying object-based indications in an autonomous driving system |
USD750663S1 (en) | 2013-03-12 | 2016-03-01 | Google Inc. | Display screen or a portion thereof with graphical user interface |
US9304203B1 (en) * | 2013-03-13 | 2016-04-05 | Google Inc. | Methods, devices, and systems for improving dynamic range of signal receiver |
USD754190S1 (en) | 2013-03-13 | 2016-04-19 | Google Inc. | Display screen or portion thereof with graphical user interface |
US9141583B2 (en) | 2013-03-13 | 2015-09-22 | Ford Global Technologies, Llc | Method and system for supervising information communication based on occupant and vehicle environment |
USD754189S1 (en) | 2013-03-13 | 2016-04-19 | Google Inc. | Display screen or portion thereof with graphical user interface |
US9103628B1 (en) * | 2013-03-14 | 2015-08-11 | Lockheed Martin Corporation | System, method, and computer program product for hostile fire strike indication |
US9146251B2 (en) | 2013-03-14 | 2015-09-29 | Lockheed Martin Corporation | System, method, and computer program product for indicating hostile fire |
US9196041B2 (en) | 2013-03-14 | 2015-11-24 | Lockheed Martin Corporation | System, method, and computer program product for indicating hostile fire |
US9002536B2 (en) | 2013-03-14 | 2015-04-07 | Ford Global Technologies, Llc | Key fob security copy to a mobile phone |
US9081382B2 (en) | 2013-03-15 | 2015-07-14 | Fresenius Medical Care Holdings, Inc. | Autonomous vehicle comprising extracorporeal blood treatment machine |
KR101803164B1 (en) | 2013-03-15 | 2017-12-28 | 우버 테크놀로지스, 인크. | Methods, systems, and apparatus for multi-sensory stereo vision for robotics |
US9111156B2 (en) | 2013-03-15 | 2015-08-18 | Mojix, Inc. | Systems and methods for compressive sensing ranging evaluation |
SE537265C2 (en) * | 2013-03-19 | 2015-03-17 | Scania Cv Ab | Control system and method for controlling vehicles when detecting obstacles |
CN105073542B (en) * | 2013-04-01 | 2018-03-16 | 朴秀旼 | Automatic vehicle control system |
US20140309836A1 (en) * | 2013-04-16 | 2014-10-16 | Neya Systems, Llc | Position Estimation and Vehicle Control in Autonomous Multi-Vehicle Convoys |
US9753137B2 (en) | 2013-05-26 | 2017-09-05 | Intel Corporation | Apparatus, system and method of communicating positioning information |
EP2821307B1 (en) * | 2013-07-03 | 2016-09-28 | Volvo Car Corporation | A vehicle system, a vehicle and a method for autonomous road irregularity avoidance |
US9062979B1 (en) | 2013-07-08 | 2015-06-23 | Google Inc. | Pose estimation using long range features |
WO2015005912A1 (en) | 2013-07-10 | 2015-01-15 | Intel Corporation | Apparatus, system and method of communicating positioning transmissions |
US8930060B1 (en) * | 2013-07-15 | 2015-01-06 | Ford Global Technologies | Post-impact path assist for vehicles |
US9286520B1 (en) | 2013-07-16 | 2016-03-15 | Google Inc. | Real-time road flare detection using templates and appropriate color spaces |
US9558584B1 (en) * | 2013-07-29 | 2017-01-31 | Google Inc. | 3D position estimation of objects from a monocular camera using a set of known 3D points on an underlying surface |
US9776632B2 (en) * | 2013-07-31 | 2017-10-03 | Elwha Llc | Systems and methods for adaptive vehicle sensing systems |
US9230442B2 (en) | 2013-07-31 | 2016-01-05 | Elwha Llc | Systems and methods for adaptive vehicle sensing systems |
US9269268B2 (en) | 2013-07-31 | 2016-02-23 | Elwha Llc | Systems and methods for adaptive vehicle sensing systems |
CN104442552B (en) * | 2013-09-13 | 2018-03-09 | 富泰华工业(深圳)有限公司 | Vehicle fitting and its CAS |
US20150100191A1 (en) * | 2013-10-09 | 2015-04-09 | Ford Global Technologies, Llc | Monitoring autonomous vehicle steering |
GB201318185D0 (en) | 2013-10-14 | 2013-11-27 | Guidance Navigation Ltd | Tracking device |
US9501878B2 (en) | 2013-10-16 | 2016-11-22 | Smartdrive Systems, Inc. | Vehicle event playback apparatus and methods |
US9174672B2 (en) * | 2013-10-28 | 2015-11-03 | GM Global Technology Operations LLC | Path planning for evasive steering maneuver in presence of target vehicle and surrounding objects |
US9199668B2 (en) * | 2013-10-28 | 2015-12-01 | GM Global Technology Operations LLC | Path planning for evasive steering maneuver employing a virtual potential field technique |
DE102013018724B4 (en) * | 2013-11-08 | 2015-09-03 | LASE Industrielle Lasertechnik GmbH | Agricultural harvester |
US9610955B2 (en) | 2013-11-11 | 2017-04-04 | Smartdrive Systems, Inc. | Vehicle fuel consumption monitor and feedback systems |
US9552735B2 (en) | 2013-11-22 | 2017-01-24 | Ford Global Technologies, Llc | Autonomous vehicle identification |
US9364178B2 (en) * | 2013-11-26 | 2016-06-14 | Elwha Llc | Robotic vehicle control |
US8989944B1 (en) | 2013-11-26 | 2015-03-24 | Google Inc. | Methods and devices for determining movements of an object in an environment |
JP5939238B2 (en) * | 2013-11-29 | 2016-06-22 | トヨタ自動車株式会社 | Vehicle steering control device |
JP6042794B2 (en) * | 2013-12-03 | 2016-12-14 | 本田技研工業株式会社 | Vehicle control method |
RU2550907C1 (en) * | 2013-12-04 | 2015-05-20 | Олег Иванович Завалишин | Aircraft landing with help of uv receivers abd radiators (2 versions) |
US9393998B2 (en) | 2013-12-04 | 2016-07-19 | Mobileye Vision Technologies Ltd. | Systems and methods for vehicle offset navigation |
US9340207B2 (en) * | 2014-01-16 | 2016-05-17 | Toyota Motor Engineering & Manufacturing North America, Inc. | Lateral maneuver planner for automated driving system |
US9104205B1 (en) | 2014-01-22 | 2015-08-11 | Fred Caver | Automated return shopping cart |
CN103761637B (en) * | 2014-01-27 | 2016-09-28 | 林兴志 | Multiple spatial locations information superposition location models management method with the storage of navigation |
US9493170B2 (en) * | 2014-01-29 | 2016-11-15 | Continental Automotive Systems, Inc. | Method for reducing false activations in reverse collision avoidance systems |
EP3100206B1 (en) * | 2014-01-30 | 2020-09-09 | Mobileye Vision Technologies Ltd. | Systems and methods for lane end recognition |
US8892310B1 (en) | 2014-02-21 | 2014-11-18 | Smartdrive Systems, Inc. | System and method to detect execution of driving maneuvers |
KR102173994B1 (en) * | 2014-03-18 | 2020-11-04 | 주식회사 만도 | Tuning method and control device of vehicle radar |
US9377314B2 (en) * | 2014-04-02 | 2016-06-28 | Ford Global Technologies, Llc | Reduced network flow and computational load using a spatial and temporal variable scheduler |
US9972054B1 (en) | 2014-05-20 | 2018-05-15 | State Farm Mutual Automobile Insurance Company | Accident fault determination for autonomous vehicles |
US10319039B1 (en) | 2014-05-20 | 2019-06-11 | State Farm Mutual Automobile Insurance Company | Accident fault determination for autonomous vehicles |
US10185999B1 (en) | 2014-05-20 | 2019-01-22 | State Farm Mutual Automobile Insurance Company | Autonomous feature use monitoring and telematics |
US10599155B1 (en) | 2014-05-20 | 2020-03-24 | State Farm Mutual Automobile Insurance Company | Autonomous vehicle operation feature monitoring and evaluation of effectiveness |
US9646428B1 (en) | 2014-05-20 | 2017-05-09 | State Farm Mutual Automobile Insurance Company | Accident response using autonomous vehicle monitoring |
US11669090B2 (en) | 2014-05-20 | 2023-06-06 | State Farm Mutual Automobile Insurance Company | Autonomous vehicle operation feature monitoring and evaluation of effectiveness |
US10373259B1 (en) | 2014-05-20 | 2019-08-06 | State Farm Mutual Automobile Insurance Company | Fully autonomous vehicle insurance pricing |
US9436182B2 (en) | 2014-05-23 | 2016-09-06 | Google Inc. | Autonomous vehicles |
US9631933B1 (en) | 2014-05-23 | 2017-04-25 | Google Inc. | Specifying unavailable locations for autonomous vehicles |
US9643615B2 (en) * | 2014-06-04 | 2017-05-09 | International Business Machines Corporation | Automotive dynamic virtual network |
US9798007B2 (en) | 2014-07-01 | 2017-10-24 | Sikorsky Aircraft Corporation | Obstacle data model construction system with range sensor shadows and use in motion planning |
US10102587B1 (en) | 2014-07-21 | 2018-10-16 | State Farm Mutual Automobile Insurance Company | Methods of pre-generating insurance claims |
DE102014111126A1 (en) * | 2014-08-05 | 2016-02-11 | Valeo Schalter Und Sensoren Gmbh | Method for generating an environment map of an environmental area of a motor vehicle, driver assistance system and motor vehicle |
US9321461B1 (en) | 2014-08-29 | 2016-04-26 | Google Inc. | Change detection using curve alignment |
KR101610502B1 (en) * | 2014-09-02 | 2016-04-07 | 현대자동차주식회사 | Apparatus and method for recognizing driving enviroment for autonomous vehicle |
JP5883489B1 (en) * | 2014-09-30 | 2016-03-15 | 富士重工業株式会社 | Vehicle control apparatus and vehicle control method |
US9248834B1 (en) | 2014-10-02 | 2016-02-02 | Google Inc. | Predicting trajectories of objects based on contextual information |
TWI564193B (en) * | 2014-10-15 | 2017-01-01 | 華創車電技術中心股份有限公司 | System and method for vehicle steering control |
US9663127B2 (en) | 2014-10-28 | 2017-05-30 | Smartdrive Systems, Inc. | Rail vehicle event detection and recording system |
JP6025268B2 (en) | 2014-10-31 | 2016-11-16 | 富士重工業株式会社 | Vehicle travel control device |
KR101664582B1 (en) * | 2014-11-12 | 2016-10-10 | 현대자동차주식회사 | Path Planning Apparatus and Method for Autonomous Vehicle |
US10431018B1 (en) | 2014-11-13 | 2019-10-01 | State Farm Mutual Automobile Insurance Company | Autonomous vehicle operating status assessment |
US11069257B2 (en) | 2014-11-13 | 2021-07-20 | Smartdrive Systems, Inc. | System and method for detecting a vehicle event and generating review criteria |
DE102014223900A1 (en) * | 2014-11-24 | 2016-05-25 | Conti Temic Microelectronic Gmbh | Vehicle environment scanning using a phased array laser |
US10183659B2 (en) * | 2014-11-24 | 2019-01-22 | Ford Global Technologies, Llc | Vehicle underside impact avoidance |
CN107624155B (en) * | 2014-12-05 | 2021-09-28 | 苹果公司 | Autonomous Navigation System |
US10126745B2 (en) * | 2015-01-04 | 2018-11-13 | Hangzhou Zero Zero Technology Co., Ltd. | System and method for automated aerial system operation |
US10719080B2 (en) | 2015-01-04 | 2020-07-21 | Hangzhou Zero Zero Technology Co., Ltd. | Aerial system and detachable housing |
US10358214B2 (en) | 2015-01-04 | 2019-07-23 | Hangzhou Zero Zro Technology Co., Ltd. | Aerial vehicle and method of operation |
US10220954B2 (en) | 2015-01-04 | 2019-03-05 | Zero Zero Robotics Inc | Aerial system thermal control system and method |
US9836053B2 (en) | 2015-01-04 | 2017-12-05 | Zero Zero Robotics Inc. | System and method for automated aerial system operation |
WO2016113904A1 (en) * | 2015-01-16 | 2016-07-21 | 株式会社日立製作所 | Three-dimensional-information-calculating device, method for calculating three-dimensional information, and autonomous mobile device |
JP6176263B2 (en) * | 2015-01-19 | 2017-08-09 | トヨタ自動車株式会社 | Automatic driving device |
US10216196B2 (en) * | 2015-02-01 | 2019-02-26 | Prosper Technology, Llc | Methods to operate autonomous vehicles to pilot vehicles in groups or convoys |
CN107438754A (en) * | 2015-02-10 | 2017-12-05 | 御眼视觉技术有限公司 | Sparse map for autonomous vehicle navigation |
DE102015102080A1 (en) * | 2015-02-13 | 2016-08-18 | Horsch Leeb Application Systems Gmbh | Device for dispensing liquids and method for controlling the movement of at least two extension arms of an agricultural field sprayer |
DE102015205032A1 (en) | 2015-03-19 | 2016-09-22 | Kuka Roboter Gmbh | Vehicle network and method for forming and operating a vehicle network |
US9625582B2 (en) | 2015-03-25 | 2017-04-18 | Google Inc. | Vehicle with multiple light detection and ranging devices (LIDARs) |
US9679420B2 (en) | 2015-04-01 | 2017-06-13 | Smartdrive Systems, Inc. | Vehicle event recording system and method |
JP6299651B2 (en) * | 2015-04-02 | 2018-03-28 | 株式会社デンソー | Image processing device |
US10249123B2 (en) | 2015-04-09 | 2019-04-02 | Ford Global Technologies, Llc | Systems and methods for mobile phone key fob management |
US9505404B2 (en) * | 2015-04-10 | 2016-11-29 | Jaguar Land Rover Limited | Collision avoidance system |
US9883337B2 (en) * | 2015-04-24 | 2018-01-30 | Mijix, Inc. | Location based services for RFID and sensor networks |
EP3304520B1 (en) | 2015-05-27 | 2023-09-20 | Dov Moran | Alerting predicted accidents between driverless cars |
US10031522B2 (en) | 2015-05-27 | 2018-07-24 | Dov Moran | Alerting predicted accidents between driverless cars |
DE102015210015A1 (en) * | 2015-06-01 | 2016-12-01 | Robert Bosch Gmbh | Method and device for determining the position of a vehicle |
US20160356010A1 (en) * | 2015-06-02 | 2016-12-08 | Christopher Wilkinson | Apparatus for removal and processing of frozen or semi-frozen materials |
WO2016194168A1 (en) * | 2015-06-03 | 2016-12-08 | 日産自動車株式会社 | Travel control device and method |
US20160368336A1 (en) | 2015-06-19 | 2016-12-22 | Paccar Inc | Use of laser scanner for autonomous truck operation |
KR20170015115A (en) | 2015-07-30 | 2017-02-08 | 삼성전자주식회사 | Autonomous vehicle and method for controlling the autonomous vehicle |
US9869560B2 (en) | 2015-07-31 | 2018-01-16 | International Business Machines Corporation | Self-driving vehicle's response to a proximate emergency vehicle |
US9594378B2 (en) | 2015-07-31 | 2017-03-14 | Delphi Technologies, Inc. | Variable object detection field-of-focus for automated vehicle control |
JP6899369B2 (en) | 2015-08-03 | 2021-07-07 | トムトム グローバル コンテント ベスローテン フエンノートシャップ | Methods and systems for generating and using localization reference data |
US9785145B2 (en) | 2015-08-07 | 2017-10-10 | International Business Machines Corporation | Controlling driving modes of self-driving vehicles |
US9721397B2 (en) | 2015-08-11 | 2017-08-01 | International Business Machines Corporation | Automatic toll booth interaction with self-driving vehicles |
US9718471B2 (en) | 2015-08-18 | 2017-08-01 | International Business Machines Corporation | Automated spatial separation of self-driving vehicles from manually operated vehicles |
US9896100B2 (en) | 2015-08-24 | 2018-02-20 | International Business Machines Corporation | Automated spatial separation of self-driving vehicles from other vehicles based on occupant preferences |
US9805601B1 (en) | 2015-08-28 | 2017-10-31 | State Farm Mutual Automobile Insurance Company | Vehicular traffic alerts for avoidance of abnormal traffic conditions |
US9731726B2 (en) | 2015-09-02 | 2017-08-15 | International Business Machines Corporation | Redirecting self-driving vehicles to a product provider based on physiological states of occupants of the self-driving vehicles |
US9952304B2 (en) | 2015-09-10 | 2018-04-24 | Ford Global Technologies, Llc | Vehicle positioning system |
US9969329B2 (en) * | 2015-09-16 | 2018-05-15 | Sony Corporation | System and method for generation of a preventive alert |
US10061020B2 (en) * | 2015-09-20 | 2018-08-28 | Qualcomm Incorporated | Light detection and ranging (LIDAR) system with dual beam steering |
US9566986B1 (en) | 2015-09-25 | 2017-02-14 | International Business Machines Corporation | Controlling driving modes of self-driving vehicles |
EP3356899B1 (en) | 2015-09-28 | 2021-12-29 | Uatc, Llc | Method of operating an autonomous vehicle having independent auxiliary control unit |
US10293653B2 (en) * | 2015-09-29 | 2019-05-21 | Ford Global Technologies, Llc | Obstacle avoidance system with active suspensions |
US9834224B2 (en) | 2015-10-15 | 2017-12-05 | International Business Machines Corporation | Controlling driving modes of self-driving vehicles |
US10557939B2 (en) | 2015-10-19 | 2020-02-11 | Luminar Technologies, Inc. | Lidar system with improved signal-to-noise ratio in the presence of solar background noise |
US10527730B2 (en) * | 2015-10-22 | 2020-01-07 | Toyota Motor Engineering & Manufacturing North America, Inc. | Object detection system |
US9751532B2 (en) | 2015-10-27 | 2017-09-05 | International Business Machines Corporation | Controlling spacing of self-driving vehicles based on social network relationships |
US9944291B2 (en) | 2015-10-27 | 2018-04-17 | International Business Machines Corporation | Controlling driving modes of self-driving vehicles |
US10607293B2 (en) | 2015-10-30 | 2020-03-31 | International Business Machines Corporation | Automated insurance toggling for self-driving vehicles |
DE102015118767A1 (en) * | 2015-11-03 | 2017-05-04 | Claas Selbstfahrende Erntemaschinen Gmbh | Environment detection device for agricultural machine |
US10488496B2 (en) | 2015-11-05 | 2019-11-26 | Luminar Technologies, Inc. | Lidar system with improved scanning speed for high-resolution depth mapping |
US10176525B2 (en) | 2015-11-09 | 2019-01-08 | International Business Machines Corporation | Dynamically adjusting insurance policy parameters for a self-driving vehicle |
EP4001111A3 (en) | 2015-11-10 | 2022-08-17 | Matternet, Inc. | Methods and system for transportation using unmanned aerial vehicles |
US9791861B2 (en) | 2015-11-12 | 2017-10-17 | International Business Machines Corporation | Autonomously servicing self-driving vehicles |
US10520602B2 (en) | 2015-11-30 | 2019-12-31 | Luminar Technologies, Inc. | Pulsed laser for lidar system |
US9862364B2 (en) * | 2015-12-04 | 2018-01-09 | Waymo Llc | Collision mitigated braking for autonomous vehicles |
US10061326B2 (en) | 2015-12-09 | 2018-08-28 | International Business Machines Corporation | Mishap amelioration based on second-order sensing by a self-driving vehicle |
US10338225B2 (en) * | 2015-12-15 | 2019-07-02 | Uber Technologies, Inc. | Dynamic LIDAR sensor controller |
US9946259B2 (en) * | 2015-12-18 | 2018-04-17 | Raytheon Company | Negative obstacle detector |
DE102015122844A1 (en) | 2015-12-27 | 2017-06-29 | Faro Technologies, Inc. | 3D measuring device with battery pack |
US9571980B1 (en) * | 2015-12-28 | 2017-02-14 | Cisco Technology, Inc. | Augmenting Wi-Fi localization with auxiliary sensor information |
JP2017122990A (en) * | 2016-01-05 | 2017-07-13 | 株式会社リコー | Travel device, travel control device, and travel control system |
US10481173B2 (en) * | 2016-01-05 | 2019-11-19 | Pasco Scientific | Wireless smart devices having integrated force, position, acceleration, and rotational sensing for science education |
US10482789B2 (en) * | 2016-01-05 | 2019-11-19 | Pasco Scientific | Controllable propulsive force devices including parameter adjustments for wireless smart devices |
US10324463B1 (en) | 2016-01-22 | 2019-06-18 | State Farm Mutual Automobile Insurance Company | Autonomous vehicle operation adjustment based upon route |
US10395332B1 (en) | 2016-01-22 | 2019-08-27 | State Farm Mutual Automobile Insurance Company | Coordinated autonomous vehicle automatic area scanning |
US11441916B1 (en) | 2016-01-22 | 2022-09-13 | State Farm Mutual Automobile Insurance Company | Autonomous vehicle trip routing |
US10386845B1 (en) | 2016-01-22 | 2019-08-20 | State Farm Mutual Automobile Insurance Company | Autonomous vehicle parking |
US11242051B1 (en) | 2016-01-22 | 2022-02-08 | State Farm Mutual Automobile Insurance Company | Autonomous vehicle action communications |
US11719545B2 (en) | 2016-01-22 | 2023-08-08 | Hyundai Motor Company | Autonomous vehicle component damage and salvage assessment |
US10134278B1 (en) | 2016-01-22 | 2018-11-20 | State Farm Mutual Automobile Insurance Company | Autonomous vehicle application |
US9940834B1 (en) | 2016-01-22 | 2018-04-10 | State Farm Mutual Automobile Insurance Company | Autonomous vehicle application |
US9836973B2 (en) | 2016-01-27 | 2017-12-05 | International Business Machines Corporation | Selectively controlling a self-driving vehicle's access to a roadway |
SG11201806223XA (en) * | 2016-01-29 | 2018-08-30 | Univ Meiji | Laser scanning system, laser scanning method, movable laser scanning system, and program |
US9896123B2 (en) * | 2016-01-29 | 2018-02-20 | Tk Holdings Inc. | Systems and methods for detecting steering wheel contact |
US10627490B2 (en) | 2016-01-31 | 2020-04-21 | Velodyne Lidar, Inc. | Multiple pulse, LIDAR based 3-D imaging |
WO2017139917A1 (en) * | 2016-02-15 | 2017-08-24 | 吴伟民 | Vehicle obstacle detection method and system |
US10281923B2 (en) | 2016-03-03 | 2019-05-07 | Uber Technologies, Inc. | Planar-beam, light detection and ranging system |
US9802599B2 (en) * | 2016-03-08 | 2017-10-31 | Ford Global Technologies, Llc | Vehicle lane placement |
JP6945785B2 (en) * | 2016-03-14 | 2021-10-06 | イムラ ウーロプ ソシエテ・パ・アクシオンス・シンプリフィエ | 3D point cloud processing method |
US9580080B1 (en) * | 2016-03-15 | 2017-02-28 | Uber Technologies, Inc. | Drive-by-wire control system |
EP3430428B1 (en) | 2016-03-19 | 2025-04-02 | Velodyne Lidar USA, Inc. | Integrated illumination and detection for lidar based 3-d imaging |
US10359779B2 (en) | 2016-03-22 | 2019-07-23 | Aurora Flight Sciences Corporation | Aircrew automation system and method |
US10179583B2 (en) | 2016-03-29 | 2019-01-15 | Ford Global Technologies, Llc | Vehicle subsystem coordination and control |
US10761195B2 (en) | 2016-04-22 | 2020-09-01 | OPSYS Tech Ltd. | Multi-wavelength LIDAR system |
WO2017187275A2 (en) | 2016-04-24 | 2017-11-02 | Hangzhou Zero Zero Technology Co., Ltd. | Aerial system propulsion assembly and method of use |
US10451740B2 (en) * | 2016-04-26 | 2019-10-22 | Cepton Technologies, Inc. | Scanning lidar systems for three-dimensional sensing |
US10255816B2 (en) * | 2016-04-27 | 2019-04-09 | Uber Technologies, Inc. | Transport vehicle configuration for impaired riders |
CN111469793A (en) * | 2016-04-28 | 2020-07-31 | 赵忠尧 | Vehicle-mounted front anti-collision system based on laser radar |
US10152891B2 (en) | 2016-05-02 | 2018-12-11 | Cnh Industrial America Llc | System for avoiding collisions between autonomous vehicles conducting agricultural operations |
US10838062B2 (en) * | 2016-05-24 | 2020-11-17 | Veoneer Us, Inc. | Direct detection LiDAR system and method with pulse amplitude modulation (AM) transmitter and quadrature receiver |
US10685391B2 (en) | 2016-05-24 | 2020-06-16 | International Business Machines Corporation | Directing movement of a self-driving vehicle based on sales activity |
US9952317B2 (en) | 2016-05-27 | 2018-04-24 | Uber Technologies, Inc. | Vehicle sensor calibration system |
US10393877B2 (en) | 2016-06-01 | 2019-08-27 | Velodyne Lidar, Inc. | Multiple pixel scanning LIDAR |
IL246386B (en) * | 2016-06-22 | 2018-03-29 | Tarantula Tech Ltd | Apparatus for detecting hazardous objects within a designated distance from a surface |
US9747804B1 (en) * | 2016-06-23 | 2017-08-29 | GM Global Technology Operations LLC | Object detection-based directional control of light and sound |
US10165231B2 (en) | 2016-06-29 | 2018-12-25 | International Business Machines Corporation | Visualization of navigation information for connected autonomous vehicles |
US10832331B1 (en) * | 2016-07-11 | 2020-11-10 | State Farm Mutual Automobile Insurance Company | Systems and methods for allocating fault to autonomous vehicles |
US10007264B2 (en) | 2016-07-14 | 2018-06-26 | Baidu Usa Llc | Autonomous vehicle human driver takeover mechanism using electrodes |
US10162354B2 (en) * | 2016-07-21 | 2018-12-25 | Baidu Usa Llc | Controlling error corrected planning methods for operating autonomous vehicles |
US10203696B2 (en) * | 2016-07-26 | 2019-02-12 | Waymo Llc | Determining drivability of objects for autonomous vehicles |
US10175361B2 (en) * | 2016-07-28 | 2019-01-08 | Sharp Laboratories Of America, Inc. | System and method for three-dimensional mapping using two-dimensional LiDAR laser ranging |
BR112019003807A2 (en) | 2016-08-26 | 2019-05-21 | Crown Equipment Corporation | material handling vehicle, and method for performing scan logic for a material handling vehicle. |
CN115328158A (en) | 2016-08-26 | 2022-11-11 | 克朗设备公司 | Multi-field scanning tool in a materials handling vehicle |
EP3504602B1 (en) | 2016-08-26 | 2022-03-16 | Crown Equipment Corporation | Materials handling vehicle path validation and dynamic path modification |
CN106272562A (en) * | 2016-08-31 | 2017-01-04 | 贺宜 | A kind of machine vision and the fitting method of robot |
WO2018039975A1 (en) * | 2016-08-31 | 2018-03-08 | SZ DJI Technology Co., Ltd. | Laser radar scanning and positioning mechanisms for uavs and other objects, and associated systems and methods |
US10108194B1 (en) | 2016-09-02 | 2018-10-23 | X Development Llc | Object placement verification |
US20180067195A1 (en) * | 2016-09-08 | 2018-03-08 | Qualcomm Incorporated | Multi-tier light-based ranging systems and methods |
US10394237B2 (en) * | 2016-09-08 | 2019-08-27 | Ford Global Technologies, Llc | Perceiving roadway conditions from fused sensor data |
WO2018050224A1 (en) * | 2016-09-15 | 2018-03-22 | Connaught Electronics Ltd. | Method for determining a plane of a roadway, driver assistance system as well as motor vehicle |
US10093322B2 (en) | 2016-09-15 | 2018-10-09 | International Business Machines Corporation | Automatically providing explanations for actions taken by a self-driving vehicle |
US10643256B2 (en) | 2016-09-16 | 2020-05-05 | International Business Machines Corporation | Configuring a self-driving vehicle for charitable donations pickup and delivery |
US10502574B2 (en) * | 2016-09-20 | 2019-12-10 | Waymo Llc | Devices and methods for a sensor platform of a vehicle |
WO2018093460A2 (en) * | 2016-09-22 | 2018-05-24 | NOAA Technology Partnerships Office | System, apparatus and methodology for flight operation imaging |
DE102016118481A1 (en) * | 2016-09-29 | 2018-03-29 | Valeo Schalter Und Sensoren Gmbh | Scanning unit of an optical transmitting and receiving device of an optical detection device of a vehicle |
WO2018064727A1 (en) * | 2016-10-07 | 2018-04-12 | Superior Pak Holdings Pty Ltd | Detection system for objects at the side of a vehicle |
FR3057534B1 (en) * | 2016-10-17 | 2018-11-30 | Metalliance | AUTONOMOUS TRANSPORT SYSTEM FOR TUNNEL |
JP6854620B2 (en) * | 2016-10-26 | 2021-04-07 | 株式会社クボタ | Travel route generator |
US10565682B2 (en) * | 2016-11-07 | 2020-02-18 | Ford Global Technologies, Llc | Constructing map data using laser scanned images |
US10620011B2 (en) | 2016-11-10 | 2020-04-14 | International Business Machines Corporation | Autonomous vehicle routing |
IT201600114161A1 (en) * | 2016-11-11 | 2018-05-11 | Info Solution S P A | METHOD AND DEVICE FOR PILOTING A SELF-PROPELLED VEHICLE AND ITS PILOT SYSTEM |
US10259454B2 (en) * | 2016-11-16 | 2019-04-16 | Nio Usa, Inc. | System for controlling a vehicle based on wheel angle tracking |
US11042161B2 (en) * | 2016-11-16 | 2021-06-22 | Symbol Technologies, Llc | Navigation control method and apparatus in a mobile automation system |
CN106515713A (en) * | 2016-11-18 | 2017-03-22 | 华南农业大学 | Car evadable system and method thereof |
KR101911703B1 (en) * | 2016-12-09 | 2018-10-25 | 엘지전자 주식회사 | Driving control apparatus for vehicle and vehicle |
JP6597585B2 (en) * | 2016-12-15 | 2019-10-30 | トヨタ自動車株式会社 | Driving assistance device |
EP3339999B1 (en) * | 2016-12-22 | 2021-10-27 | Panasonic Intellectual Property Corporation of America | Information processing apparatus, information processing method, and recording medium storing programm |
US10742088B2 (en) | 2016-12-30 | 2020-08-11 | Panosense Inc. | Support assembly for rotating body |
US10109183B1 (en) | 2016-12-30 | 2018-10-23 | Panosense Inc. | Interface for transferring data between a non-rotating body and a rotating body |
US10591740B2 (en) | 2016-12-30 | 2020-03-17 | Panosense Inc. | Lens assembly for a LIDAR system |
US10359507B2 (en) | 2016-12-30 | 2019-07-23 | Panosense Inc. | Lidar sensor assembly calibration based on reference surface |
US10830878B2 (en) | 2016-12-30 | 2020-11-10 | Panosense Inc. | LIDAR system |
US10048358B2 (en) | 2016-12-30 | 2018-08-14 | Panosense Inc. | Laser power calibration and correction |
US10122416B2 (en) | 2016-12-30 | 2018-11-06 | Panosense Inc. | Interface for transferring power and data between a non-rotating body and a rotating body |
US11255951B1 (en) | 2016-12-30 | 2022-02-22 | Zoox, Inc. | Aligning optical components in LIDAR systems |
US10942257B2 (en) | 2016-12-31 | 2021-03-09 | Innovusion Ireland Limited | 2D scanning high precision LiDAR using combination of rotating concave mirror and beam steering devices |
US10520592B2 (en) | 2016-12-31 | 2019-12-31 | Waymo Llc | Light detection and ranging (LIDAR) device with an off-axis receiver |
US10259452B2 (en) | 2017-01-04 | 2019-04-16 | International Business Machines Corporation | Self-driving vehicle collision management system |
US10529147B2 (en) | 2017-01-05 | 2020-01-07 | International Business Machines Corporation | Self-driving vehicle road safety flare deploying system |
US10363893B2 (en) | 2017-01-05 | 2019-07-30 | International Business Machines Corporation | Self-driving vehicle contextual lock control system |
US10332292B1 (en) * | 2017-01-17 | 2019-06-25 | Zoox, Inc. | Vision augmentation for supplementing a person's view |
US10801177B2 (en) | 2017-01-23 | 2020-10-13 | Built Robotics Inc. | Excavating earth from a dig site using an excavation vehicle |
US10152060B2 (en) | 2017-03-08 | 2018-12-11 | International Business Machines Corporation | Protecting contents of a smart vault being transported by a self-driving vehicle |
US10147193B2 (en) | 2017-03-10 | 2018-12-04 | TuSimple | System and method for semantic segmentation using hybrid dilated convolution (HDC) |
WO2018169758A1 (en) | 2017-03-13 | 2018-09-20 | OPSYS Tech Ltd. | Eye-safe scanning lidar system |
US9810775B1 (en) | 2017-03-16 | 2017-11-07 | Luminar Technologies, Inc. | Q-switched laser for LIDAR system |
US9905992B1 (en) | 2017-03-16 | 2018-02-27 | Luminar Technologies, Inc. | Self-Raman laser for lidar system |
US9810786B1 (en) | 2017-03-16 | 2017-11-07 | Luminar Technologies, Inc. | Optical parametric oscillator for lidar system |
US9869754B1 (en) | 2017-03-22 | 2018-01-16 | Luminar Technologies, Inc. | Scan patterns for lidar systems |
US10479376B2 (en) | 2017-03-23 | 2019-11-19 | Uatc, Llc | Dynamic sensor selection for self-driving vehicles |
US10114111B2 (en) | 2017-03-28 | 2018-10-30 | Luminar Technologies, Inc. | Method for dynamically controlling laser power |
US11119198B2 (en) | 2017-03-28 | 2021-09-14 | Luminar, Llc | Increasing operational safety of a lidar system |
US10139478B2 (en) | 2017-03-28 | 2018-11-27 | Luminar Technologies, Inc. | Time varying gain in an optical detector operating in a lidar system |
US10254388B2 (en) | 2017-03-28 | 2019-04-09 | Luminar Technologies, Inc. | Dynamically varying laser output in a vehicle in view of weather conditions |
US10061019B1 (en) | 2017-03-28 | 2018-08-28 | Luminar Technologies, Inc. | Diffractive optical element in a lidar system to correct for backscan |
US10121813B2 (en) | 2017-03-28 | 2018-11-06 | Luminar Technologies, Inc. | Optical detector having a bandpass filter in a lidar system |
US10732281B2 (en) | 2017-03-28 | 2020-08-04 | Luminar Technologies, Inc. | Lidar detector system having range walk compensation |
US10007001B1 (en) | 2017-03-28 | 2018-06-26 | Luminar Technologies, Inc. | Active short-wave infrared four-dimensional camera |
US10209359B2 (en) | 2017-03-28 | 2019-02-19 | Luminar Technologies, Inc. | Adaptive pulse rate in a lidar system |
US10545240B2 (en) | 2017-03-28 | 2020-01-28 | Luminar Technologies, Inc. | LIDAR transmitter and detector system using pulse encoding to reduce range ambiguity |
US10267899B2 (en) | 2017-03-28 | 2019-04-23 | Luminar Technologies, Inc. | Pulse timing based on angle of view |
US10191155B2 (en) | 2017-03-29 | 2019-01-29 | Luminar Technologies, Inc. | Optical resolution in front of a vehicle |
US10969488B2 (en) | 2017-03-29 | 2021-04-06 | Luminar Holdco, Llc | Dynamically scanning a field of regard using a limited number of output beams |
US10088559B1 (en) | 2017-03-29 | 2018-10-02 | Luminar Technologies, Inc. | Controlling pulse timing to compensate for motor dynamics |
US10976417B2 (en) | 2017-03-29 | 2021-04-13 | Luminar Holdco, Llc | Using detectors with different gains in a lidar system |
US11002853B2 (en) | 2017-03-29 | 2021-05-11 | Luminar, Llc | Ultrasonic vibrations on a window in a lidar system |
US10254762B2 (en) | 2017-03-29 | 2019-04-09 | Luminar Technologies, Inc. | Compensating for the vibration of the vehicle |
US10663595B2 (en) | 2017-03-29 | 2020-05-26 | Luminar Technologies, Inc. | Synchronized multiple sensor head system for a vehicle |
US10641874B2 (en) | 2017-03-29 | 2020-05-05 | Luminar Technologies, Inc. | Sizing the field of view of a detector to improve operation of a lidar system |
US10983213B2 (en) | 2017-03-29 | 2021-04-20 | Luminar Holdco, Llc | Non-uniform separation of detector array elements in a lidar system |
US11181622B2 (en) | 2017-03-29 | 2021-11-23 | Luminar, Llc | Method for controlling peak and average power through laser receiver |
US10401481B2 (en) | 2017-03-30 | 2019-09-03 | Luminar Technologies, Inc. | Non-uniform beam power distribution for a laser operating in a vehicle |
US9989629B1 (en) | 2017-03-30 | 2018-06-05 | Luminar Technologies, Inc. | Cross-talk mitigation using wavelength switching |
US10241198B2 (en) | 2017-03-30 | 2019-03-26 | Luminar Technologies, Inc. | Lidar receiver calibration |
US10684360B2 (en) | 2017-03-30 | 2020-06-16 | Luminar Technologies, Inc. | Protecting detector in a lidar system using off-axis illumination |
US10295668B2 (en) | 2017-03-30 | 2019-05-21 | Luminar Technologies, Inc. | Reducing the number of false detections in a lidar system |
JP7014523B2 (en) * | 2017-03-30 | 2022-02-01 | 本田技研工業株式会社 | Autonomous mobile system |
EP3605459A4 (en) * | 2017-03-31 | 2021-01-20 | Pioneer Corporation | Three-dimensional data generation device, three-dimensional data generation method, three-dimensional data generation program, and computer-readable recording medium having three-dimensional data generation program recorded thereon |
US11022688B2 (en) | 2017-03-31 | 2021-06-01 | Luminar, Llc | Multi-eye lidar system |
CN110914705B (en) | 2017-03-31 | 2024-04-26 | 威力登激光雷达美国有限公司 | Devices, systems and methods for integrated LIDAR illumination power control |
US20180284246A1 (en) | 2017-03-31 | 2018-10-04 | Luminar Technologies, Inc. | Using Acoustic Signals to Modify Operation of a Lidar System |
US10556585B1 (en) | 2017-04-13 | 2020-02-11 | Panosense Inc. | Surface normal determination for LIDAR range samples by detecting probe pulse stretching |
US10677897B2 (en) | 2017-04-14 | 2020-06-09 | Luminar Technologies, Inc. | Combining lidar and camera data |
US10392012B2 (en) | 2017-04-24 | 2019-08-27 | Adam Benjamin Tannenbaum | System and method of use for vehicular driving safety |
US10538421B2 (en) | 2017-05-05 | 2020-01-21 | Atlantic Corporation | Systems, devices, and methods for inventory management of carpet rolls in a warehouse |
CN115575928A (en) | 2017-05-08 | 2023-01-06 | 威力登激光雷达美国有限公司 | LIDAR data acquisition and control |
CN107423841A (en) * | 2017-05-09 | 2017-12-01 | 深圳实现创新科技有限公司 | The planning driving path method and system for planning of fire fighting truck |
US10324473B2 (en) * | 2017-06-01 | 2019-06-18 | Wipro Limited | Method and system for generating a safe navigation path for navigating a driverless vehicle |
US11255959B2 (en) * | 2017-06-02 | 2022-02-22 | Sony Corporation | Apparatus, method and computer program for computer vision |
JP6564423B2 (en) * | 2017-06-09 | 2019-08-21 | 株式会社Subaru | Outside environment recognition device and outside environment recognition method |
US10762635B2 (en) | 2017-06-14 | 2020-09-01 | Tusimple, Inc. | System and method for actively selecting and labeling images for semantic segmentation |
US10816970B2 (en) | 2017-06-15 | 2020-10-27 | Aurora Flight Sciences Corporation | System and method for performing an emergency descent and landing |
DE102017114571A1 (en) * | 2017-06-29 | 2019-01-03 | Volkswagen Aktiengesellschaft | Method and device for determining a condition of a surface in the vicinity of a vehicle |
US10386856B2 (en) | 2017-06-29 | 2019-08-20 | Uber Technologies, Inc. | Autonomous vehicle collision mitigation systems and methods |
US10453351B2 (en) | 2017-07-17 | 2019-10-22 | Aurora Flight Sciences Corporation | System and method for detecting obstacles in aerial systems |
US10509415B2 (en) | 2017-07-27 | 2019-12-17 | Aurora Flight Sciences Corporation | Aircrew automation system and method with integrated imaging and force sensing modalities |
KR102218679B1 (en) | 2017-07-28 | 2021-02-23 | 옵시스 테크 엘티디 | VCSEL Array LIDAR Transmitter with Small Angle Divergence |
KR102050632B1 (en) * | 2017-08-02 | 2019-12-03 | 주식회사 에스오에스랩 | Multi-channel LiDAR sensor module |
US10065638B1 (en) | 2017-08-03 | 2018-09-04 | Uber Technologies, Inc. | Multi-model switching on a collision mitigation system |
US10636298B2 (en) | 2017-08-11 | 2020-04-28 | Cubic Corporation | Adaptive traffic control using object tracking and identity details |
US10803740B2 (en) | 2017-08-11 | 2020-10-13 | Cubic Corporation | System and method of navigating vehicles |
US10373489B2 (en) | 2017-08-11 | 2019-08-06 | Cubic Corporation | System and method of adaptive controlling of traffic using camera data |
US10636299B2 (en) | 2017-08-11 | 2020-04-28 | Cubic Corporation | System and method for controlling vehicular traffic |
US11100336B2 (en) | 2017-08-14 | 2021-08-24 | Cubic Corporation | System and method of adaptive traffic management at an intersection |
US10395522B2 (en) | 2017-08-14 | 2019-08-27 | Cubic Corporation | Adaptive traffic optimization using unmanned aerial vehicles |
US10935388B2 (en) | 2017-08-14 | 2021-03-02 | Cubic Corporation | Adaptive optimization of navigational routes using traffic data |
US11250699B2 (en) | 2017-08-14 | 2022-02-15 | Cubic Corporation | System and method of adaptive traffic management at an intersection |
CN107577230B (en) * | 2017-08-16 | 2020-01-14 | 武汉理工大学 | Intelligent collision avoidance system for unmanned ship |
US10746858B2 (en) | 2017-08-17 | 2020-08-18 | Uatc, Llc | Calibration for an autonomous vehicle LIDAR module |
US10775488B2 (en) | 2017-08-17 | 2020-09-15 | Uatc, Llc | Calibration for an autonomous vehicle LIDAR module |
US10503171B2 (en) * | 2017-08-17 | 2019-12-10 | Wipro Limited | Method and system for determining drivable navigation path for an autonomous vehicle |
US10431023B1 (en) * | 2017-08-21 | 2019-10-01 | Uber Technologies, Inc. | Systems and methods to test an autonomous vehicle |
US10816354B2 (en) | 2017-08-22 | 2020-10-27 | Tusimple, Inc. | Verification module system and method for motion-based lane detection with multiple sensors |
US10762673B2 (en) | 2017-08-23 | 2020-09-01 | Tusimple, Inc. | 3D submap reconstruction system and method for centimeter precision localization using camera-based submap and LiDAR-based global map |
US10565457B2 (en) | 2017-08-23 | 2020-02-18 | Tusimple, Inc. | Feature matching and correspondence refinement and 3D submap position refinement system and method for centimeter precision localization using camera-based submap and LiDAR-based global map |
DE102017215307A1 (en) * | 2017-09-01 | 2019-03-07 | Robert Bosch Gmbh | Lidar arrangement and vehicle and robot with such a lidar arrangement |
DE102017215592A1 (en) * | 2017-09-05 | 2019-03-07 | Volkswagen Aktiengesellschaft | Driver assistance device set up for partially autonomous and fully autonomous guidance of a motor vehicle, method and motor vehicle |
US10649458B2 (en) | 2017-09-07 | 2020-05-12 | Tusimple, Inc. | Data-driven prediction-based system and method for trajectory planning of autonomous vehicles |
US10953880B2 (en) | 2017-09-07 | 2021-03-23 | Tusimple, Inc. | System and method for automated lane change control for autonomous vehicles |
US10953881B2 (en) | 2017-09-07 | 2021-03-23 | Tusimple, Inc. | System and method for automated lane change control for autonomous vehicles |
US10671083B2 (en) | 2017-09-13 | 2020-06-02 | Tusimple, Inc. | Neural network architecture system for deep odometry assisted by static scene optical flow |
US10552979B2 (en) | 2017-09-13 | 2020-02-04 | TuSimple | Output of a neural network method for deep odometry assisted by static scene optical flow |
US10679493B2 (en) * | 2017-09-18 | 2020-06-09 | International Business Machines Corporation | Cognitive-based incident response |
US10613200B2 (en) | 2017-09-19 | 2020-04-07 | Veoneer, Inc. | Scanning lidar system and method |
US10838043B2 (en) | 2017-11-15 | 2020-11-17 | Veoneer Us, Inc. | Scanning LiDAR system and method with spatial filtering for reduction of ambient light |
US11460550B2 (en) | 2017-09-19 | 2022-10-04 | Veoneer Us, Llc | Direct detection LiDAR system and method with synthetic doppler processing |
US11194022B2 (en) | 2017-09-29 | 2021-12-07 | Veoneer Us, Inc. | Detection system with reflection member and offset detection array |
US10612932B2 (en) * | 2017-09-29 | 2020-04-07 | Wipro Limited | Method and system for correcting a pre-generated navigation path for an autonomous vehicle |
US10229596B1 (en) * | 2017-10-05 | 2019-03-12 | Analog Devices, Inc. | Systems and methods for measuring a bridge clearance |
US10014390B1 (en) | 2017-10-10 | 2018-07-03 | Globalfoundries Inc. | Inner spacer formation for nanosheet field-effect transistors with tall suspensions |
US20190107622A1 (en) * | 2017-10-11 | 2019-04-11 | Veoneer Us, Inc. | Scanning LiDAR System and Method with Source Laser Beam Splitting Apparatus and Method |
US10473772B2 (en) * | 2017-10-12 | 2019-11-12 | Ford Global Technologies, Llc | Vehicle sensor operation |
CN107831761B (en) * | 2017-10-16 | 2020-07-17 | 中国科学院电工研究所 | Path tracking control method of intelligent vehicle |
IL255050B (en) | 2017-10-16 | 2022-03-01 | Israel Aerospace Ind Ltd | Control over an autonomic vehicle |
US10003168B1 (en) | 2017-10-18 | 2018-06-19 | Luminar Technologies, Inc. | Fiber laser with free-space components |
CN107908104B (en) * | 2017-10-24 | 2020-12-18 | 上海交通大学 | Method and system for UAV sweep trajectory generation and speed planning in isolated island area |
EP3477616A1 (en) | 2017-10-27 | 2019-05-01 | Sigra Technologies GmbH | Method for controlling a vehicle using a machine learning system |
US10967862B2 (en) | 2017-11-07 | 2021-04-06 | Uatc, Llc | Road anomaly detection for autonomous vehicle |
WO2019221776A2 (en) | 2017-11-15 | 2019-11-21 | OPSYS Tech Ltd. | Noise adaptive solid-state lidar system |
US11585901B2 (en) | 2017-11-15 | 2023-02-21 | Veoneer Us, Llc | Scanning lidar system and method with spatial filtering for reduction of ambient light |
CN107745697A (en) | 2017-11-16 | 2018-03-02 | 北京图森未来科技有限公司 | A kind of auto cleaning system and method |
US10451716B2 (en) | 2017-11-22 | 2019-10-22 | Luminar Technologies, Inc. | Monitoring rotation of a mirror in a lidar system |
US10324185B2 (en) * | 2017-11-22 | 2019-06-18 | Luminar Technologies, Inc. | Reducing audio noise in a lidar scanner with a polygon mirror |
EP3717973A4 (en) | 2017-11-28 | 2021-01-06 | Elta Systems Ltd. | FAULT DETECTION IN AN AUTONOMOUS VEHICLE |
US11294041B2 (en) | 2017-12-08 | 2022-04-05 | Velodyne Lidar Usa, Inc. | Systems and methods for improving detection of a return signal in a light ranging and detection system |
KR20190078664A (en) * | 2017-12-11 | 2019-07-05 | 삼성전자주식회사 | Method and apparatus for displaying content |
US10942244B2 (en) | 2017-12-12 | 2021-03-09 | Waymo Llc | Systems and methods for LIDARs with adjustable resolution and failsafe operation |
US11493601B2 (en) | 2017-12-22 | 2022-11-08 | Innovusion, Inc. | High density LIDAR scanning |
US10859389B2 (en) * | 2018-01-03 | 2020-12-08 | Wipro Limited | Method for generation of a safe navigation path for a vehicle and system thereof |
US10684131B2 (en) | 2018-01-04 | 2020-06-16 | Wipro Limited | Method and system for generating and updating vehicle navigation maps with features of navigation paths |
US10967875B2 (en) | 2018-01-05 | 2021-04-06 | Honda Motor Co., Ltd. | Control system for autonomous all-terrain vehicle (ATV) |
JP6548754B2 (en) * | 2018-01-09 | 2019-07-24 | キヤノン株式会社 | Optical apparatus, in-vehicle system and mobile apparatus having the same |
EP3737595B1 (en) | 2018-01-09 | 2023-12-27 | TuSimple, Inc. | Real-time remote control of vehicles with high redundancy |
EP3737968A1 (en) * | 2018-01-11 | 2020-11-18 | Apple Inc. | Architecture for vehicle automation and fail operational automation |
WO2019140277A2 (en) | 2018-01-11 | 2019-07-18 | TuSimple | Monitoring system for autonomous vehicle operation |
CN108256577B (en) * | 2018-01-18 | 2021-09-28 | 东南大学 | Obstacle clustering method based on multi-line laser radar |
CN108270970B (en) | 2018-01-24 | 2020-08-25 | 北京图森智途科技有限公司 | An image acquisition control method and device, and an image acquisition system |
US10914820B2 (en) | 2018-01-31 | 2021-02-09 | Uatc, Llc | Sensor assembly for vehicles |
US12270661B2 (en) | 2018-02-14 | 2025-04-08 | Tusimple, Inc. | Lane marking localization and fusion |
US11009356B2 (en) | 2018-02-14 | 2021-05-18 | Tusimple, Inc. | Lane marking localization and fusion |
US11009365B2 (en) | 2018-02-14 | 2021-05-18 | Tusimple, Inc. | Lane marking localization |
US11808888B2 (en) | 2018-02-23 | 2023-11-07 | Innovusion, Inc. | Multi-wavelength pulse steering in LiDAR systems |
WO2019165294A1 (en) | 2018-02-23 | 2019-08-29 | Innovusion Ireland Limited | 2-dimensional steering system for lidar systems |
US10685244B2 (en) | 2018-02-27 | 2020-06-16 | Tusimple, Inc. | System and method for online real-time multi-object tracking |
JP2019151207A (en) * | 2018-03-02 | 2019-09-12 | 本田技研工業株式会社 | Vehicle control device, vehicle control method, and program |
EP3767367A4 (en) * | 2018-03-13 | 2021-05-12 | Mitsubishi Electric Corporation | Optical scanning device and method of control therefor |
US10901424B2 (en) * | 2018-03-28 | 2021-01-26 | Wipro Limited | Method and system for generating a safe navigation path in real-time for navigating a vehicle |
KR102634880B1 (en) * | 2018-04-01 | 2024-02-08 | 옵시스 테크 엘티디 | Noise adaptive solid-state lidar system |
US10943485B2 (en) * | 2018-04-03 | 2021-03-09 | Baidu Usa Llc | Perception assistant for autonomous driving vehicles (ADVs) |
KR102409522B1 (en) * | 2018-04-05 | 2022-06-16 | 현대자동차주식회사 | SIGNAL COMPRESSION METHOD, SIGNAL COMPRESSION APPARATUS, VEHICLE AND LiDAR SYSTEM |
US10578720B2 (en) | 2018-04-05 | 2020-03-03 | Luminar Technologies, Inc. | Lidar system with a polygon mirror and a noise-reducing feature |
US11029406B2 (en) | 2018-04-06 | 2021-06-08 | Luminar, Llc | Lidar system with AlInAsSb avalanche photodiode |
US10906536B2 (en) | 2018-04-11 | 2021-02-02 | Aurora Innovation, Inc. | Control of autonomous vehicle based on determined yaw parameter(s) of additional vehicle |
CN110378185A (en) | 2018-04-12 | 2019-10-25 | 北京图森未来科技有限公司 | A kind of image processing method applied to automatic driving vehicle, device |
US11126199B2 (en) * | 2018-04-16 | 2021-09-21 | Baidu Usa Llc | Learning based speed planner for autonomous driving vehicles |
US10850397B2 (en) | 2018-04-19 | 2020-12-01 | Aurora Flight Sciences Corporation | System and method for providing in-cockpit actuation of aircraft controls |
US10875662B2 (en) | 2018-04-19 | 2020-12-29 | Aurora Flight Sciences Corporation | Method of robot manipulation in a vibration environment |
US11282385B2 (en) * | 2018-04-24 | 2022-03-22 | Qualcomm Incorproated | System and method of object-based navigation |
CN116129376A (en) | 2018-05-02 | 2023-05-16 | 北京图森未来科技有限公司 | Road edge detection method and device |
US10348051B1 (en) | 2018-05-18 | 2019-07-09 | Luminar Technologies, Inc. | Fiber-optic amplifier |
US10565728B2 (en) | 2018-06-01 | 2020-02-18 | Tusimple, Inc. | Smoothness constraint for camera pose estimation |
US11333740B2 (en) | 2018-06-04 | 2022-05-17 | Uatc, Llc | Determining specular reflectivity characteristics using Lidar |
US10591601B2 (en) | 2018-07-10 | 2020-03-17 | Luminar Technologies, Inc. | Camera-gated lidar system |
CN108764218A (en) * | 2018-07-17 | 2018-11-06 | 江西精骏电控技术有限公司 | Comprehensive pedestrian detecting system |
US10627516B2 (en) * | 2018-07-19 | 2020-04-21 | Luminar Technologies, Inc. | Adjustable pulse characteristics for ground detection in lidar systems |
CN110780665B (en) * | 2018-07-26 | 2022-02-08 | 比亚迪股份有限公司 | Vehicle unmanned control method and device |
US10598791B2 (en) * | 2018-07-31 | 2020-03-24 | Uatc, Llc | Object detection based on Lidar intensity |
US10706724B2 (en) | 2018-08-01 | 2020-07-07 | GM Global Technology Operations LLC | Controlling articulating sensors of an autonomous vehicle |
JP2021532368A (en) | 2018-08-03 | 2021-11-25 | オプシス テック リミテッド | Distributed modular solid-state lidar system |
US11112795B1 (en) | 2018-08-07 | 2021-09-07 | GM Global Technology Operations LLC | Maneuver-based interaction system for an autonomous vehicle |
US10559198B1 (en) | 2018-08-08 | 2020-02-11 | Cubic Corporation | System and method of adaptive controlling of traffic using zone based occupancy |
US10551501B1 (en) | 2018-08-09 | 2020-02-04 | Luminar Technologies, Inc. | Dual-mode lidar system |
US11399137B2 (en) | 2018-08-10 | 2022-07-26 | Aurora Flight Sciences Corporation | Object-tracking system |
US10782691B2 (en) | 2018-08-10 | 2020-09-22 | Buffalo Automation Group Inc. | Deep learning and intelligent sensing system integration |
WO2020033967A1 (en) | 2018-08-10 | 2020-02-13 | Buffalo Automation Group Inc. | Training a deep learning system for maritime applications |
US10340651B1 (en) | 2018-08-21 | 2019-07-02 | Luminar Technologies, Inc. | Lidar system with optical trigger |
US11971507B2 (en) | 2018-08-24 | 2024-04-30 | Velodyne Lidar Usa, Inc. | Systems and methods for mitigating optical crosstalk in a light ranging and detection system |
CN109191487B (en) * | 2018-08-30 | 2022-03-25 | 阿波罗智能技术(北京)有限公司 | Unmanned vehicle-based collision detection method, device, equipment and storage medium |
JP7014106B2 (en) * | 2018-09-04 | 2022-02-01 | トヨタ自動車株式会社 | Vehicle control system |
US11023742B2 (en) | 2018-09-07 | 2021-06-01 | Tusimple, Inc. | Rear-facing perception system for vehicles |
US11019274B2 (en) | 2018-09-10 | 2021-05-25 | Tusimple, Inc. | Adaptive illumination for a time-of-flight camera on a vehicle |
CN112689586B (en) | 2018-09-13 | 2024-04-16 | 图森有限公司 | Remote safe driving method and system |
US10712434B2 (en) | 2018-09-18 | 2020-07-14 | Velodyne Lidar, Inc. | Multi-channel LIDAR illumination driver |
CN109358618B (en) * | 2018-09-28 | 2021-10-26 | 安徽工程大学 | Path planning method of mobile robot |
US10474930B1 (en) * | 2018-10-05 | 2019-11-12 | StradVision, Inc. | Learning method and testing method for monitoring blind spot of vehicle, and learning device and testing device using the same |
US11692811B2 (en) * | 2018-10-08 | 2023-07-04 | Faro Technologies, Inc. | System and method of defining a path and scanning an environment |
US10989532B2 (en) | 2018-10-08 | 2021-04-27 | Faro Technologies, Inc. | System and method of defining a path and scanning an environment |
US10914569B2 (en) * | 2018-10-08 | 2021-02-09 | Faro Technologies, Inc. | System and method of defining a path and scanning an environment |
US11037453B2 (en) | 2018-10-12 | 2021-06-15 | Aurora Flight Sciences Corporation | Adaptive sense and avoid system |
US11151810B2 (en) | 2018-10-12 | 2021-10-19 | Aurora Flight Sciences Corporation | Adaptable vehicle monitoring system |
GB2578108A (en) * | 2018-10-15 | 2020-04-22 | Atlantic Inertial Systems Ltd | A navigation system |
US20200133308A1 (en) * | 2018-10-18 | 2020-04-30 | Cartica Ai Ltd | Vehicle to vehicle (v2v) communication less truck platooning |
US10976747B2 (en) * | 2018-10-29 | 2021-04-13 | Here Global B.V. | Method and apparatus for generating a representation of an environment |
US10942271B2 (en) | 2018-10-30 | 2021-03-09 | Tusimple, Inc. | Determining an angle between a tow vehicle and a trailer |
US11092458B2 (en) * | 2018-10-30 | 2021-08-17 | Telenav, Inc. | Navigation system with operation obstacle alert mechanism and method of operation thereof |
US11536845B2 (en) | 2018-10-31 | 2022-12-27 | Waymo Llc | LIDAR systems with multi-faceted mirrors |
JP6838028B2 (en) * | 2018-10-31 | 2021-03-03 | ファナック株式会社 | Automatic program correction device and automatic program correction method |
US11829143B2 (en) | 2018-11-02 | 2023-11-28 | Aurora Operations, Inc. | Labeling autonomous vehicle data |
US11256263B2 (en) | 2018-11-02 | 2022-02-22 | Aurora Operations, Inc. | Generating targeted training instances for autonomous vehicles |
US11403492B2 (en) | 2018-11-02 | 2022-08-02 | Aurora Operations, Inc. | Generating labeled training instances for autonomous vehicles |
US11163312B2 (en) * | 2018-11-02 | 2021-11-02 | Aurora Operations, Inc. | Removable automotive LIDAR data collection POD |
US11086319B2 (en) | 2018-11-02 | 2021-08-10 | Aurora Operations, Inc. | Generating testing instances for autonomous vehicles |
US11209821B2 (en) | 2018-11-02 | 2021-12-28 | Aurora Operations, Inc. | Labeling autonomous vehicle data |
US11082010B2 (en) | 2018-11-06 | 2021-08-03 | Velodyne Lidar Usa, Inc. | Systems and methods for TIA base current detection and compensation |
TWI674984B (en) * | 2018-11-15 | 2019-10-21 | 財團法人車輛研究測試中心 | Driving track planning system and method for self-driving vehicles |
US11119192B2 (en) * | 2018-12-07 | 2021-09-14 | Here Global B.V. | Automatic detection of overhead obstructions |
CN111366938B (en) | 2018-12-10 | 2023-03-14 | 北京图森智途科技有限公司 | Trailer included angle measuring method and device and vehicle |
US11221399B2 (en) * | 2018-12-12 | 2022-01-11 | Waymo Llc | Detecting spurious objects for autonomous vehicles |
CN111319629B (en) | 2018-12-14 | 2021-07-16 | 北京图森智途科技有限公司 | A method, device and system for forming an autonomous vehicle fleet |
CN109460035B (en) * | 2018-12-18 | 2021-10-15 | 国家海洋局北海海洋工程勘察研究院(青岛环海海洋工程勘察研究院) | A two-level autonomous obstacle avoidance method for unmanned boats at high speed |
US11460578B2 (en) * | 2018-12-28 | 2022-10-04 | Robert Bosch Gmbh | 3D lidar sensing unit with oscillating axis of rotation for autonomous driving |
EP3889721B1 (en) * | 2018-12-29 | 2024-08-07 | Great Wall Motor Company Limited | Obstacle avoidance method and system during automatic driving of vehicle, and vehicle |
US11885958B2 (en) | 2019-01-07 | 2024-01-30 | Velodyne Lidar Usa, Inc. | Systems and methods for a dual axis resonant scanning mirror |
US12061263B2 (en) | 2019-01-07 | 2024-08-13 | Velodyne Lidar Usa, Inc. | Systems and methods for a configurable sensor system |
CN109753073A (en) * | 2019-01-25 | 2019-05-14 | 温州大学 | An unmanned intelligent vehicle speed management system |
US11500075B2 (en) * | 2019-02-01 | 2022-11-15 | Zoox, Inc. | Identifying and/or removing ghost detections from lidar sensor output |
CA3226839A1 (en) | 2019-02-01 | 2020-08-06 | Crown Equipment Corporation | On-board charging station for a remote control device |
US11641121B2 (en) | 2019-02-01 | 2023-05-02 | Crown Equipment Corporation | On-board charging station for a remote control device |
US11774561B2 (en) | 2019-02-08 | 2023-10-03 | Luminar Technologies, Inc. | Amplifier input protection circuits |
JP7167751B2 (en) * | 2019-02-08 | 2022-11-09 | 株式会社アイシン | object detector |
DE102019204260A1 (en) * | 2019-03-27 | 2020-10-01 | Zf Friedrichshafen Ag | Controlling a motor vehicle |
JP7535313B2 (en) | 2019-04-09 | 2024-08-16 | オプシス テック リミテッド | Solid-state LIDAR transmitter with laser control |
CN110208819A (en) * | 2019-05-14 | 2019-09-06 | 江苏大学 | A kind of processing method of multiple barrier three-dimensional laser radar data |
JP7276437B2 (en) * | 2019-05-23 | 2023-05-18 | 日本電信電話株式会社 | 3D point cloud label learning and estimating device, 3D point cloud label learning and estimating method, and 3D point cloud label learning and estimating program |
CN113906316A (en) | 2019-05-30 | 2022-01-07 | 欧普赛斯技术有限公司 | Eye-safe long-range LIDAR system using actuators |
US11823460B2 (en) | 2019-06-14 | 2023-11-21 | Tusimple, Inc. | Image fusion for autonomous vehicle operation |
US11227490B2 (en) | 2019-06-18 | 2022-01-18 | Toyota Motor North America, Inc. | Identifying changes in the condition of a transport |
US20200402149A1 (en) | 2019-06-18 | 2020-12-24 | Toyota Motor North America, Inc. | Identifying changes in the condition of a transport |
WO2020263735A1 (en) | 2019-06-25 | 2020-12-30 | OPSYS Tech Ltd. | Adaptive multiple-pulse lidar system |
US10613203B1 (en) | 2019-07-01 | 2020-04-07 | Velodyne Lidar, Inc. | Interference mitigation for light detection and ranging |
US11300677B2 (en) * | 2019-07-08 | 2022-04-12 | GM Global Technology Operations LLC | Automated driving systems and control logic for host vehicle velocity estimation using wide aperture radar |
US11579257B2 (en) | 2019-07-15 | 2023-02-14 | Veoneer Us, Llc | Scanning LiDAR system and method with unitary optical element |
US11474218B2 (en) | 2019-07-15 | 2022-10-18 | Veoneer Us, Llc | Scanning LiDAR system and method with unitary optical element |
EP3966654B1 (en) * | 2019-07-31 | 2024-07-31 | Hewlett-Packard Development Company, L.P. | Positioning autonomous vehicles |
WO2021021872A1 (en) | 2019-07-31 | 2021-02-04 | OPSYS Tech Ltd. | High-resolution solid-state lidar transmitter |
CN110386230B (en) * | 2019-08-20 | 2024-08-13 | 广州拓浪智能应急科技有限公司 | Barrier device and system are kept away to radar |
CN112630799B (en) * | 2019-09-24 | 2022-11-29 | 阿波罗智能技术(北京)有限公司 | Method and apparatus for outputting information |
US11341781B2 (en) * | 2019-10-18 | 2022-05-24 | Toyota Motor Engineering And Manufacturing North America, Inc. | Vehicular communications through identifiers and online systems |
US11313969B2 (en) | 2019-10-28 | 2022-04-26 | Veoneer Us, Inc. | LiDAR homodyne transceiver using pulse-position modulation |
US11320830B2 (en) | 2019-10-28 | 2022-05-03 | Deere & Company | Probabilistic decision support for obstacle detection and classification in a working area |
CN113580155B (en) * | 2019-10-31 | 2023-07-18 | 灵动科技(北京)有限公司 | Autonomous mobile device and warehouse logistics system |
US11385058B2 (en) * | 2019-11-26 | 2022-07-12 | Toyota Motor Engineering & Manufacturing North America, Inc. | Systems, vehicles, and methods for detecting and mapping off-road obstacles |
US11614742B2 (en) * | 2019-11-27 | 2023-03-28 | Zoox, Inc. | Height estimation using sensor data |
US11592575B2 (en) | 2019-12-20 | 2023-02-28 | Waymo Llc | Sensor steering for multi-directional long-range perception |
US11544936B2 (en) * | 2019-12-20 | 2023-01-03 | Zoox, Inc. | In-path obstacle detection and avoidance system |
CN114489044B (en) * | 2019-12-31 | 2025-01-17 | 深圳引望智能技术有限公司 | Track planning method and device |
US10732261B1 (en) | 2019-12-31 | 2020-08-04 | Aurora Innovation, Inc. | Generating data using radar observation model based on machine learning |
US11351961B2 (en) * | 2020-01-29 | 2022-06-07 | Ford Global Technologies, Llc | Proximity-based vehicle security systems and methods |
US11520009B2 (en) * | 2020-02-05 | 2022-12-06 | Caterpillar Inc. | Method and system for detecting an obstacle |
US11538343B2 (en) | 2020-03-23 | 2022-12-27 | Toyota Motor North America, Inc. | Automatic warning of atypical audio indicating a transport event |
US11443624B2 (en) | 2020-03-23 | 2022-09-13 | Toyota Motor North America, Inc. | Automatic warning of navigating towards a dangerous area or event |
EP3893150A1 (en) | 2020-04-09 | 2021-10-13 | Tusimple, Inc. | Camera pose estimation techniques |
JP7327257B2 (en) * | 2020-04-13 | 2023-08-16 | トヨタ自動車株式会社 | Automotive sensor system |
FR3109452B1 (en) * | 2020-04-16 | 2022-03-11 | Renault Sas | Method for detecting a traffic jam situation on a motor vehicle. |
CA3116108A1 (en) * | 2020-04-26 | 2021-10-26 | Potential Motors Inc. | Electric vehicle control system |
WO2021231984A1 (en) * | 2020-05-14 | 2021-11-18 | Raven Industries, Inc. | Obstacle monitoring systems and methods for same |
US11847919B2 (en) | 2020-05-19 | 2023-12-19 | Toyota Motor North America, Inc. | Control of transport en route |
AU2021203567A1 (en) | 2020-06-18 | 2022-01-20 | Tusimple, Inc. | Angle and orientation measurements for vehicles with multiple drivable sections |
US12135565B2 (en) | 2020-06-26 | 2024-11-05 | Tusimple, Inc. | Adaptive sensor control |
CN113928307A (en) | 2020-06-29 | 2022-01-14 | 图森有限公司 | Automatic parking technology |
US11879745B2 (en) * | 2020-07-20 | 2024-01-23 | Ford Global Technologies, Llc | Systems and methods for facilitating a final leg of a journey |
EP4185924A4 (en) | 2020-07-21 | 2024-08-14 | Leddartech Inc. | BEAM STEERING DEVICE, ESPECIALLY FOR LIDAR SYSTEMS |
CA3210182A1 (en) | 2020-07-21 | 2021-10-06 | Leddartech Inc. | Beam-steering devices and methods for lidar applications |
EP4185894A4 (en) | 2020-07-21 | 2024-07-24 | Leddartech Inc. | SYSTEMS AND METHODS FOR WIDE-ANGLE LIDAR WITH NON-UNIFORM MAGNIFICATION OPTICS |
CN111950420A (en) * | 2020-08-04 | 2020-11-17 | 国以贤智能科技(上海)股份有限公司 | Obstacle avoidance method, device, equipment and storage medium |
CN116057491A (en) | 2020-08-11 | 2023-05-02 | 克朗设备公司 | Remote control device |
AU2021355468B2 (en) * | 2020-09-30 | 2024-05-02 | Autonomous Solutions, Inc. | Shared obstacles in autonomous vehicle systems |
US12181878B2 (en) * | 2020-10-22 | 2024-12-31 | Waymo Llc | Velocity estimation and object tracking for autonomous vehicle applications |
US11841439B2 (en) | 2020-11-02 | 2023-12-12 | Waymo Llc | Point cloud segmentation using a coherent lidar for autonomous vehicle applications |
US12233905B2 (en) | 2020-11-02 | 2025-02-25 | Waymo Llc | Classification of objects based on motion patterns for autonomous vehicle applications |
US12050267B2 (en) | 2020-11-09 | 2024-07-30 | Waymo Llc | Doppler-assisted object mapping for autonomous vehicle applications |
US11483070B2 (en) * | 2020-12-04 | 2022-10-25 | Eric Clifton Roberts | Systems, methods, and devices for infrared communications |
US11656629B1 (en) | 2020-12-08 | 2023-05-23 | Waymo Llc | Detection of particulate matter in autonomous vehicle applications |
US11427226B1 (en) * | 2020-12-15 | 2022-08-30 | Waymo Llc | Positional gaps for driver controllability |
US12044800B2 (en) | 2021-01-14 | 2024-07-23 | Magna Electronics, Llc | Scanning LiDAR system and method with compensation for transmit laser pulse effects |
US11733369B2 (en) | 2021-02-11 | 2023-08-22 | Waymo Llc | Methods and systems for three dimensional object detection and localization |
US11326758B1 (en) | 2021-03-12 | 2022-05-10 | Veoneer Us, Inc. | Spotlight illumination system using optical element |
CN113064444B (en) * | 2021-03-15 | 2022-09-09 | 中国南方电网有限责任公司超高压输电公司天生桥局 | Method, device, device and storage medium for predicting next flight position of unmanned aerial vehicle |
CN113176777B (en) * | 2021-03-16 | 2023-11-03 | 深圳市人工智能与机器人研究院 | Automatic sailing ship running monitoring method, device, equipment and readable storage medium |
US12216474B1 (en) | 2021-05-04 | 2025-02-04 | Waymo Llc | Vibrometry-based behavior prediction for autonomous vehicle applications |
US11732858B2 (en) | 2021-06-18 | 2023-08-22 | Veoneer Us, Llc | Headlight illumination system using optical element |
CN113665590B (en) * | 2021-09-24 | 2024-04-19 | 阿波罗智联(北京)科技有限公司 | Method, device, equipment, medium and vehicle for controlling vehicle |
DE102021125525A1 (en) * | 2021-10-01 | 2023-04-06 | Deere & Company | Method for issuing collision warnings for an agricultural vehicle combination |
US20230128651A1 (en) * | 2021-10-22 | 2023-04-27 | Mobile Industrial Robots A/S | Determining scanner error |
CN114194210A (en) * | 2021-11-19 | 2022-03-18 | 岚图汽车科技有限公司 | A highway pilot system, control method and device thereof |
US11851052B2 (en) * | 2022-01-31 | 2023-12-26 | Ford Global Technologies, Llc | Vehicle path verification |
CN114577233B (en) * | 2022-05-05 | 2022-07-29 | 腾讯科技(深圳)有限公司 | A vehicle navigation method, device, computer equipment, and storage medium |
WO2024035714A1 (en) * | 2022-08-09 | 2024-02-15 | Pete Bitar | Compact and lightweight drone delivery device called an arcspear electric jet drone system having an electric ducted air propulsion system and being relatively difficult to track in flight |
US12092278B2 (en) | 2022-10-07 | 2024-09-17 | Magna Electronics, Llc | Generating a spotlight |
US12228653B2 (en) | 2022-10-07 | 2025-02-18 | Magna Electronics, Llc | Integrating a sensing system into headlight optics |
US12202396B1 (en) | 2023-12-19 | 2025-01-21 | Magna Electronics, Llc | Line-scan-gated imaging for LiDAR headlight |
CN118493408B (en) * | 2024-07-17 | 2024-09-13 | 中科云谷科技有限公司 | Control method, device and system for mechanical arm based on self-adaptive controller and storage medium |
Citations (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5821860A (en) * | 1996-05-20 | 1998-10-13 | Honda Giken Kogyo Kabushiki Kaisha | Driving condition-monitoring apparatus for automotive vehicles |
US5928299A (en) * | 1995-02-22 | 1999-07-27 | Honda Giken Kogyo Kabushiki Kaisha | Vehicle control system using navigation system steering |
US20030095046A1 (en) * | 2001-11-19 | 2003-05-22 | Volvo Trucks North America, Inc. | System for ensuring driver competency |
WO2004101306A1 (en) * | 2003-05-16 | 2004-11-25 | Daimlerchrysler Ag | Method and device for influencing the demands on a driver in a motor vehicle |
US6870478B2 (en) * | 2001-08-28 | 2005-03-22 | Pioneer Corporation | Information providing system and information providing method |
US7088250B2 (en) * | 2002-08-29 | 2006-08-08 | Pioneer Corporation | Apparatus and method for estimating fatigue level |
US20060220915A1 (en) * | 2005-03-21 | 2006-10-05 | Bauer James A | Inter-vehicle drowsy driver advisory system |
US20060267781A1 (en) * | 2005-05-24 | 2006-11-30 | Coulter Jeffery R | Process and method for safer vehicle navigation through facial gesture recognition and operator condition monitoring |
US20070063855A1 (en) * | 2002-03-08 | 2007-03-22 | Alexander Maass | Method and device for warning a driver |
US20070168128A1 (en) * | 2004-01-28 | 2007-07-19 | Setsuo Tokoro | Running support system for vehicle |
US7254439B2 (en) * | 2004-01-06 | 2007-08-07 | Monebo Technologies, Inc. | Method and system for contactless evaluation of fatigue of an operator |
US7304580B2 (en) * | 2003-12-04 | 2007-12-04 | Hoana Medical, Inc. | Intelligent medical vigilance system |
US7526333B2 (en) * | 2002-08-30 | 2009-04-28 | Pioneer Corporation | Sound reproduction system and method based on physical and mental states of a drive |
USRE42471E1 (en) * | 1996-08-19 | 2011-06-21 | Torch William C | System and method for monitoring eye movement |
US8096946B2 (en) * | 1999-01-27 | 2012-01-17 | Compumedics Limited | Vigilance monitoring system |
US8164464B2 (en) * | 2005-03-24 | 2012-04-24 | Matos Jeffrey A | Method and system of aircraft pilot assessment |
Family Cites Families (44)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4342997A (en) * | 1980-07-03 | 1982-08-03 | Westinghouse Electric Corp. | Array modification that adds height capability to a 2D array radar |
US4649390A (en) * | 1983-08-05 | 1987-03-10 | Hughes Aircraft Company | Two dimension radar system with selectable three dimension target data extraction |
US4646244A (en) * | 1984-02-02 | 1987-02-24 | Sundstrand Data Control, Inc. | Terrain advisory system |
US5557282A (en) * | 1988-10-11 | 1996-09-17 | Itt Corporation | Height finding antenna apparatus and method of operation |
US5202695A (en) * | 1990-09-27 | 1993-04-13 | Sperry Marine Inc. | Orientation stabilization by software simulated stabilized platform |
NL9002717A (en) * | 1990-12-11 | 1992-07-01 | Hollandse Signaalapparaten Bv | RADAR SYSTEM. |
US5243553A (en) * | 1991-07-02 | 1993-09-07 | Loral Vought Systems Corporation | Gate array pulse capture device |
US5200606A (en) * | 1991-07-02 | 1993-04-06 | Ltv Missiles And Electronics Group | Laser radar scanning system |
US5198819A (en) * | 1992-03-02 | 1993-03-30 | Thermwood Corporation | Weather radar display system |
JPH05265547A (en) * | 1992-03-23 | 1993-10-15 | Fuji Heavy Ind Ltd | On-vehicle outside monitoring device |
IL104542A (en) * | 1993-01-28 | 1996-05-14 | Israel State | Airborne obstacle collision avoidance apparatus |
US5424823A (en) * | 1993-08-17 | 1995-06-13 | Loral Vought Systems Corporation | System for identifying flat orthogonal objects using reflected energy signals |
US5793310A (en) * | 1994-02-04 | 1998-08-11 | Nissan Motor Co., Ltd. | Portable or vehicular navigating apparatus and method capable of displaying bird's eye view |
US5644386A (en) * | 1995-01-11 | 1997-07-01 | Loral Vought Systems Corp. | Visual recognition system for LADAR sensors |
US6720920B2 (en) * | 1997-10-22 | 2004-04-13 | Intelligent Technologies International Inc. | Method and arrangement for communicating between vehicles |
US7085637B2 (en) * | 1997-10-22 | 2006-08-01 | Intelligent Technologies International, Inc. | Method and system for controlling a vehicle |
US7202776B2 (en) * | 1997-10-22 | 2007-04-10 | Intelligent Technologies International, Inc. | Method and system for detecting objects external to a vehicle |
JPH09142236A (en) * | 1995-11-17 | 1997-06-03 | Mitsubishi Electric Corp | Periphery monitoring method and device for vehicle, and trouble deciding method and device for periphery monitoring device |
JP3619628B2 (en) * | 1996-12-19 | 2005-02-09 | 株式会社日立製作所 | Driving environment recognition device |
US5870181A (en) * | 1997-10-28 | 1999-02-09 | Alliant Defense Electronics Systems, Inc. | Acoustic optical scanning of linear detector array for laser radar |
AUPP299498A0 (en) * | 1998-04-15 | 1998-05-07 | Commonwealth Scientific And Industrial Research Organisation | Method of tracking and sensing position of objects |
US7068211B2 (en) * | 2000-02-08 | 2006-06-27 | Cambridge Consultants Limited | Methods and apparatus for obtaining positional information |
JP4233723B2 (en) * | 2000-02-28 | 2009-03-04 | 本田技研工業株式会社 | Obstacle detection device, obstacle detection method, and recording medium recording an obstacle detection program |
US6664529B2 (en) * | 2000-07-19 | 2003-12-16 | Utah State University | 3D multispectral lidar |
FR2817339B1 (en) * | 2000-11-24 | 2004-05-14 | Mensi | THREE-DIMENSIONAL LIFTING DEVICE OF A LASER EMISSION SCENE |
US20030069693A1 (en) * | 2001-01-16 | 2003-04-10 | Snapp Douglas N. | Geographic pointing device |
US6844924B2 (en) * | 2001-06-29 | 2005-01-18 | The United States Of America As Represented By The Secretary Of The Army | Ladar system for detecting objects |
US6727807B2 (en) * | 2001-12-14 | 2004-04-27 | Koninklijke Philips Electronics N.V. | Driver's aid using image processing |
US20030140775A1 (en) * | 2002-01-30 | 2003-07-31 | Stewart John R. | Method and apparatus for sighting and targeting a controlled system from a common three-dimensional data set |
US7248342B1 (en) * | 2003-02-14 | 2007-07-24 | United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration | Three-dimension imaging lidar |
US6810321B1 (en) * | 2003-03-17 | 2004-10-26 | Sprint Communications Company L.P. | Vehicle traffic monitoring using cellular telephone location and velocity data |
US7076365B2 (en) * | 2003-05-12 | 2006-07-11 | Circumnav Networks, Inc. | Enhanced dead reckoning method |
JP3915742B2 (en) * | 2003-06-20 | 2007-05-16 | 株式会社デンソー | Vehicle object recognition device |
NZ545627A (en) * | 2003-08-29 | 2009-02-28 | Drs Sustainment Systems Inc | Electronically programmable actively damped sensor mount |
US7818127B1 (en) * | 2004-06-18 | 2010-10-19 | Geneva Aerospace, Inc. | Collision avoidance for vehicle control systems |
US7266447B2 (en) * | 2004-06-30 | 2007-09-04 | Navteq North America, Llc. | Method of collecting information for a geographic database for use with a navigation system |
JP2006151125A (en) * | 2004-11-26 | 2006-06-15 | Omron Corp | On-vehicle image processing device |
US20060197867A1 (en) * | 2005-03-02 | 2006-09-07 | Peter Johnson | Imaging head and imaging system |
US7301497B2 (en) * | 2005-04-05 | 2007-11-27 | Eastman Kodak Company | Stereo display for position sensing systems |
US7250902B2 (en) * | 2005-07-19 | 2007-07-31 | Raytheon Company | Method of generating accurate estimates of azimuth and elevation angles of a target for a phased—phased array rotating radar |
AU2006304838B2 (en) * | 2005-10-21 | 2011-07-28 | Deere & Company | Systems and methods for obstacle avoidance |
CN101688774A (en) * | 2006-07-13 | 2010-03-31 | 威力登音响公司 | High definition lidar system |
US20110040427A1 (en) * | 2007-10-31 | 2011-02-17 | Pinhas Ben-Tzvi | Hybrid mobile robot |
US8126642B2 (en) | 2008-10-24 | 2012-02-28 | Gray & Company, Inc. | Control and systems for autonomously driven vehicles |
-
2006
- 2006-03-16 US US11/376,160 patent/US8050863B2/en active Active
-
2007
- 2007-03-16 WO PCT/US2007/064128 patent/WO2008048707A2/en active Application Filing
-
2011
- 2011-09-22 US US13/240,576 patent/US8346480B2/en active Active
-
2012
- 2012-12-14 US US13/715,586 patent/US20130131908A1/en not_active Abandoned
Patent Citations (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5928299A (en) * | 1995-02-22 | 1999-07-27 | Honda Giken Kogyo Kabushiki Kaisha | Vehicle control system using navigation system steering |
US5821860A (en) * | 1996-05-20 | 1998-10-13 | Honda Giken Kogyo Kabushiki Kaisha | Driving condition-monitoring apparatus for automotive vehicles |
USRE42471E1 (en) * | 1996-08-19 | 2011-06-21 | Torch William C | System and method for monitoring eye movement |
US8096946B2 (en) * | 1999-01-27 | 2012-01-17 | Compumedics Limited | Vigilance monitoring system |
US6870478B2 (en) * | 2001-08-28 | 2005-03-22 | Pioneer Corporation | Information providing system and information providing method |
US20030095046A1 (en) * | 2001-11-19 | 2003-05-22 | Volvo Trucks North America, Inc. | System for ensuring driver competency |
US20070063855A1 (en) * | 2002-03-08 | 2007-03-22 | Alexander Maass | Method and device for warning a driver |
US7088250B2 (en) * | 2002-08-29 | 2006-08-08 | Pioneer Corporation | Apparatus and method for estimating fatigue level |
US7526333B2 (en) * | 2002-08-30 | 2009-04-28 | Pioneer Corporation | Sound reproduction system and method based on physical and mental states of a drive |
US20070182529A1 (en) * | 2003-05-16 | 2007-08-09 | Daimlerchrysler Ag | Method and apparatus for influencing the load of a driver in a motor vehicle |
WO2004101306A1 (en) * | 2003-05-16 | 2004-11-25 | Daimlerchrysler Ag | Method and device for influencing the demands on a driver in a motor vehicle |
US7304580B2 (en) * | 2003-12-04 | 2007-12-04 | Hoana Medical, Inc. | Intelligent medical vigilance system |
US7254439B2 (en) * | 2004-01-06 | 2007-08-07 | Monebo Technologies, Inc. | Method and system for contactless evaluation of fatigue of an operator |
US20070168128A1 (en) * | 2004-01-28 | 2007-07-19 | Setsuo Tokoro | Running support system for vehicle |
US20060220915A1 (en) * | 2005-03-21 | 2006-10-05 | Bauer James A | Inter-vehicle drowsy driver advisory system |
US8164464B2 (en) * | 2005-03-24 | 2012-04-24 | Matos Jeffrey A | Method and system of aircraft pilot assessment |
US20060267781A1 (en) * | 2005-05-24 | 2006-11-30 | Coulter Jeffery R | Process and method for safer vehicle navigation through facial gesture recognition and operator condition monitoring |
Cited By (70)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8706394B2 (en) | 2008-10-24 | 2014-04-22 | Gray & Company, Inc. | Control and systems for autonomously driven vehicles |
US8744661B2 (en) * | 2009-10-21 | 2014-06-03 | Berthold K. P. Horn | Method and apparatus for reducing motor vehicle traffic flow instabilities and increasing vehicle throughput |
US20110093177A1 (en) * | 2009-10-21 | 2011-04-21 | Horn Berthold K P | Method and Apparatus for Reducing Motor Vehicle Traffic Flow Instabilities and Increasing Vehicle Throughput |
US8849483B2 (en) * | 2011-04-13 | 2014-09-30 | California Institute Of Technology | Target trailing with safe navigation with colregs for maritime autonomous surface vehicles |
US20120265380A1 (en) * | 2011-04-13 | 2012-10-18 | California Institute Of Technology | Target Trailing with Safe Navigation with colregs for Maritime Autonomous Surface Vehicles |
US9047770B2 (en) * | 2012-04-06 | 2015-06-02 | Thales | Method for determining an instantaneous or anticipated probable zone of occupancy of an aircraft in an airport navigation zone |
US20140236470A1 (en) * | 2012-04-06 | 2014-08-21 | Thales | Method for determining an instantaneous or anticipated probable zone of occupancy of an aircraft in an airport navigation zone |
US9043069B1 (en) * | 2012-11-07 | 2015-05-26 | Google Inc. | Methods and systems for scan matching approaches for vehicle heading estimation |
US9043072B1 (en) * | 2013-04-04 | 2015-05-26 | Google Inc. | Methods and systems for correcting an estimated heading using a map |
US20150088373A1 (en) * | 2013-09-23 | 2015-03-26 | The Boeing Company | Optical communications and obstacle sensing for autonomous vehicles |
US9784835B1 (en) * | 2013-09-27 | 2017-10-10 | Waymo Llc | Laser diode timing feedback using trace loop |
US11460575B1 (en) | 2013-09-27 | 2022-10-04 | Waymo Llc | Laser diode timing feedback using trace loop |
US12248067B1 (en) | 2013-09-27 | 2025-03-11 | Waymo Llc | Laser diode timing feedback using trace loop |
US10007000B1 (en) | 2013-09-27 | 2018-06-26 | Waymo Llc | Laser diode timing feedback using trace loop |
US10732647B2 (en) | 2013-11-27 | 2020-08-04 | The Trustees Of The University Of Pennsylvania | Multi-sensor fusion for robust autonomous flight in indoor and outdoor environments with a rotorcraft micro-aerial vehicle (MAV) |
US9804600B1 (en) | 2014-07-24 | 2017-10-31 | Google Inc. | Systems and methods for ground plane estimation |
US11921508B2 (en) | 2014-07-24 | 2024-03-05 | Boston Dynamics, Inc. | Systems and methods for ground plane estimation |
US9259838B1 (en) * | 2014-07-24 | 2016-02-16 | Google Inc. | Systems and methods for ground plane estimation |
US10496094B1 (en) | 2014-07-24 | 2019-12-03 | Boston Dynamics, Inc. | Systems and methods for ground plane estimation |
US11287819B2 (en) | 2014-07-24 | 2022-03-29 | Boston Dynamics, Inc. | Systems and methods for ground plane estimation |
US10395115B2 (en) | 2015-01-27 | 2019-08-27 | The Trustees Of The University Of Pennsylvania | Systems, devices, and methods for robotic remote sensing for precision agriculture |
US10991247B2 (en) | 2015-02-06 | 2021-04-27 | Aptiv Technologies Limited | Method of automatically controlling an autonomous vehicle based on electronic messages from roadside infrastructure or other vehicles |
WO2016126316A1 (en) * | 2015-02-06 | 2016-08-11 | Delphi Technologies, Inc. | Autonomous guidance system |
US11763670B2 (en) | 2015-02-06 | 2023-09-19 | Aptiv Technologies Limited | Method of automatically controlling an autonomous vehicle based on electronic messages from roadside infrastructure or other vehicles |
US11543832B2 (en) | 2015-02-06 | 2023-01-03 | Aptiv Technologies Limited | Method and apparatus for controlling an autonomous vehicle |
US10948924B2 (en) | 2015-02-06 | 2021-03-16 | Aptiv Technologies Limited | Method and apparatus for controlling an autonomous vehicle |
US20180004221A1 (en) * | 2015-02-06 | 2018-01-04 | Delphi Technologies, Inc. | Autonomous guidance system |
US10037028B2 (en) | 2015-07-24 | 2018-07-31 | The Trustees Of The University Of Pennsylvania | Systems, devices, and methods for on-board sensing and control of micro aerial vehicles |
US20190072980A1 (en) * | 2015-09-11 | 2019-03-07 | The Trustees Of The University Of Pennsylvania | Systems and methods for generating safe trajectories for multi-vehicle teams |
WO2017095493A3 (en) * | 2015-09-11 | 2017-07-20 | The Trustees Of The University Of Pennsylvania | Systems and methods for generating safe trajectories for multi-vehicle teams |
US10884430B2 (en) | 2015-09-11 | 2021-01-05 | The Trustees Of The University Of Pennsylvania | Systems and methods for generating safe trajectories for multi-vehicle teams |
US20170110022A1 (en) * | 2015-10-14 | 2017-04-20 | Toyota Motor Engineering & Manufacturing North America, Inc. | Assessing driver readiness for transition between operational modes of an autonomous vehicle |
US9786192B2 (en) * | 2015-10-14 | 2017-10-10 | Toyota Motor Engineering & Manufacturing North America, Inc. | Assessing driver readiness for transition between operational modes of an autonomous vehicle |
US11022974B2 (en) * | 2015-11-04 | 2021-06-01 | Zoox, Inc. | Sensor-based object-detection optimization for autonomous vehicles |
US9720415B2 (en) * | 2015-11-04 | 2017-08-01 | Zoox, Inc. | Sensor-based object-detection optimization for autonomous vehicles |
US20170351261A1 (en) * | 2015-11-04 | 2017-12-07 | Zoox, Inc. | Sensor-Based Object-Detection Optimization For Autonomous Vehicles |
US11099559B2 (en) * | 2016-05-19 | 2021-08-24 | Denso Corporation | Automated driving system and automated driving switch determination program product |
US20190072957A1 (en) * | 2016-05-19 | 2019-03-07 | Denso Corporation | Automated driving system and automated driving switch determination program product |
US11860305B2 (en) | 2016-10-17 | 2024-01-02 | Waymo Llc | Light detection and ranging (LIDAR) device having multiple receivers |
US10379540B2 (en) | 2016-10-17 | 2019-08-13 | Waymo Llc | Light detection and ranging (LIDAR) device having multiple receivers |
US10496890B2 (en) | 2016-10-28 | 2019-12-03 | International Business Machines Corporation | Vehicular collaboration for vehicular blind spot detection |
CN106774312A (en) * | 2016-12-05 | 2017-05-31 | 遨博(北京)智能科技有限公司 | A kind of method and apparatus that motion track is determined for mobile robot |
WO2018134296A1 (en) * | 2017-01-19 | 2018-07-26 | Robert Bosch Gmbh | System and method of using crowd-sourced driving path data in an autonmous or semi-autonomous driving system |
US10209715B2 (en) | 2017-01-19 | 2019-02-19 | Robert Bosch Gmbh | System and method of using crowd-sourced driving path data in an autonomous or semi-autonomous driving system |
US10338594B2 (en) | 2017-03-13 | 2019-07-02 | Nio Usa, Inc. | Navigation of autonomous vehicles to enhance safety under one or more fault conditions |
US10423162B2 (en) | 2017-05-08 | 2019-09-24 | Nio Usa, Inc. | Autonomous vehicle logic to identify permissioned parking relative to multiple classes of restricted parking |
US11790551B2 (en) | 2017-06-06 | 2023-10-17 | Plusai, Inc. | Method and system for object centric stereo in autonomous driving vehicles |
WO2018224874A1 (en) * | 2017-06-06 | 2018-12-13 | PlusAI Corp | Method and system for distributed learning and adaptation in autonomous driving vehicles |
US12093821B2 (en) | 2017-06-06 | 2024-09-17 | Plusai, Inc. | Method and system for closed loop perception in autonomous driving vehicles |
US11042155B2 (en) | 2017-06-06 | 2021-06-22 | Plusai Limited | Method and system for closed loop perception in autonomous driving vehicles |
US12039445B2 (en) | 2017-06-06 | 2024-07-16 | Plusai, Inc. | Method and system for on-the-fly object labeling via cross modality validation in autonomous driving vehicles |
US11573573B2 (en) | 2017-06-06 | 2023-02-07 | Plusai, Inc. | Method and system for distributed learning and adaptation in autonomous driving vehicles |
US11392133B2 (en) | 2017-06-06 | 2022-07-19 | Plusai, Inc. | Method and system for object centric stereo in autonomous driving vehicles |
US11435750B2 (en) | 2017-06-06 | 2022-09-06 | Plusai, Inc. | Method and system for object centric stereo via cross modality validation in autonomous driving vehicles |
US11550334B2 (en) | 2017-06-06 | 2023-01-10 | Plusai, Inc. | Method and system for integrated global and distributed learning in autonomous driving vehicles |
US11537126B2 (en) | 2017-06-06 | 2022-12-27 | Plusai, Inc. | Method and system for on-the-fly object labeling via cross modality validation in autonomous driving vehicles |
US10369974B2 (en) | 2017-07-14 | 2019-08-06 | Nio Usa, Inc. | Control and coordination of driverless fuel replenishment for autonomous vehicles |
US10710633B2 (en) | 2017-07-14 | 2020-07-14 | Nio Usa, Inc. | Control of complex parking maneuvers and autonomous fuel replenishment of driverless vehicles |
US20200208993A1 (en) * | 2017-07-31 | 2020-07-02 | Israel Aerospace Industries Ltd. | Path planning within a traversed area |
US11519737B2 (en) * | 2017-07-31 | 2022-12-06 | Elta Systems Ltd. | Path planning within a traversed area |
CN107578427A (en) * | 2017-07-31 | 2018-01-12 | 深圳市易成自动驾驶技术有限公司 | Detection method, device and the computer-readable recording medium of dynamic barrier |
US20190041859A1 (en) * | 2017-08-04 | 2019-02-07 | Aptiv Technologies Limited | Sensor failure compensation system for an automated vehicle |
WO2019084130A1 (en) * | 2017-10-25 | 2019-05-02 | Bae Systems Information And Electronic Systems Integration Inc. | Method and system of digital light processing and light detection and ranging for guided autonomous vehicles |
US12093042B2 (en) | 2018-01-16 | 2024-09-17 | Nio Usa, Inc. | Event data recordation to identify and resolve anomalies associated with control of driverless vehicles |
US11022971B2 (en) | 2018-01-16 | 2021-06-01 | Nio Usa, Inc. | Event data recordation to identify and resolve anomalies associated with control of driverless vehicles |
US20190353783A1 (en) * | 2018-05-18 | 2019-11-21 | Koito Manufacturing Co., Ltd. | Sensor system |
US10955553B2 (en) * | 2018-05-18 | 2021-03-23 | Koito Manufacturing Co., Ltd. | Sensor system for compensating information processing rate |
CN110320919A (en) * | 2019-07-31 | 2019-10-11 | 河海大学常州校区 | A kind of circulating robot method for optimizing route in unknown geographical environment |
EP4300237A4 (en) * | 2021-02-23 | 2024-12-18 | Jingdong Technology Information Technology Co., Ltd. | Robot traveling method and apparatus, and electronic device, storage medium and program product |
US20240061119A1 (en) * | 2022-08-16 | 2024-02-22 | Qualcomm Incorporated | Cellular radio frequency (rf) sensing in automobile navigation |
Also Published As
Publication number | Publication date |
---|---|
US8050863B2 (en) | 2011-11-01 |
US20070219720A1 (en) | 2007-09-20 |
WO2008048707A3 (en) | 2008-07-31 |
US20120035788A1 (en) | 2012-02-09 |
US8346480B2 (en) | 2013-01-01 |
WO2008048707A2 (en) | 2008-04-24 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8346480B2 (en) | Navigation and control system for autonomous vehicles | |
EP2338029B1 (en) | Control and systems for autonomously driven vehicles | |
EP0936518B1 (en) | Integrated vehicle positioning and navigation system, apparatus and method | |
US5170352A (en) | Multi-purpose autonomous vehicle with path plotting | |
Bohren et al. | Little ben: The ben franklin racing team's entry in the 2007 DARPA urban challenge | |
US5555503A (en) | System and method for providing accurate vehicle positioning using spatial bias techniques | |
US5629855A (en) | System and method for using parabolic models to improve position estimates from a global positioning system | |
US5438517A (en) | Vehicle position determination system and method | |
US20080059015A1 (en) | Software architecture for high-speed traversal of prescribed routes | |
Ng et al. | Vehicle following with obstacle avoidance capabilities in natural environments | |
AU679321B2 (en) | System and method for detecting obstacles in a road | |
Spofford et al. | Stopping distance analysis of ladar and stereo for unmanned ground vehicles |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |