US20060087416A1 - Information providing device for vehicle - Google Patents
Information providing device for vehicle Download PDFInfo
- Publication number
- US20060087416A1 US20060087416A1 US11/232,914 US23291405A US2006087416A1 US 20060087416 A1 US20060087416 A1 US 20060087416A1 US 23291405 A US23291405 A US 23291405A US 2006087416 A1 US2006087416 A1 US 2006087416A1
- Authority
- US
- United States
- Prior art keywords
- vehicle
- information
- unit
- visual information
- providing
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 230000000007 visual effect Effects 0.000 claims abstract description 46
- 238000000034 method Methods 0.000 claims description 4
- 238000001514 detection method Methods 0.000 description 8
- 230000001133 acceleration Effects 0.000 description 3
- 230000037007 arousal Effects 0.000 description 3
- 230000001678 irradiating effect Effects 0.000 description 3
- 238000012544 monitoring process Methods 0.000 description 3
- 239000003550 marker Substances 0.000 description 2
- 230000007423 decrease Effects 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 239000000725 suspension Substances 0.000 description 1
Images
Classifications
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60Q—ARRANGEMENT OF SIGNALLING OR LIGHTING DEVICES, THE MOUNTING OR SUPPORTING THEREOF OR CIRCUITS THEREFOR, FOR VEHICLES IN GENERAL
- B60Q1/00—Arrangement of optical signalling or lighting devices, the mounting or supporting thereof or circuits therefor
- B60Q1/26—Arrangement of optical signalling or lighting devices, the mounting or supporting thereof or circuits therefor the devices being primarily intended to indicate the vehicle, or parts thereof, or to give signals, to other traffic
- B60Q1/50—Arrangement of optical signalling or lighting devices, the mounting or supporting thereof or circuits therefor the devices being primarily intended to indicate the vehicle, or parts thereof, or to give signals, to other traffic for indicating other intentions or conditions, e.g. request for waiting or overtaking
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60Q—ARRANGEMENT OF SIGNALLING OR LIGHTING DEVICES, THE MOUNTING OR SUPPORTING THEREOF OR CIRCUITS THEREFOR, FOR VEHICLES IN GENERAL
- B60Q2400/00—Special features or arrangements of exterior signal lamps for vehicles
- B60Q2400/50—Projected symbol or information, e.g. onto the road or car body
Definitions
- the present invention relates to an information providing device for a vehicle.
- JP-H06-255399-A a conventional device is described which warns a vehicle's driver of an dangerous object by illuminating the dangerous object using a spotlight or by projecting a warning sign to a position which is on the windshield of the vehicle and corresponds to a position of the dangerous object.
- the driver moves his/her sight line to the warning sign and adjusts his/her focus to the warning sign on the windshield to notice presence of the dangerous object. Subsequently, the driver adjusts his/her focus from the windshield to the dangerous object itself to estimate its distance from the vehicle.
- the driver needs to adjust his/her focus from the windshield to an actual position of a dangerous object in the process of noticing the dangerous object.
- the device projects the spotlight only when it detects a dangerous object. Therefore, the driver possibly has difficulty in immediately reacting to the spotlight which comes up at an unexpected position. Moreover, a sight line of the driver and a heading direction of the vehicle are almost the same; therefore, the driver possibly fails to recognize the spotlight if the spotlight deviates from the heading direction.
- the driver sometimes has difficulty in recognizing a dangerous object even in a case where the driver needs to react against the dangerous object as quickly as possible.
- an information providing device in a vehicle is provided with the following.
- An object detecting unit is included for detecting an object outside the vehicle.
- a providing unit is included for providing, outside the vehicle, visual information which is visually perceptible.
- a controlling unit is included for controlling the providing unit.
- the providing unit is usually caused to provide the visual information to indicate a heading direction of the vehicle.
- the providing unit is caused to move the visual information to indicate a direction towards the detected object.
- the information providing device can usually lead a driver's sight line to the heading direction of the vehicle.
- driver's eyes can focus on the heading direction, which is proper for usual driving operation.
- the visual information is moved while continuing to be perceptible so that the detected object can be indicated by the moved visual information.
- the driver watching the heading direction notices the movement of the visual information and the sight line of the driver is led to the detected object by following the movement. Consequently, this helps prevent a delay in movement of the sight line or an oversight of the dangerous object.
- an amount of adjustment of the focus in this case using the device of the present invention is small compared to that in a case using the conventional device where the focus moves from the windshield of the vehicle to the dangerous object outside the vehicle.
- FIG. 1 shows a structure of an information providing device for a vehicle
- FIG. 2 shows a diagram of units included in a controller in the information providing device
- FIG. 3 shows a situation where the information providing device projects visual information of an estimated path for a subject vehicle
- FIG. 4 shows a situation where the information providing device projects visual information of danger information towards a dangerous object
- FIG. 5 shows a flowchart of a process which the controller executes.
- an information providing device 100 mounted in a subject vehicle includes surrounding information detectors 10 , driving operation state detectors 20 , vehicle running state detectors 30 , a controller 40 , and a laser projection device 50 .
- the surrounding information detectors 10 include a laser or radar sensor, an image sensor, an ultrasonic sensor, and a navigation device. Using the detectors 10 , the controller 40 detects characteristics of an object located outside the vehicle, specifically, frontward, sideward, and rearward of the vehicle. The detected characteristics of the object include, with respect to the object, a kind, a position, a heading direction, or a relative velocity relative to the vehicle.
- the kind of the object is, for example, another vehicle (further, a passenger automobile, a truck, a two-wheeled motorcycle, or the like), a pedestrian, a traffic control device, a road traffic sign, or a lane marker or a border between lanes on a road (e.g., a white line, a yellow line, or a central reserve).
- another vehicle further, a passenger automobile, a truck, a two-wheeled motorcycle, or the like
- a pedestrian e.g., a passenger automobile, a truck, a two-wheeled motorcycle, or the like
- a traffic control device e.g., a pedestrian, a traffic control device, a road traffic sign, or a lane marker or a border between lanes on a road (e.g., a white line, a yellow line, or a central reserve).
- the laser or radar sensor irradiates an area within a predetermined distance outside the vehicle using laser beams or the like and calculates a distance from a vehicle, a velocity relative to the vehicle, and a direction as seen from the vehicle, with respect to the object which reflects the laser beams or the like. Then the laser or radar sensor outputs the calculated values as electric signals to the controller 40 .
- the image sensor includes an optical camera which operates as an imaging unit for photographing an area frontward, sideward, or rearward of the vehicle to take an image. According to the image, the image sensor detects characteristics of an object in the image and outputs the characteristics as electric signals to the controller 40 .
- the characteristics include a kind, a size, and a state (e.g., a lighting color of a traffic light), with respect to the object.
- the ultrasonic sensor sends ultrasonic sound waves to locations surrounding the vehicle.
- the locations are about several meters away from the vehicle.
- the ultrasonic sensor detects a distance between the vehicle and the object which reflects the ultrasonic sound waves, and outputs the detected distance as electric signals to the controller 40 .
- the navigation device displays a map around a current position of the vehicle, finds an optimum route to a destination, and shows the route to a driver of the vehicle.
- the navigation device has a storage device memorizing map data which includes roads data (e.g., data regarding road traffic signs and traffic control devices) and facilities data.
- the roads data includes, for example, data regarding locations or kinds of road signs, data regarding locations of traffic control devices relative to intersections or crosswalks, and data regarding kinds or shapes of traffic control devices.
- the navigation device detects a current position of the vehicle and outputs information such as a kind and a distance of a structure on a road within a predetermined distance from the current position.
- the structure is, for example, a traffic control device and a traffic sign which directs or indicates a vehicle to stop or go.
- the driving operation state detectors 20 include a throttle sensor, a brake sensor, a steering sensor, a shift state sensor, and a turn signal state sensor, which are not illustrated, and outputs detection signals of each sensor to the controller 40 .
- the throttle sensor detects a throttle position depending on an amount of operation of a throttle pedal.
- the brake sensor detects a depression force of a brake pedal.
- the steering sensor detects a direction of operation of a steering wheel and an amount of rotation of the steering wheel relative to the central position.
- the turn signal state sensor detects a direction (i.e., right or left) indicated by a turn signal.
- the vehicle running state detectors 30 include a speed sensor, an acceleration sensor, a yew rate sensor, a tire direction sensor, and a wheel loading sensor, which are not illustrated, and outputs detection signals of each sensor to the controller 40 .
- the speed sensor is a wheel speed sensor which detects speed of the vehicle.
- the acceleration sensor detects acceleration of the vehicle in the anteroposterior direction, the lateral direction, and the vertical direction.
- the tire direction sensor detects a turn angle of a vehicle wheel around a vertical axis against a reference angle that is an angle of the wheel when the vehicle is going in a straight line.
- the turn angle to be detected depends on operation of a steering wheel. Therefore, the turn angle can be also detected by the steering sensor.
- the wheel loading sensor detects a load on each wheel suspension of the vehicle along the vertical direction.
- the laser projection device 50 radiates laser beams to provide or project visual information by irradiating a target position or a target object outside the vehicle.
- the irradiated target position or object can be seen through the projected transparent visual information.
- the driver can view a surrounding area and the visual information as a part of the surrounding area.
- the laser projection device 50 is controlled by the controller 50 as described below.
- the controller 40 is a microcomputer which has a CPU, a RAM, a ROM, and an input/output interface, which are not illustrated.
- the controller 40 includes a surrounding information detection unit 41 , a driving operation information generation unit 42 , a running information generation unit 43 , a driver intention estimation unit 44 , a path estimation unit 45 , a danger determination unit 46 , an object projection calculation unit 47 , and a path projection calculation unit 48 .
- the CPU achieves the above units by executing programs corresponding to the units, by reading the programs from the ROM and using the RAM as a work area.
- the surrounding information detection unit 41 detects surrounding information relating to an object surrounding the vehicle, by comprehensively considering the signals from the surrounding information detectors 10 .
- the object is, for example, another vehicle (a passenger automobile, a truck, a two-wheeled motorcycle, or the like), a pedestrian, a traffic control device, a road traffic sign, or a lane marker or a border between lanes on a road (e.g., a white line, a yellow line, or a central reserve).
- the surrounding information with respect to the object includes a kind, a position, a heading direction, and a relative velocity to the vehicle.
- the detected surrounding information is used by the danger determination unit 46 and the object projection calculation unit 47 .
- the driving operation information generation unit 42 generates, based on the detection signals from the driving operation state detectors 20 , driving operation information which indicates a state of driving operation by the driver.
- the driving operation information is used by the driver intention estimation unit 44 .
- the driver intention estimation unit 44 estimates, based on the generated driving operation information, driver's intention regarding driving operation such as starting, accelerating, changing lanes, turning left/right, decelerating, or stopping.
- Driver intention information which indicates the estimated driver's intention is used by the danger determination unit 46 and the path target calculation unit 48 .
- the running information generation unit 43 generates, based on the detection signals from the vehicle running state detectors 30 , running information which indicates a state of movement of the vehicle.
- the driving operation information is used by the path estimation unit 45 .
- the path estimation unit 45 estimates, based on the generated running information, an estimated path which the vehicle is estimated or expected to follow. In detail, the path estimation unit 45 applies the running information to parameters in a predetermined movement model of the subject vehicle to obtain the estimated path. Information on the estimated path is used by the danger determination unit 46 and the path projection calculation unit 48 .
- the danger determination unit 46 determines whether the detected object is dangerous to advancement of the vehicle and whether the detected structure on a road indicates the vehicle to stop. Such an indication is made, for example, by a stop signal from a traffic control device and a brief stop sign on a road. In addition, the unit 46 determines a danger level (e.g., high or low) of the detected object depending on a distance between the detected object and the vehicle.
- a danger level e.g., high or low
- the determination is executed based on the surrounding information, the driver intention information, and the information on the estimated path. However, the determination can be executed without the driver intention information. The determination can be also executed without the information on the estimated path.
- the unit 46 determines that the object is dangerous to advancement of the vehicle or that the structure indicates the vehicle to stop, it generates danger information which includes a kind, a relative position to the vehicle, and a danger level, with respect to the object or the structure (hereafter dangerous object).
- the danger information is used by the object projection calculation unit 47 .
- the path projection calculation unit 48 calculates path-target positions (e.g., directions and/or distances) to irradiate using laser beams.
- the path-target positions correspond to an estimated path from a current position of the vehicle to a position which the vehicle is estimated to reach after a period (e.g., 3 seconds) by following the estimated path.
- the unit 48 outputs the calculated path-target positions to the laser projection device 50 .
- the laser projection device 50 While the vehicle is running, the laser projection device 50 usually provides or projects an estimated path LL, LR (in FIG. 3 ) of the visual information toward the heading direction of the vehicle as seen from the driver. Therefore, the information providing device 100 can lead the driver's sight line to the path-target positions of the estimated path.
- the path projection calculation unit 48 may change the way of the projection, based on the running information (e.g., a speed of the vehicle and an amount of rotation of the steering wheel), by changing at least one of a direction, a length, and a shape, with respect to the projection.
- the running information e.g., a speed of the vehicle and an amount of rotation of the steering wheel
- the driver generally has difficulty in properly maintaining his/her sight line or focus when his/her arousal level is reduced.
- characterless and unchanging visual information possibly decreases his/her arousal level further and causes his/her lack of attention.
- the diver possibly overlooks information necessary for driving the vehicle, or fails to react to the information in time.
- the change of the way of the projection based on the running state of the vehicle gives variations to the visual information, and possibly suppresses the reduction of the arousal level.
- the object projection calculation unit 47 calculates object-target positions (e.g., directions and/or distances) that enable visual information as the danger information to indicate the dangerous object itself and a direction towards a position of the dangerous object, when the unit 46 generates the danger information.
- the object projection calculation unit 47 moves the estimated path of the visual information from the lines LL, LR in FIG. 3 to lines LT, LT in FIG. 4 that correspond to the calculated object-target positions.
- the object projection calculation unit 47 also determines the way of projection so that the laser projection device 50 can provide the visual information corresponding to the kind of the dangerous object. For example, when the danger information indicates that the dangerous object is a nearby vehicle or a pedestrian, the unit 47 determines the way of projection so that the projected visual information as the danger information includes a shape of an exclamation mark in addition to the lines LT, LT, as seen in FIG. 4 .
- the controller 40 operates as shown in FIG. 5 as a whole. First, the controller 40 detects the surrounding information using the surrounding information detection unit 41 at Step S 10 .
- Step S 20 vehicle states of the driving operation state and the vehicle running state are detected using the driving operation information generation unit 42 and the running information generation unit 43 , respectively.
- Step S 30 vehicle information of the driving operation information and the running information is then generated based on the detections of Step S 20 also using the driving operation information generation unit 42 and the running information generation unit 43 , respectively.
- Step S 40 the driver's intention is estimated based on the driving operation information using the driver intention estimation unit 44 .
- Step S 50 the estimated path is estimated based on the running information using the path estimation unit 45 .
- Step S 60 the path-target positions are calculated for providing the visual information regarding the estimated path using the path projection calculation unit 48 .
- Step S 70 the detected object or structure is determined to be the dangerous object or not; namely, it is determined whether the detected object is dangerous to advancement of the vehicle and whether the detected structure on a road indicates the vehicle to stop, using the danger determination unit 46 .
- Step S 80 it is determined whether there is a dangerous object outside the vehicle also using the danger determination unit 46 .
- Step S 110 takes place.
- the controller 40 causes the laser projection device 50 to project the estimated path of the lines LL, LR towards the heading direction of the vehicle which is calculated at Step S 60 , as shown in FIG. 3 .
- Step S 100 the object-target positions and the way of the projection are calculated for providing the visual information of the danger information regarding the dangerous object using the object projection calculation unit 47 .
- Step S 110 takes place.
- the controller 40 causes the laser projection device 50 to move the visual information of the estimated path from the lines LL, LR shown in FIG. 3 to the lines LT, LT including the exclamation mark shown in FIG. 4 , based on the object-target positions and way of the projection calculated at Step 100 .
- Steps S 10 to S 110 the information providing device 100 continuously projects the estimated path as the visual information toward the heading direction of the vehicle while a dangerous object is not around the vehicle. Once the dangerous object appears, the information providing device 100 starts to move the projected estimated path from the positions of the lines LL, LR in FIG. 3 to the positions of the lines LT, LT in FIG. 4 that is approaching the dangerous object. During this movement, the lines continue being projected on a relevant road.
- the driver can smoothly follow the movement of the lines of the estimated path as the visual information. Therefore, the delay of movement of the driver's sight line can be suppressed, and the oversight of the dangerous object can be reduced.
- the information providing device 100 may project visual information on a lane maker between two lanes of a road when the vehicle is about to deviate from one of the lanes to the other.
- the device 100 can draw the driver's attention to the situation.
- the information providing device 100 may have a monitoring unit for monitoring a sight line of the driver.
- the device 100 when the device 100 detects, using the monitoring unit, that the driver is looking aside, it can provide the visual information such as an exclamation mark by irradiating the location which the driver is looking at.
- the device 100 can urge the driver to gaze at a proper direction for driving.
- the device 100 may retrieve a route, which the vehicle should go along, from the navigation device and provide visual information indicating a direction along the route by irradiating positions in the direction.
- the device 100 can provide the information of the route to the driver.
- the device 100 may project visual information on an approaching vehicle to notify the approaching vehicle of presence of the vehicle where the device 100 is mounted.
- the device 100 may project visual information on a pedestrian who is determined to be not a dangerous object but to be present in front of the vehicle. Thus, the device 100 can notify the pedestrian of presence of the approaching vehicle.
- the device 100 may project visual information indicating an intention of changing to an adjacent lane, in conjunction with operation of a turn signal switch and an amount of rotation of the steering wheel.
- the device 100 can notify following vehicles in the adjacent lane of the intention of changing lanes of the vehicle where the device 100 is mounted.
- the device 100 does not need to always project visual information toward the heading direction while a dangerous object is not present around the vehicle.
- the device 100 may need to project the visual information toward the heading direction at a certain period just before the projected visual information starts to move to the detected dangerous object.
Landscapes
- Engineering & Computer Science (AREA)
- Mechanical Engineering (AREA)
- Traffic Control Systems (AREA)
Abstract
An information providing device provided in a vehicle includes an object detecting unit for detecting an object outside the vehicle; a providing unit for providing, outside of the vehicle, visual information which is visually perceptible; and a controlling unit for controlling the providing unit. The providing unit is usually caused to provide the visual information to indicate a heading direction of the vehicle. In contrast, when the detected object is determined to include a danger to advancement of the vehicle or an indication of stopping the vehicle, the providing unit is caused to then move the visual information to indicate a direction towards the detected object.
Description
- This application is based on and incorporates herein by reference Japanese patent application No. 2004-312849 filed on Oct. 27, 2004.
- The present invention relates to an information providing device for a vehicle.
- In JP-H06-255399-A, a conventional device is described which warns a vehicle's driver of an dangerous object by illuminating the dangerous object using a spotlight or by projecting a warning sign to a position which is on the windshield of the vehicle and corresponds to a position of the dangerous object.
- When the warning sign is shown on the windshield, the driver moves his/her sight line to the warning sign and adjusts his/her focus to the warning sign on the windshield to notice presence of the dangerous object. Subsequently, the driver adjusts his/her focus from the windshield to the dangerous object itself to estimate its distance from the vehicle. Thus, in a case where a warning sign is shown on the windshield, the driver needs to adjust his/her focus from the windshield to an actual position of a dangerous object in the process of noticing the dangerous object.
- In addition, the device projects the spotlight only when it detects a dangerous object. Therefore, the driver possibly has difficulty in immediately reacting to the spotlight which comes up at an unexpected position. Moreover, a sight line of the driver and a heading direction of the vehicle are almost the same; therefore, the driver possibly fails to recognize the spotlight if the spotlight deviates from the heading direction.
- Thus, using the conventional device, the driver sometimes has difficulty in recognizing a dangerous object even in a case where the driver needs to react against the dangerous object as quickly as possible.
- It is therefore an object of the present invention to provide an information providing device which enables a driver to quickly react to a dangerous object.
- To achieve the above object, an information providing device in a vehicle is provided with the following. An object detecting unit is included for detecting an object outside the vehicle. A providing unit is included for providing, outside the vehicle, visual information which is visually perceptible. Furthermore, a controlling unit is included for controlling the providing unit. Here, the providing unit is usually caused to provide the visual information to indicate a heading direction of the vehicle. In contrast, when the detected object is determined to include a danger to advancement of the vehicle or an indication of stopping the vehicle, the providing unit is caused to move the visual information to indicate a direction towards the detected object.
- Under the above structure, the information providing device can usually lead a driver's sight line to the heading direction of the vehicle. Thus, driver's eyes can focus on the heading direction, which is proper for usual driving operation. In addition, when the detected object is determined to be requiring driver's attention, the visual information is moved while continuing to be perceptible so that the detected object can be indicated by the moved visual information.
- Thus, the driver watching the heading direction notices the movement of the visual information and the sight line of the driver is led to the detected object by following the movement. Consequently, this helps prevent a delay in movement of the sight line or an oversight of the dangerous object.
- Furthermore, when the visual information moves from the heading direction to the direction towards the detected dangerous object, the focus of the driver changes also from the heading direction to the direction towards the detected object, within scenery outside of the vehicle. Therefore, an amount of adjustment of the focus in this case using the device of the present invention is small compared to that in a case using the conventional device where the focus moves from the windshield of the vehicle to the dangerous object outside the vehicle.
- The above and other objects, features, and advantages of the present invention will become more apparent from the following detailed description made with reference to the accompanying drawings. In the drawings:
-
FIG. 1 shows a structure of an information providing device for a vehicle; -
FIG. 2 shows a diagram of units included in a controller in the information providing device; -
FIG. 3 shows a situation where the information providing device projects visual information of an estimated path for a subject vehicle; -
FIG. 4 shows a situation where the information providing device projects visual information of danger information towards a dangerous object; and -
FIG. 5 shows a flowchart of a process which the controller executes. - As shown in
FIG. 1 , aninformation providing device 100 mounted in a subject vehicle includes surroundinginformation detectors 10, drivingoperation state detectors 20, vehicle runningstate detectors 30, acontroller 40, and alaser projection device 50. - The surrounding
information detectors 10 include a laser or radar sensor, an image sensor, an ultrasonic sensor, and a navigation device. Using thedetectors 10, thecontroller 40 detects characteristics of an object located outside the vehicle, specifically, frontward, sideward, and rearward of the vehicle. The detected characteristics of the object include, with respect to the object, a kind, a position, a heading direction, or a relative velocity relative to the vehicle. The kind of the object is, for example, another vehicle (further, a passenger automobile, a truck, a two-wheeled motorcycle, or the like), a pedestrian, a traffic control device, a road traffic sign, or a lane marker or a border between lanes on a road (e.g., a white line, a yellow line, or a central reserve). - The laser or radar sensor irradiates an area within a predetermined distance outside the vehicle using laser beams or the like and calculates a distance from a vehicle, a velocity relative to the vehicle, and a direction as seen from the vehicle, with respect to the object which reflects the laser beams or the like. Then the laser or radar sensor outputs the calculated values as electric signals to the
controller 40. - The image sensor includes an optical camera which operates as an imaging unit for photographing an area frontward, sideward, or rearward of the vehicle to take an image. According to the image, the image sensor detects characteristics of an object in the image and outputs the characteristics as electric signals to the
controller 40. The characteristics include a kind, a size, and a state (e.g., a lighting color of a traffic light), with respect to the object. - The ultrasonic sensor sends ultrasonic sound waves to locations surrounding the vehicle. The locations are about several meters away from the vehicle. Then the ultrasonic sensor detects a distance between the vehicle and the object which reflects the ultrasonic sound waves, and outputs the detected distance as electric signals to the
controller 40. - The navigation device displays a map around a current position of the vehicle, finds an optimum route to a destination, and shows the route to a driver of the vehicle.
- In addition, the navigation device has a storage device memorizing map data which includes roads data (e.g., data regarding road traffic signs and traffic control devices) and facilities data. The roads data includes, for example, data regarding locations or kinds of road signs, data regarding locations of traffic control devices relative to intersections or crosswalks, and data regarding kinds or shapes of traffic control devices.
- Moreover, the navigation device detects a current position of the vehicle and outputs information such as a kind and a distance of a structure on a road within a predetermined distance from the current position. The structure is, for example, a traffic control device and a traffic sign which directs or indicates a vehicle to stop or go.
- The driving
operation state detectors 20 include a throttle sensor, a brake sensor, a steering sensor, a shift state sensor, and a turn signal state sensor, which are not illustrated, and outputs detection signals of each sensor to thecontroller 40. - The throttle sensor detects a throttle position depending on an amount of operation of a throttle pedal. The brake sensor detects a depression force of a brake pedal. The steering sensor detects a direction of operation of a steering wheel and an amount of rotation of the steering wheel relative to the central position. The turn signal state sensor detects a direction (i.e., right or left) indicated by a turn signal.
- The vehicle running
state detectors 30 include a speed sensor, an acceleration sensor, a yew rate sensor, a tire direction sensor, and a wheel loading sensor, which are not illustrated, and outputs detection signals of each sensor to thecontroller 40. - The speed sensor is a wheel speed sensor which detects speed of the vehicle. The acceleration sensor detects acceleration of the vehicle in the anteroposterior direction, the lateral direction, and the vertical direction. The tire direction sensor detects a turn angle of a vehicle wheel around a vertical axis against a reference angle that is an angle of the wheel when the vehicle is going in a straight line. The turn angle to be detected depends on operation of a steering wheel. Therefore, the turn angle can be also detected by the steering sensor. The wheel loading sensor detects a load on each wheel suspension of the vehicle along the vertical direction.
- The
laser projection device 50 radiates laser beams to provide or project visual information by irradiating a target position or a target object outside the vehicle. Here, the irradiated target position or object can be seen through the projected transparent visual information. Thus, the driver can view a surrounding area and the visual information as a part of the surrounding area. Thelaser projection device 50 is controlled by thecontroller 50 as described below. - The
controller 40 is a microcomputer which has a CPU, a RAM, a ROM, and an input/output interface, which are not illustrated. - As shown in
FIG. 2 , thecontroller 40 includes a surroundinginformation detection unit 41, a driving operationinformation generation unit 42, a runninginformation generation unit 43, a driverintention estimation unit 44, apath estimation unit 45, adanger determination unit 46, an objectprojection calculation unit 47, and a pathprojection calculation unit 48. Here, the CPU achieves the above units by executing programs corresponding to the units, by reading the programs from the ROM and using the RAM as a work area. - The surrounding
information detection unit 41 detects surrounding information relating to an object surrounding the vehicle, by comprehensively considering the signals from the surroundinginformation detectors 10. Here, the object is, for example, another vehicle (a passenger automobile, a truck, a two-wheeled motorcycle, or the like), a pedestrian, a traffic control device, a road traffic sign, or a lane marker or a border between lanes on a road (e.g., a white line, a yellow line, or a central reserve). The surrounding information with respect to the object includes a kind, a position, a heading direction, and a relative velocity to the vehicle. The detected surrounding information is used by thedanger determination unit 46 and the objectprojection calculation unit 47. - The driving operation
information generation unit 42 generates, based on the detection signals from the drivingoperation state detectors 20, driving operation information which indicates a state of driving operation by the driver. The driving operation information is used by the driverintention estimation unit 44. - The driver
intention estimation unit 44 estimates, based on the generated driving operation information, driver's intention regarding driving operation such as starting, accelerating, changing lanes, turning left/right, decelerating, or stopping. Driver intention information which indicates the estimated driver's intention is used by thedanger determination unit 46 and the pathtarget calculation unit 48. - The running
information generation unit 43 generates, based on the detection signals from the vehicle runningstate detectors 30, running information which indicates a state of movement of the vehicle. The driving operation information is used by thepath estimation unit 45. - The
path estimation unit 45 estimates, based on the generated running information, an estimated path which the vehicle is estimated or expected to follow. In detail, thepath estimation unit 45 applies the running information to parameters in a predetermined movement model of the subject vehicle to obtain the estimated path. Information on the estimated path is used by thedanger determination unit 46 and the pathprojection calculation unit 48. - The
danger determination unit 46 determines whether the detected object is dangerous to advancement of the vehicle and whether the detected structure on a road indicates the vehicle to stop. Such an indication is made, for example, by a stop signal from a traffic control device and a brief stop sign on a road. In addition, theunit 46 determines a danger level (e.g., high or low) of the detected object depending on a distance between the detected object and the vehicle. - The determination is executed based on the surrounding information, the driver intention information, and the information on the estimated path. However, the determination can be executed without the driver intention information. The determination can be also executed without the information on the estimated path.
- When the
unit 46 determines that the object is dangerous to advancement of the vehicle or that the structure indicates the vehicle to stop, it generates danger information which includes a kind, a relative position to the vehicle, and a danger level, with respect to the object or the structure (hereafter dangerous object). The danger information is used by the objectprojection calculation unit 47. - The path
projection calculation unit 48 calculates path-target positions (e.g., directions and/or distances) to irradiate using laser beams. The path-target positions correspond to an estimated path from a current position of the vehicle to a position which the vehicle is estimated to reach after a period (e.g., 3 seconds) by following the estimated path. Theunit 48 outputs the calculated path-target positions to thelaser projection device 50. - While the vehicle is running, the
laser projection device 50 usually provides or projects an estimated path LL, LR (inFIG. 3 ) of the visual information toward the heading direction of the vehicle as seen from the driver. Therefore, theinformation providing device 100 can lead the driver's sight line to the path-target positions of the estimated path. - In addition, the path
projection calculation unit 48 may change the way of the projection, based on the running information (e.g., a speed of the vehicle and an amount of rotation of the steering wheel), by changing at least one of a direction, a length, and a shape, with respect to the projection. - The driver generally has difficulty in properly maintaining his/her sight line or focus when his/her arousal level is reduced. In this case, characterless and unchanging visual information possibly decreases his/her arousal level further and causes his/her lack of attention. As a result, the diver possibly overlooks information necessary for driving the vehicle, or fails to react to the information in time.
- Therefore, the change of the way of the projection based on the running state of the vehicle gives variations to the visual information, and possibly suppresses the reduction of the arousal level.
- The object
projection calculation unit 47 calculates object-target positions (e.g., directions and/or distances) that enable visual information as the danger information to indicate the dangerous object itself and a direction towards a position of the dangerous object, when theunit 46 generates the danger information. The objectprojection calculation unit 47 moves the estimated path of the visual information from the lines LL, LR inFIG. 3 to lines LT, LT inFIG. 4 that correspond to the calculated object-target positions. - The object
projection calculation unit 47 also determines the way of projection so that thelaser projection device 50 can provide the visual information corresponding to the kind of the dangerous object. For example, when the danger information indicates that the dangerous object is a nearby vehicle or a pedestrian, theunit 47 determines the way of projection so that the projected visual information as the danger information includes a shape of an exclamation mark in addition to the lines LT, LT, as seen inFIG. 4 . - The
controller 40 operates as shown inFIG. 5 as a whole. First, thecontroller 40 detects the surrounding information using the surroundinginformation detection unit 41 at Step S10. - At Step S20, vehicle states of the driving operation state and the vehicle running state are detected using the driving operation
information generation unit 42 and the runninginformation generation unit 43, respectively. At Step S30, vehicle information of the driving operation information and the running information is then generated based on the detections of Step S20 also using the driving operationinformation generation unit 42 and the runninginformation generation unit 43, respectively. - Subsequently, at Step S40, the driver's intention is estimated based on the driving operation information using the driver
intention estimation unit 44. Step S50, the estimated path is estimated based on the running information using thepath estimation unit 45. At Step S60, the path-target positions are calculated for providing the visual information regarding the estimated path using the pathprojection calculation unit 48. - Then, at Step S70, the detected object or structure is determined to be the dangerous object or not; namely, it is determined whether the detected object is dangerous to advancement of the vehicle and whether the detected structure on a road indicates the vehicle to stop, using the
danger determination unit 46. Then, at Step S80, it is determined whether there is a dangerous object outside the vehicle also using thedanger determination unit 46. - When the determination of Step S80 is negated, Step S110 takes place. Here, the
controller 40 causes thelaser projection device 50 to project the estimated path of the lines LL, LR towards the heading direction of the vehicle which is calculated at Step S60, as shown inFIG. 3 . - In contrast, when the determination at Step S80 is affirmed, the danger information is generated at Step S90. At Step S100, the object-target positions and the way of the projection are calculated for providing the visual information of the danger information regarding the dangerous object using the object
projection calculation unit 47. After Step S100, Step S110 takes place. Here, thecontroller 40 causes thelaser projection device 50 to move the visual information of the estimated path from the lines LL, LR shown inFIG. 3 to the lines LT, LT including the exclamation mark shown inFIG. 4 , based on the object-target positions and way of the projection calculated atStep 100. - By executing Steps S10 to S110, the
information providing device 100 continuously projects the estimated path as the visual information toward the heading direction of the vehicle while a dangerous object is not around the vehicle. Once the dangerous object appears, theinformation providing device 100 starts to move the projected estimated path from the positions of the lines LL, LR inFIG. 3 to the positions of the lines LT, LT inFIG. 4 that is approaching the dangerous object. During this movement, the lines continue being projected on a relevant road. - Thus, the driver can smoothly follow the movement of the lines of the estimated path as the visual information. Therefore, the delay of movement of the driver's sight line can be suppressed, and the oversight of the dangerous object can be reduced.
- The present invention should not be limited to the embodiment discussed above and shown in the figures, but may be implemented in various ways without departing from the spirit of the invention.
- For example, the
information providing device 100 may project visual information on a lane maker between two lanes of a road when the vehicle is about to deviate from one of the lanes to the other. Thus, thedevice 100 can draw the driver's attention to the situation. - In addition, the
information providing device 100 may have a monitoring unit for monitoring a sight line of the driver. In this case, when thedevice 100 detects, using the monitoring unit, that the driver is looking aside, it can provide the visual information such as an exclamation mark by irradiating the location which the driver is looking at. Thus, thedevice 100 can urge the driver to gaze at a proper direction for driving. - In addition, the
device 100 may retrieve a route, which the vehicle should go along, from the navigation device and provide visual information indicating a direction along the route by irradiating positions in the direction. Thus, thedevice 100 can provide the information of the route to the driver. - In addition, when the vehicle is about to enter a blind intersection, the
device 100 may project visual information on an approaching vehicle to notify the approaching vehicle of presence of the vehicle where thedevice 100 is mounted. - In addition, the
device 100 may project visual information on a pedestrian who is determined to be not a dangerous object but to be present in front of the vehicle. Thus, thedevice 100 can notify the pedestrian of presence of the approaching vehicle. - In addition, the
device 100 may project visual information indicating an intention of changing to an adjacent lane, in conjunction with operation of a turn signal switch and an amount of rotation of the steering wheel. Thus, thedevice 100 can notify following vehicles in the adjacent lane of the intention of changing lanes of the vehicle where thedevice 100 is mounted. - In addition, the
device 100 does not need to always project visual information toward the heading direction while a dangerous object is not present around the vehicle. Thedevice 100 may need to project the visual information toward the heading direction at a certain period just before the projected visual information starts to move to the detected dangerous object.
Claims (8)
1. An information providing device provided in a vehicle, the device comprising:
an object detecting unit for detecting an object outside the vehicle;
a providing unit for providing, outside of the vehicle, visual information which is visually perceptible; and
a controlling unit for causing the providing unit
to provide the visual information to indicate a heading direction of the vehicle and
to move the visual information to indicate a direction towards the detected object when the detected object is determined to include at least one of (i) a danger to advancement of the vehicle and (ii) an indication of stopping the vehicle.
2. The information providing device of claim 1 , further comprising:
a running state detecting unit for detecting a running state of the vehicle,
wherein the controlling unit
estimates a path which the vehicle is estimated to follow according to the detected running state and
causes the providing unit to provide the visual information that shows the estimated path to indicate the heading direction of the vehicle.
3. The information providing device of claim 2 ,
wherein the controlling unit changes, with respect to the shown estimated path, at least one of a direction, a length, and a shape, based on the detected running state.
4. The information providing device of claim 1 , further comprising:
a running state detecting unit for detecting a running state of the vehicle,
wherein the controlling unit
estimates a path which the vehicle is estimated to follow based on the detected running state and
determines whether the detected object includes at least one of the danger to advancement of the vehicle and the indication of stopping the vehicle, based on the estimated path.
5. The information providing device of claim 1 , further comprising:
a driving operation state detecting unit for detecting a driving operation state of a driver of the vehicle,
wherein the controlling unit
estimates intention of the driver regarding the driving operation based on the detected driving operation state,
determines whether the detected object includes at least one of the danger to advancement of the vehicle and the indication of stopping the vehicle, based on the estimated intention.
6. The information providing device of claim 1 ,
wherein the object detecting unit detects at least one of another vehicle, a pedestrian, a traffic control device, a road traffic sign, and a boarder between lanes of a road, and
wherein, when the providing unit is caused to move the visual information to indicate the direction towards the detected object, the providing unit provides the visual information which varies depending on a kind of a detected object by the object detecting unit.
7. The information providing device of claim 1 ,
wherein the providing unit provides the visual information that is formed using laser beams.
8. An information providing method used in a vehicle, the vehicle including:
an object detecting unit for detecting an object outside the vehicle; and
a providing unit for providing, outside of the vehicle, visual information which is visually perceptible,
the information providing method comprising steps of:
causing the providing unit to provide the visual information to indicate a heading direction of the vehicle;
executing a determination of whether or not the detected object includes at least one of (i) a danger to advancement of the vehicle and (ii) an indication of stopping the vehicle; and
causing the providing unit to move the visual information to indicate a direction towards the detected object, when the determination is affirmed.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/287,709 US7394356B2 (en) | 2004-10-27 | 2005-11-28 | Information presenting apparatus and method |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2004312849A JP2006127055A (en) | 2004-10-27 | 2004-10-27 | Information presentation device for vehicle |
JP2004-312849 | 2004-10-27 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/287,709 Continuation-In-Part US7394356B2 (en) | 2004-10-27 | 2005-11-28 | Information presenting apparatus and method |
Publications (1)
Publication Number | Publication Date |
---|---|
US20060087416A1 true US20060087416A1 (en) | 2006-04-27 |
Family
ID=36202056
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/232,914 Abandoned US20060087416A1 (en) | 2004-10-27 | 2005-09-23 | Information providing device for vehicle |
US11/287,709 Active 2026-05-29 US7394356B2 (en) | 2004-10-27 | 2005-11-28 | Information presenting apparatus and method |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/287,709 Active 2026-05-29 US7394356B2 (en) | 2004-10-27 | 2005-11-28 | Information presenting apparatus and method |
Country Status (4)
Country | Link |
---|---|
US (2) | US20060087416A1 (en) |
JP (1) | JP2006127055A (en) |
CN (1) | CN100408383C (en) |
DE (1) | DE102005051049A1 (en) |
Cited By (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE102006050548A1 (en) * | 2006-10-26 | 2008-04-30 | Bayerische Motoren Werke Ag | Motorist warning method for motor vehicle i.e. passenger car , involves determining projection point outside of motor vehicle for production of projection object based on situation information, and selectively producing object in point |
WO2008155006A1 (en) * | 2007-06-16 | 2008-12-24 | Bayerische Motoren Werke Aktiengesellschaft | Method for supporting a driver of a motor vehicle when driving the motor vehicle through a narrow passage and/or for maintaining a safe distance from a preceding vehicle |
WO2009066149A3 (en) * | 2007-11-20 | 2009-09-17 | Toyota Jidosha Kabushiki Kaisha | Collision prediction device and method of controlling the collision prediction device |
US8248219B2 (en) | 2007-01-19 | 2012-08-21 | Denso Corporation | Interior information display apparatus and light irradiation apparatus used for the same |
US20140028451A1 (en) * | 2011-04-20 | 2014-01-30 | Toyota Jidosha Kabushiki Kaisha | Vehicle periphery alert device |
DE102012219572A1 (en) * | 2012-10-25 | 2014-04-30 | Bayerische Motoren Werke Aktiengesellschaft | Device for representing reference on display of vehicle such as passenger car for other road users, has display with spatial arrangement for viewer outside of display to determine reference for display, based on received information |
DE102015200131A1 (en) | 2015-01-08 | 2016-07-14 | Volkswagen Aktiengesellschaft | Method and driver assistance system for assisting a driver of a vehicle |
CN106335512A (en) * | 2016-11-04 | 2017-01-18 | 大连文森特软件科技有限公司 | Driving Assistance System Based on AR Augmented Reality and Vehicle Violation Record Query |
US9583002B2 (en) | 2011-12-14 | 2017-02-28 | Toyota Jidosha Kabushiki Kaisha | Vehicle information transmitting device |
US10082869B2 (en) * | 2017-02-03 | 2018-09-25 | Qualcomm Incorporated | Maintaining occupant awareness in vehicles |
CN109643490A (en) * | 2017-03-29 | 2019-04-16 | 松下电器(美国)知识产权公司 | Controller of vehicle, control method for vehicle and program |
WO2019139959A1 (en) | 2018-01-12 | 2019-07-18 | Intelligent Security Systems Corporation | Systems and methods for image-based light output |
US10408634B2 (en) * | 2014-03-25 | 2019-09-10 | Jaguar Land Rover Limited | Navigation system |
US20190307632A1 (en) * | 2016-08-05 | 2019-10-10 | Sony Corporation | Information processing device, information processing method, and program |
US10522054B2 (en) | 2013-09-05 | 2019-12-31 | Crown Equipment Corporation | Dynamic operator behavior analyzer |
US11120691B2 (en) | 2019-12-02 | 2021-09-14 | Denso Corporation | Systems and methods for providing warnings to surrounding vehicles to avoid collisions |
US11238737B2 (en) | 2017-03-09 | 2022-02-01 | Bayerische Motoren Werke Aktiengesellschaft | Motor vehicle comprising a lighting module for generating a set of symbols |
US20240005795A1 (en) * | 2021-03-18 | 2024-01-04 | Huawei Technologies Co., Ltd. | Light Projection Apparatus and Method, and Storage Medium |
Families Citing this family (42)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1820695A3 (en) * | 2006-02-15 | 2007-11-21 | MAGNETI MARELLI SISTEMI ELETTRONICI S.p.A. | Control system for automotive display devices |
US8965685B1 (en) | 2006-04-07 | 2015-02-24 | Here Global B.V. | Method and system for enabling precautionary actions in a vehicle |
JP4665841B2 (en) * | 2006-06-14 | 2011-04-06 | トヨタ自動車株式会社 | Warning device for vehicle |
EP1892688B1 (en) * | 2006-08-24 | 2010-09-01 | Valeo Vision | Method for determining the passing of a vehicle in a bottleneck |
DE102006050547A1 (en) * | 2006-10-26 | 2008-04-30 | Bayerische Motoren Werke Ag | Method for displaying information |
DE102006050546B4 (en) | 2006-10-26 | 2020-07-23 | Bayerische Motoren Werke Aktiengesellschaft | Information presentation method |
JP5262057B2 (en) * | 2006-11-17 | 2013-08-14 | 株式会社豊田中央研究所 | Irradiation device |
US8027029B2 (en) | 2007-11-07 | 2011-09-27 | Magna Electronics Inc. | Object detection and tracking system |
JP2009225322A (en) * | 2008-03-18 | 2009-10-01 | Hyundai Motor Co Ltd | Vehicle information display system |
JP5332356B2 (en) * | 2008-07-08 | 2013-11-06 | 日産自動車株式会社 | Vehicle driving support device |
DE102009009473A1 (en) * | 2009-02-19 | 2010-08-26 | Volkswagen Ag | Driver assisting method for vehicle, involves projecting light signal at road surface by vehicle for alerting road users about collision danger when one road user represents collision danger for vehicle |
DE102011081380A1 (en) | 2011-08-23 | 2013-02-28 | Robert Bosch Gmbh | Method for controlling a light emission of a headlamp of a vehicle |
DE202009012589U1 (en) * | 2009-09-16 | 2011-02-03 | Sick Ag | Optoelectronic sensor |
US8502860B2 (en) * | 2009-09-29 | 2013-08-06 | Toyota Motor Engineering & Manufacturing North America (Tema) | Electronic control system, electronic control unit and associated methodology of adapting 3D panoramic views of vehicle surroundings by predicting driver intent |
US10703299B2 (en) * | 2010-04-19 | 2020-07-07 | SMR Patents S.à.r.l. | Rear view mirror simulation |
CN101890947A (en) * | 2010-07-05 | 2010-11-24 | 邹崇毓 | Automobile driving assisting system |
US8717192B2 (en) | 2010-10-08 | 2014-05-06 | Navteq B.V. | Method and system for using intersecting electronic horizons |
DE102010063567A1 (en) * | 2010-12-20 | 2012-06-21 | Bayerische Motoren Werke Aktiengesellschaft | Display system for a motor vehicle |
US8781644B2 (en) * | 2011-03-21 | 2014-07-15 | Denso Corporation | Method and apparatus for recognizing shape of road for vehicles |
US8890674B2 (en) * | 2011-06-07 | 2014-11-18 | Continental Automotive Systems, Inc. | Driver assistance detection system |
DE102011081382A1 (en) * | 2011-08-23 | 2013-02-28 | Robert Bosch Gmbh | Method and device for changing a light emission of at least one headlight of a vehicle |
DE102011081371A1 (en) | 2011-08-23 | 2013-02-28 | Robert Bosch Gmbh | Method and device for controlling a light emission of a headlamp of a vehicle |
KR101288069B1 (en) * | 2011-11-14 | 2013-07-18 | 현대모비스 주식회사 | Parking assistance system using projector and method thereof |
DE102012002333A1 (en) * | 2012-02-07 | 2013-08-08 | Daimler Ag | Method for controlling optical axis of beam of headlight distribution of vehicle, involves directing optical axis of beam as warning function to border lane boundary, when lane boundary is exceeded |
DE102012003158A1 (en) * | 2012-02-17 | 2013-08-22 | Daimler Ag | Method for projecting laser light image in environment around vehicle during driving vehicle, involves projecting manipulated contents of captured image to position in predetermined environment by using laser light beams |
JP5492962B2 (en) * | 2012-09-28 | 2014-05-14 | 富士重工業株式会社 | Gaze guidance system |
DE102012024627A1 (en) * | 2012-12-17 | 2014-06-18 | GM Global Technology Operations LLC (n. d. Gesetzen des Staates Delaware) | Method for controlling headlamp with adjustable light direction in motor vehicle, involves detecting unlighted road user over camera system, lidar system or radar system |
DE102013001282A1 (en) * | 2013-01-25 | 2014-07-31 | GM Global Technology Operations LLC (n. d. Ges. d. Staates Delaware) | Method for controlling bending light in motor car, involves performing temporary change of color emitted by cornering light, if road users relative to motor car is in direction in which cornering light performs lighting |
DE102014000935A1 (en) * | 2014-01-22 | 2015-07-23 | Johannes Reusch | Method and arrangement for generating a traffic light display for vehicles |
DE102014016336A1 (en) * | 2014-11-05 | 2016-05-12 | Audi Ag | System and method for controlling the light distribution of a laser-based pixel light system of at least one vehicle-side headlamp device |
KR102383425B1 (en) * | 2014-12-01 | 2022-04-07 | 현대자동차주식회사 | Electronic apparatus, control method of electronic apparatus, computer program and computer readable recording medium |
DE102014019420A1 (en) * | 2014-12-22 | 2016-06-23 | GM Global Technology Operations LLC (n. d. Ges. d. Staates Delaware) | Motor vehicle headlight system, motor vehicle, method for operating a motor vehicle headlight system and computer program product |
DE102015201764B4 (en) * | 2015-02-02 | 2018-04-05 | Volkswagen Aktiengesellschaft | Method and driver assistance system for generating a light distribution by a vehicle for outputting a driving instruction |
DE102015201767B4 (en) | 2015-02-02 | 2021-12-16 | Volkswagen Aktiengesellschaft | Method for generating a light distribution for the output of driving instructions for another vehicle |
DE102015115354B4 (en) * | 2015-09-11 | 2025-02-20 | HELLA GmbH & Co. KGaA | Method and device for projecting a temporary road marking onto a road |
DE102016001692B4 (en) | 2016-02-13 | 2022-01-27 | Audi Ag | Driver assistance system for controlling light emission from a vehicle headlight device |
US10328896B2 (en) * | 2017-05-18 | 2019-06-25 | Ford Global Technologies, Llc | Vehicle theft avoidance systems and associated methods |
CN109887334A (en) * | 2017-12-01 | 2019-06-14 | 奥迪股份公司 | Vehicle drive assist system and method |
JP2019220009A (en) * | 2018-06-21 | 2019-12-26 | ルネサスエレクトロニクス株式会社 | Semiconductor device, message image output method, and program |
KR102087359B1 (en) * | 2018-10-05 | 2020-03-10 | 성균관대학교산학협력단 | Methods and apparatuses for attracting driver's eyes based on vehicle path |
JP6642681B2 (en) * | 2018-11-06 | 2020-02-12 | 日本精機株式会社 | Vehicle information projection system and vehicle information projection method |
DE102020101020B4 (en) | 2020-01-17 | 2024-03-21 | Audi Aktiengesellschaft | motor vehicle |
Family Cites Families (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH02231245A (en) | 1989-03-03 | 1990-09-13 | Nariyasu Shiba | Advance path forecasting device utilizing light beam for automobile |
JPH06255399A (en) | 1993-03-04 | 1994-09-13 | Mazda Motor Corp | Vehicle display |
JP3522317B2 (en) * | 1993-12-27 | 2004-04-26 | 富士重工業株式会社 | Travel guide device for vehicles |
US5719568A (en) * | 1995-11-28 | 1998-02-17 | Adams; Peter J. | Perimeter light system |
JP3630252B2 (en) * | 1996-02-07 | 2005-03-16 | 株式会社エクォス・リサーチ | Navigation device |
US6272431B1 (en) * | 1997-04-29 | 2001-08-07 | Thomas Zamojdo | Method for displaying a map in a vehicle en-route guidance system |
JP3617307B2 (en) * | 1998-04-20 | 2005-02-02 | 日産自動車株式会社 | Vehicle lighting device |
US6198386B1 (en) * | 1998-10-02 | 2001-03-06 | White, Ii Locke | Vehicle guidance assembly and method |
JP3864406B2 (en) * | 1999-01-26 | 2006-12-27 | マツダ株式会社 | Vehicle display device |
CN2379782Y (en) * | 1999-02-26 | 2000-05-24 | 三水市好帮手电子科技有限公司 | Radar type instrument for guiding driver to back a car |
US6486798B2 (en) * | 2000-05-11 | 2002-11-26 | Rastar Corporation | System and method of preventing aircraft wing damage |
US6963293B1 (en) * | 2000-05-11 | 2005-11-08 | Rastar Corporation | System and method of preventing aircraft wingtip ground incursion |
JP2002156233A (en) * | 2000-11-16 | 2002-05-31 | Fuji Heavy Ind Ltd | In-vehicle information display device |
US6741186B2 (en) * | 2001-05-24 | 2004-05-25 | Phillip N. Ross | Infrared road line detector |
JP3700614B2 (en) | 2001-06-22 | 2005-09-28 | 株式会社豊田自動織機 | Parking assistance device |
JP2003123196A (en) * | 2001-10-10 | 2003-04-25 | Denso Corp | Device and program for monitoring circumference of vehicle |
JP4161584B2 (en) * | 2002-02-07 | 2008-10-08 | トヨタ自動車株式会社 | Safety device for moving objects |
JP2003291688A (en) * | 2002-04-03 | 2003-10-15 | Denso Corp | Display method, driving support device, program |
US6801843B2 (en) * | 2002-05-24 | 2004-10-05 | Ford Global Technologies, Llc | Vehicle pre-crash sensing based conic target threat assessment system |
JP2004185105A (en) * | 2002-11-29 | 2004-07-02 | Nissan Motor Co Ltd | Obstacle informing device |
JP2004220257A (en) * | 2003-01-14 | 2004-08-05 | Toyota Motor Corp | Vehicle safety assist device and vehicle safety assist system |
JP4134785B2 (en) * | 2003-03-28 | 2008-08-20 | 株式会社デンソー | Display device |
JP4239771B2 (en) * | 2003-09-18 | 2009-03-18 | 日産自動車株式会社 | VEHICLE DRIVE OPERATION ASSISTANCE DEVICE AND VEHICLE HAVING VEHICLE DRIVE OPERATION ASSISTANCE DEVICE |
-
2004
- 2004-10-27 JP JP2004312849A patent/JP2006127055A/en active Pending
-
2005
- 2005-09-23 US US11/232,914 patent/US20060087416A1/en not_active Abandoned
- 2005-10-25 DE DE102005051049A patent/DE102005051049A1/en not_active Ceased
- 2005-10-26 CN CNB2005101181541A patent/CN100408383C/en not_active Expired - Fee Related
- 2005-11-28 US US11/287,709 patent/US7394356B2/en active Active
Cited By (36)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE102006050548B4 (en) * | 2006-10-26 | 2014-05-08 | Bayerische Motoren Werke Aktiengesellschaft | Procedure for warning other road users |
DE102006050548A1 (en) * | 2006-10-26 | 2008-04-30 | Bayerische Motoren Werke Ag | Motorist warning method for motor vehicle i.e. passenger car , involves determining projection point outside of motor vehicle for production of projection object based on situation information, and selectively producing object in point |
US8248219B2 (en) | 2007-01-19 | 2012-08-21 | Denso Corporation | Interior information display apparatus and light irradiation apparatus used for the same |
WO2008155006A1 (en) * | 2007-06-16 | 2008-12-24 | Bayerische Motoren Werke Aktiengesellschaft | Method for supporting a driver of a motor vehicle when driving the motor vehicle through a narrow passage and/or for maintaining a safe distance from a preceding vehicle |
US20100094541A1 (en) * | 2007-06-16 | 2010-04-15 | Bayerische Motoren Werke Aktiengesellschaft | Method for Assisting a Motor Vehicle Driver When Driving Through a Narrow Passage and/or for Maintaining a Safe Distance from a Vehicle in Front |
US10214143B2 (en) | 2007-06-16 | 2019-02-26 | Bayerische Motoren Werke Aktiengesellschaft | Method for assisting a motor vehicle driver when driving through a narrow passage and/or for maintaining a safe distance from a vehicle in front |
WO2009066149A3 (en) * | 2007-11-20 | 2009-09-17 | Toyota Jidosha Kabushiki Kaisha | Collision prediction device and method of controlling the collision prediction device |
US20140028451A1 (en) * | 2011-04-20 | 2014-01-30 | Toyota Jidosha Kabushiki Kaisha | Vehicle periphery alert device |
US10150407B2 (en) | 2011-04-20 | 2018-12-11 | Toyota Jidosha Kabushiki Kaisha | Vehicle periphery alert device |
US9583002B2 (en) | 2011-12-14 | 2017-02-28 | Toyota Jidosha Kabushiki Kaisha | Vehicle information transmitting device |
DE102012219572A1 (en) * | 2012-10-25 | 2014-04-30 | Bayerische Motoren Werke Aktiengesellschaft | Device for representing reference on display of vehicle such as passenger car for other road users, has display with spatial arrangement for viewer outside of display to determine reference for display, based on received information |
US11935426B2 (en) | 2013-09-05 | 2024-03-19 | Crown Equipment Corporation | Dynamic operator behavior analyzer |
US11694572B2 (en) | 2013-09-05 | 2023-07-04 | Crown Equipment Corporation | Dynamic operator behavior analyzer |
EP3671174B1 (en) * | 2013-09-05 | 2023-11-01 | Crown Equipment Corporation | Dynamic operator behavior analyzer |
US10991266B2 (en) | 2013-09-05 | 2021-04-27 | Crown Equipment Corporation | Dynamic operator behavior analyzer |
EP3042174B1 (en) * | 2013-09-05 | 2020-01-15 | Crown Equipment Corporation | Dynamic operator behavior analyzer |
US10522054B2 (en) | 2013-09-05 | 2019-12-31 | Crown Equipment Corporation | Dynamic operator behavior analyzer |
US10408634B2 (en) * | 2014-03-25 | 2019-09-10 | Jaguar Land Rover Limited | Navigation system |
WO2016110365A1 (en) | 2015-01-08 | 2016-07-14 | Volkswagen Aktiengesellschaft | Method and driver assistance system for assisting a driver of a vehicle |
US10220771B2 (en) | 2015-01-08 | 2019-03-05 | Volkswagen Aktiengesellschaft | Method and driver assistance system for assisting a driver of a vehicle |
DE102015200131A1 (en) | 2015-01-08 | 2016-07-14 | Volkswagen Aktiengesellschaft | Method and driver assistance system for assisting a driver of a vehicle |
US11744766B2 (en) | 2016-08-05 | 2023-09-05 | Sony Corporation | Information processing apparatus and information processing method |
US10765588B2 (en) * | 2016-08-05 | 2020-09-08 | Sony Corporation | Information processing apparatus and information processing method |
US20190307632A1 (en) * | 2016-08-05 | 2019-10-10 | Sony Corporation | Information processing device, information processing method, and program |
CN106335512A (en) * | 2016-11-04 | 2017-01-18 | 大连文森特软件科技有限公司 | Driving Assistance System Based on AR Augmented Reality and Vehicle Violation Record Query |
KR20190112274A (en) * | 2017-02-03 | 2019-10-04 | 퀄컴 인코포레이티드 | Maintain occupant awareness in the vehicle |
EP3577606B1 (en) * | 2017-02-03 | 2022-09-07 | Qualcomm Incorporated | Maintaining occupant awareness in vehicles |
KR102487922B1 (en) | 2017-02-03 | 2023-01-12 | 퀄컴 인코포레이티드 | Maintaining occupant awareness in the vehicle |
US10082869B2 (en) * | 2017-02-03 | 2018-09-25 | Qualcomm Incorporated | Maintaining occupant awareness in vehicles |
US11238737B2 (en) | 2017-03-09 | 2022-02-01 | Bayerische Motoren Werke Aktiengesellschaft | Motor vehicle comprising a lighting module for generating a set of symbols |
CN109643490A (en) * | 2017-03-29 | 2019-04-16 | 松下电器(美国)知识产权公司 | Controller of vehicle, control method for vehicle and program |
EP3738316A4 (en) * | 2018-01-12 | 2021-12-22 | Intelligent Security Systems Corporation | SYSTEMS AND PROCESSES FOR IMAGE-BASED LIGHT OUTPUT |
WO2019139959A1 (en) | 2018-01-12 | 2019-07-18 | Intelligent Security Systems Corporation | Systems and methods for image-based light output |
US11120691B2 (en) | 2019-12-02 | 2021-09-14 | Denso Corporation | Systems and methods for providing warnings to surrounding vehicles to avoid collisions |
US20240005795A1 (en) * | 2021-03-18 | 2024-01-04 | Huawei Technologies Co., Ltd. | Light Projection Apparatus and Method, and Storage Medium |
US12307899B2 (en) * | 2021-03-18 | 2025-05-20 | Shenzhen Yinwang Intelligent Technologies Co., Ltd. | Light projection apparatus and method, and storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN100408383C (en) | 2008-08-06 |
US20060097858A1 (en) | 2006-05-11 |
CN1810547A (en) | 2006-08-02 |
DE102005051049A1 (en) | 2006-05-04 |
US7394356B2 (en) | 2008-07-01 |
JP2006127055A (en) | 2006-05-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20060087416A1 (en) | Information providing device for vehicle | |
JP6673299B2 (en) | Steering support device | |
US12275425B2 (en) | Alert apparatus for host vehicle | |
JP4044031B2 (en) | Vehicle travel support device | |
US6708087B2 (en) | Display system for vehicle | |
JP4730267B2 (en) | Visibility state determination device for vehicle | |
CN103459225B (en) | Driving assist system | |
US12103520B2 (en) | Vehicle driving support system and vehicle driving support method | |
US20180018939A1 (en) | Display system replacing side mirror and method for controlling output brightness thereof | |
JP2019001183A (en) | Lane change support device | |
WO2019021471A1 (en) | Display control method and display control device | |
US20130058116A1 (en) | Method and device for changing a light emission of at least one headlight of a vehicle | |
CN106364480A (en) | Vehicle control device | |
JP2005078414A (en) | Vehicle travel support device | |
CN110356317B (en) | Vehicle reminding method and system and vehicle | |
JP2008094151A (en) | Vehicular light control system | |
JP2005161977A (en) | Vehicular travel supporting device | |
JP4983564B2 (en) | Vehicle tracking device | |
JP2019003235A (en) | Target information acquisition device | |
JP4492725B2 (en) | Light control device | |
KR20190066115A (en) | Vehicle and method for controlling thereof | |
CN110371110B (en) | Vehicle control method and system and vehicle | |
JP4434864B2 (en) | Vehicle lamp control device | |
KR101395616B1 (en) | Device for displaying distance between cars | |
WO2018199075A1 (en) | Vehicle control device and vehicle control method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: DENSO CORPORATION, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KUMABE, HAJIME;TSURU, NAOHIKO;NISHIMURA, TAKAO;REEL/FRAME:017027/0499 Effective date: 20050826 |
|
STCB | Information on status: application discontinuation |
Free format text: EXPRESSLY ABANDONED -- DURING EXAMINATION |