+

US20060087416A1 - Information providing device for vehicle - Google Patents

Information providing device for vehicle Download PDF

Info

Publication number
US20060087416A1
US20060087416A1 US11/232,914 US23291405A US2006087416A1 US 20060087416 A1 US20060087416 A1 US 20060087416A1 US 23291405 A US23291405 A US 23291405A US 2006087416 A1 US2006087416 A1 US 2006087416A1
Authority
US
United States
Prior art keywords
vehicle
information
unit
visual information
providing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11/232,914
Inventor
Hajime Kumabe
Naohiko Tsuru
Takao Nishimura
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Denso Corp
Original Assignee
Denso Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Denso Corp filed Critical Denso Corp
Assigned to DENSO CORPORATION reassignment DENSO CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: KUMABE, HAJIME, NISHIMURA, TAKAO, TSURU, NAOHIKO
Priority to US11/287,709 priority Critical patent/US7394356B2/en
Publication of US20060087416A1 publication Critical patent/US20060087416A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60QARRANGEMENT OF SIGNALLING OR LIGHTING DEVICES, THE MOUNTING OR SUPPORTING THEREOF OR CIRCUITS THEREFOR, FOR VEHICLES IN GENERAL
    • B60Q1/00Arrangement of optical signalling or lighting devices, the mounting or supporting thereof or circuits therefor
    • B60Q1/26Arrangement of optical signalling or lighting devices, the mounting or supporting thereof or circuits therefor the devices being primarily intended to indicate the vehicle, or parts thereof, or to give signals, to other traffic
    • B60Q1/50Arrangement of optical signalling or lighting devices, the mounting or supporting thereof or circuits therefor the devices being primarily intended to indicate the vehicle, or parts thereof, or to give signals, to other traffic for indicating other intentions or conditions, e.g. request for waiting or overtaking
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60QARRANGEMENT OF SIGNALLING OR LIGHTING DEVICES, THE MOUNTING OR SUPPORTING THEREOF OR CIRCUITS THEREFOR, FOR VEHICLES IN GENERAL
    • B60Q2400/00Special features or arrangements of exterior signal lamps for vehicles
    • B60Q2400/50Projected symbol or information, e.g. onto the road or car body

Definitions

  • the present invention relates to an information providing device for a vehicle.
  • JP-H06-255399-A a conventional device is described which warns a vehicle's driver of an dangerous object by illuminating the dangerous object using a spotlight or by projecting a warning sign to a position which is on the windshield of the vehicle and corresponds to a position of the dangerous object.
  • the driver moves his/her sight line to the warning sign and adjusts his/her focus to the warning sign on the windshield to notice presence of the dangerous object. Subsequently, the driver adjusts his/her focus from the windshield to the dangerous object itself to estimate its distance from the vehicle.
  • the driver needs to adjust his/her focus from the windshield to an actual position of a dangerous object in the process of noticing the dangerous object.
  • the device projects the spotlight only when it detects a dangerous object. Therefore, the driver possibly has difficulty in immediately reacting to the spotlight which comes up at an unexpected position. Moreover, a sight line of the driver and a heading direction of the vehicle are almost the same; therefore, the driver possibly fails to recognize the spotlight if the spotlight deviates from the heading direction.
  • the driver sometimes has difficulty in recognizing a dangerous object even in a case where the driver needs to react against the dangerous object as quickly as possible.
  • an information providing device in a vehicle is provided with the following.
  • An object detecting unit is included for detecting an object outside the vehicle.
  • a providing unit is included for providing, outside the vehicle, visual information which is visually perceptible.
  • a controlling unit is included for controlling the providing unit.
  • the providing unit is usually caused to provide the visual information to indicate a heading direction of the vehicle.
  • the providing unit is caused to move the visual information to indicate a direction towards the detected object.
  • the information providing device can usually lead a driver's sight line to the heading direction of the vehicle.
  • driver's eyes can focus on the heading direction, which is proper for usual driving operation.
  • the visual information is moved while continuing to be perceptible so that the detected object can be indicated by the moved visual information.
  • the driver watching the heading direction notices the movement of the visual information and the sight line of the driver is led to the detected object by following the movement. Consequently, this helps prevent a delay in movement of the sight line or an oversight of the dangerous object.
  • an amount of adjustment of the focus in this case using the device of the present invention is small compared to that in a case using the conventional device where the focus moves from the windshield of the vehicle to the dangerous object outside the vehicle.
  • FIG. 1 shows a structure of an information providing device for a vehicle
  • FIG. 2 shows a diagram of units included in a controller in the information providing device
  • FIG. 3 shows a situation where the information providing device projects visual information of an estimated path for a subject vehicle
  • FIG. 4 shows a situation where the information providing device projects visual information of danger information towards a dangerous object
  • FIG. 5 shows a flowchart of a process which the controller executes.
  • an information providing device 100 mounted in a subject vehicle includes surrounding information detectors 10 , driving operation state detectors 20 , vehicle running state detectors 30 , a controller 40 , and a laser projection device 50 .
  • the surrounding information detectors 10 include a laser or radar sensor, an image sensor, an ultrasonic sensor, and a navigation device. Using the detectors 10 , the controller 40 detects characteristics of an object located outside the vehicle, specifically, frontward, sideward, and rearward of the vehicle. The detected characteristics of the object include, with respect to the object, a kind, a position, a heading direction, or a relative velocity relative to the vehicle.
  • the kind of the object is, for example, another vehicle (further, a passenger automobile, a truck, a two-wheeled motorcycle, or the like), a pedestrian, a traffic control device, a road traffic sign, or a lane marker or a border between lanes on a road (e.g., a white line, a yellow line, or a central reserve).
  • another vehicle further, a passenger automobile, a truck, a two-wheeled motorcycle, or the like
  • a pedestrian e.g., a passenger automobile, a truck, a two-wheeled motorcycle, or the like
  • a traffic control device e.g., a pedestrian, a traffic control device, a road traffic sign, or a lane marker or a border between lanes on a road (e.g., a white line, a yellow line, or a central reserve).
  • the laser or radar sensor irradiates an area within a predetermined distance outside the vehicle using laser beams or the like and calculates a distance from a vehicle, a velocity relative to the vehicle, and a direction as seen from the vehicle, with respect to the object which reflects the laser beams or the like. Then the laser or radar sensor outputs the calculated values as electric signals to the controller 40 .
  • the image sensor includes an optical camera which operates as an imaging unit for photographing an area frontward, sideward, or rearward of the vehicle to take an image. According to the image, the image sensor detects characteristics of an object in the image and outputs the characteristics as electric signals to the controller 40 .
  • the characteristics include a kind, a size, and a state (e.g., a lighting color of a traffic light), with respect to the object.
  • the ultrasonic sensor sends ultrasonic sound waves to locations surrounding the vehicle.
  • the locations are about several meters away from the vehicle.
  • the ultrasonic sensor detects a distance between the vehicle and the object which reflects the ultrasonic sound waves, and outputs the detected distance as electric signals to the controller 40 .
  • the navigation device displays a map around a current position of the vehicle, finds an optimum route to a destination, and shows the route to a driver of the vehicle.
  • the navigation device has a storage device memorizing map data which includes roads data (e.g., data regarding road traffic signs and traffic control devices) and facilities data.
  • the roads data includes, for example, data regarding locations or kinds of road signs, data regarding locations of traffic control devices relative to intersections or crosswalks, and data regarding kinds or shapes of traffic control devices.
  • the navigation device detects a current position of the vehicle and outputs information such as a kind and a distance of a structure on a road within a predetermined distance from the current position.
  • the structure is, for example, a traffic control device and a traffic sign which directs or indicates a vehicle to stop or go.
  • the driving operation state detectors 20 include a throttle sensor, a brake sensor, a steering sensor, a shift state sensor, and a turn signal state sensor, which are not illustrated, and outputs detection signals of each sensor to the controller 40 .
  • the throttle sensor detects a throttle position depending on an amount of operation of a throttle pedal.
  • the brake sensor detects a depression force of a brake pedal.
  • the steering sensor detects a direction of operation of a steering wheel and an amount of rotation of the steering wheel relative to the central position.
  • the turn signal state sensor detects a direction (i.e., right or left) indicated by a turn signal.
  • the vehicle running state detectors 30 include a speed sensor, an acceleration sensor, a yew rate sensor, a tire direction sensor, and a wheel loading sensor, which are not illustrated, and outputs detection signals of each sensor to the controller 40 .
  • the speed sensor is a wheel speed sensor which detects speed of the vehicle.
  • the acceleration sensor detects acceleration of the vehicle in the anteroposterior direction, the lateral direction, and the vertical direction.
  • the tire direction sensor detects a turn angle of a vehicle wheel around a vertical axis against a reference angle that is an angle of the wheel when the vehicle is going in a straight line.
  • the turn angle to be detected depends on operation of a steering wheel. Therefore, the turn angle can be also detected by the steering sensor.
  • the wheel loading sensor detects a load on each wheel suspension of the vehicle along the vertical direction.
  • the laser projection device 50 radiates laser beams to provide or project visual information by irradiating a target position or a target object outside the vehicle.
  • the irradiated target position or object can be seen through the projected transparent visual information.
  • the driver can view a surrounding area and the visual information as a part of the surrounding area.
  • the laser projection device 50 is controlled by the controller 50 as described below.
  • the controller 40 is a microcomputer which has a CPU, a RAM, a ROM, and an input/output interface, which are not illustrated.
  • the controller 40 includes a surrounding information detection unit 41 , a driving operation information generation unit 42 , a running information generation unit 43 , a driver intention estimation unit 44 , a path estimation unit 45 , a danger determination unit 46 , an object projection calculation unit 47 , and a path projection calculation unit 48 .
  • the CPU achieves the above units by executing programs corresponding to the units, by reading the programs from the ROM and using the RAM as a work area.
  • the surrounding information detection unit 41 detects surrounding information relating to an object surrounding the vehicle, by comprehensively considering the signals from the surrounding information detectors 10 .
  • the object is, for example, another vehicle (a passenger automobile, a truck, a two-wheeled motorcycle, or the like), a pedestrian, a traffic control device, a road traffic sign, or a lane marker or a border between lanes on a road (e.g., a white line, a yellow line, or a central reserve).
  • the surrounding information with respect to the object includes a kind, a position, a heading direction, and a relative velocity to the vehicle.
  • the detected surrounding information is used by the danger determination unit 46 and the object projection calculation unit 47 .
  • the driving operation information generation unit 42 generates, based on the detection signals from the driving operation state detectors 20 , driving operation information which indicates a state of driving operation by the driver.
  • the driving operation information is used by the driver intention estimation unit 44 .
  • the driver intention estimation unit 44 estimates, based on the generated driving operation information, driver's intention regarding driving operation such as starting, accelerating, changing lanes, turning left/right, decelerating, or stopping.
  • Driver intention information which indicates the estimated driver's intention is used by the danger determination unit 46 and the path target calculation unit 48 .
  • the running information generation unit 43 generates, based on the detection signals from the vehicle running state detectors 30 , running information which indicates a state of movement of the vehicle.
  • the driving operation information is used by the path estimation unit 45 .
  • the path estimation unit 45 estimates, based on the generated running information, an estimated path which the vehicle is estimated or expected to follow. In detail, the path estimation unit 45 applies the running information to parameters in a predetermined movement model of the subject vehicle to obtain the estimated path. Information on the estimated path is used by the danger determination unit 46 and the path projection calculation unit 48 .
  • the danger determination unit 46 determines whether the detected object is dangerous to advancement of the vehicle and whether the detected structure on a road indicates the vehicle to stop. Such an indication is made, for example, by a stop signal from a traffic control device and a brief stop sign on a road. In addition, the unit 46 determines a danger level (e.g., high or low) of the detected object depending on a distance between the detected object and the vehicle.
  • a danger level e.g., high or low
  • the determination is executed based on the surrounding information, the driver intention information, and the information on the estimated path. However, the determination can be executed without the driver intention information. The determination can be also executed without the information on the estimated path.
  • the unit 46 determines that the object is dangerous to advancement of the vehicle or that the structure indicates the vehicle to stop, it generates danger information which includes a kind, a relative position to the vehicle, and a danger level, with respect to the object or the structure (hereafter dangerous object).
  • the danger information is used by the object projection calculation unit 47 .
  • the path projection calculation unit 48 calculates path-target positions (e.g., directions and/or distances) to irradiate using laser beams.
  • the path-target positions correspond to an estimated path from a current position of the vehicle to a position which the vehicle is estimated to reach after a period (e.g., 3 seconds) by following the estimated path.
  • the unit 48 outputs the calculated path-target positions to the laser projection device 50 .
  • the laser projection device 50 While the vehicle is running, the laser projection device 50 usually provides or projects an estimated path LL, LR (in FIG. 3 ) of the visual information toward the heading direction of the vehicle as seen from the driver. Therefore, the information providing device 100 can lead the driver's sight line to the path-target positions of the estimated path.
  • the path projection calculation unit 48 may change the way of the projection, based on the running information (e.g., a speed of the vehicle and an amount of rotation of the steering wheel), by changing at least one of a direction, a length, and a shape, with respect to the projection.
  • the running information e.g., a speed of the vehicle and an amount of rotation of the steering wheel
  • the driver generally has difficulty in properly maintaining his/her sight line or focus when his/her arousal level is reduced.
  • characterless and unchanging visual information possibly decreases his/her arousal level further and causes his/her lack of attention.
  • the diver possibly overlooks information necessary for driving the vehicle, or fails to react to the information in time.
  • the change of the way of the projection based on the running state of the vehicle gives variations to the visual information, and possibly suppresses the reduction of the arousal level.
  • the object projection calculation unit 47 calculates object-target positions (e.g., directions and/or distances) that enable visual information as the danger information to indicate the dangerous object itself and a direction towards a position of the dangerous object, when the unit 46 generates the danger information.
  • the object projection calculation unit 47 moves the estimated path of the visual information from the lines LL, LR in FIG. 3 to lines LT, LT in FIG. 4 that correspond to the calculated object-target positions.
  • the object projection calculation unit 47 also determines the way of projection so that the laser projection device 50 can provide the visual information corresponding to the kind of the dangerous object. For example, when the danger information indicates that the dangerous object is a nearby vehicle or a pedestrian, the unit 47 determines the way of projection so that the projected visual information as the danger information includes a shape of an exclamation mark in addition to the lines LT, LT, as seen in FIG. 4 .
  • the controller 40 operates as shown in FIG. 5 as a whole. First, the controller 40 detects the surrounding information using the surrounding information detection unit 41 at Step S 10 .
  • Step S 20 vehicle states of the driving operation state and the vehicle running state are detected using the driving operation information generation unit 42 and the running information generation unit 43 , respectively.
  • Step S 30 vehicle information of the driving operation information and the running information is then generated based on the detections of Step S 20 also using the driving operation information generation unit 42 and the running information generation unit 43 , respectively.
  • Step S 40 the driver's intention is estimated based on the driving operation information using the driver intention estimation unit 44 .
  • Step S 50 the estimated path is estimated based on the running information using the path estimation unit 45 .
  • Step S 60 the path-target positions are calculated for providing the visual information regarding the estimated path using the path projection calculation unit 48 .
  • Step S 70 the detected object or structure is determined to be the dangerous object or not; namely, it is determined whether the detected object is dangerous to advancement of the vehicle and whether the detected structure on a road indicates the vehicle to stop, using the danger determination unit 46 .
  • Step S 80 it is determined whether there is a dangerous object outside the vehicle also using the danger determination unit 46 .
  • Step S 110 takes place.
  • the controller 40 causes the laser projection device 50 to project the estimated path of the lines LL, LR towards the heading direction of the vehicle which is calculated at Step S 60 , as shown in FIG. 3 .
  • Step S 100 the object-target positions and the way of the projection are calculated for providing the visual information of the danger information regarding the dangerous object using the object projection calculation unit 47 .
  • Step S 110 takes place.
  • the controller 40 causes the laser projection device 50 to move the visual information of the estimated path from the lines LL, LR shown in FIG. 3 to the lines LT, LT including the exclamation mark shown in FIG. 4 , based on the object-target positions and way of the projection calculated at Step 100 .
  • Steps S 10 to S 110 the information providing device 100 continuously projects the estimated path as the visual information toward the heading direction of the vehicle while a dangerous object is not around the vehicle. Once the dangerous object appears, the information providing device 100 starts to move the projected estimated path from the positions of the lines LL, LR in FIG. 3 to the positions of the lines LT, LT in FIG. 4 that is approaching the dangerous object. During this movement, the lines continue being projected on a relevant road.
  • the driver can smoothly follow the movement of the lines of the estimated path as the visual information. Therefore, the delay of movement of the driver's sight line can be suppressed, and the oversight of the dangerous object can be reduced.
  • the information providing device 100 may project visual information on a lane maker between two lanes of a road when the vehicle is about to deviate from one of the lanes to the other.
  • the device 100 can draw the driver's attention to the situation.
  • the information providing device 100 may have a monitoring unit for monitoring a sight line of the driver.
  • the device 100 when the device 100 detects, using the monitoring unit, that the driver is looking aside, it can provide the visual information such as an exclamation mark by irradiating the location which the driver is looking at.
  • the device 100 can urge the driver to gaze at a proper direction for driving.
  • the device 100 may retrieve a route, which the vehicle should go along, from the navigation device and provide visual information indicating a direction along the route by irradiating positions in the direction.
  • the device 100 can provide the information of the route to the driver.
  • the device 100 may project visual information on an approaching vehicle to notify the approaching vehicle of presence of the vehicle where the device 100 is mounted.
  • the device 100 may project visual information on a pedestrian who is determined to be not a dangerous object but to be present in front of the vehicle. Thus, the device 100 can notify the pedestrian of presence of the approaching vehicle.
  • the device 100 may project visual information indicating an intention of changing to an adjacent lane, in conjunction with operation of a turn signal switch and an amount of rotation of the steering wheel.
  • the device 100 can notify following vehicles in the adjacent lane of the intention of changing lanes of the vehicle where the device 100 is mounted.
  • the device 100 does not need to always project visual information toward the heading direction while a dangerous object is not present around the vehicle.
  • the device 100 may need to project the visual information toward the heading direction at a certain period just before the projected visual information starts to move to the detected dangerous object.

Landscapes

  • Engineering & Computer Science (AREA)
  • Mechanical Engineering (AREA)
  • Traffic Control Systems (AREA)

Abstract

An information providing device provided in a vehicle includes an object detecting unit for detecting an object outside the vehicle; a providing unit for providing, outside of the vehicle, visual information which is visually perceptible; and a controlling unit for controlling the providing unit. The providing unit is usually caused to provide the visual information to indicate a heading direction of the vehicle. In contrast, when the detected object is determined to include a danger to advancement of the vehicle or an indication of stopping the vehicle, the providing unit is caused to then move the visual information to indicate a direction towards the detected object.

Description

    CROSS REFERENCE TO RELATED APPLICATION
  • This application is based on and incorporates herein by reference Japanese patent application No. 2004-312849 filed on Oct. 27, 2004.
  • FIELD OF THE INVENTION
  • The present invention relates to an information providing device for a vehicle.
  • BACKGROUND OF THE INVEHTION
  • In JP-H06-255399-A, a conventional device is described which warns a vehicle's driver of an dangerous object by illuminating the dangerous object using a spotlight or by projecting a warning sign to a position which is on the windshield of the vehicle and corresponds to a position of the dangerous object.
  • When the warning sign is shown on the windshield, the driver moves his/her sight line to the warning sign and adjusts his/her focus to the warning sign on the windshield to notice presence of the dangerous object. Subsequently, the driver adjusts his/her focus from the windshield to the dangerous object itself to estimate its distance from the vehicle. Thus, in a case where a warning sign is shown on the windshield, the driver needs to adjust his/her focus from the windshield to an actual position of a dangerous object in the process of noticing the dangerous object.
  • In addition, the device projects the spotlight only when it detects a dangerous object. Therefore, the driver possibly has difficulty in immediately reacting to the spotlight which comes up at an unexpected position. Moreover, a sight line of the driver and a heading direction of the vehicle are almost the same; therefore, the driver possibly fails to recognize the spotlight if the spotlight deviates from the heading direction.
  • Thus, using the conventional device, the driver sometimes has difficulty in recognizing a dangerous object even in a case where the driver needs to react against the dangerous object as quickly as possible.
  • SUMMARY OF THE INVENTION
  • It is therefore an object of the present invention to provide an information providing device which enables a driver to quickly react to a dangerous object.
  • To achieve the above object, an information providing device in a vehicle is provided with the following. An object detecting unit is included for detecting an object outside the vehicle. A providing unit is included for providing, outside the vehicle, visual information which is visually perceptible. Furthermore, a controlling unit is included for controlling the providing unit. Here, the providing unit is usually caused to provide the visual information to indicate a heading direction of the vehicle. In contrast, when the detected object is determined to include a danger to advancement of the vehicle or an indication of stopping the vehicle, the providing unit is caused to move the visual information to indicate a direction towards the detected object.
  • Under the above structure, the information providing device can usually lead a driver's sight line to the heading direction of the vehicle. Thus, driver's eyes can focus on the heading direction, which is proper for usual driving operation. In addition, when the detected object is determined to be requiring driver's attention, the visual information is moved while continuing to be perceptible so that the detected object can be indicated by the moved visual information.
  • Thus, the driver watching the heading direction notices the movement of the visual information and the sight line of the driver is led to the detected object by following the movement. Consequently, this helps prevent a delay in movement of the sight line or an oversight of the dangerous object.
  • Furthermore, when the visual information moves from the heading direction to the direction towards the detected dangerous object, the focus of the driver changes also from the heading direction to the direction towards the detected object, within scenery outside of the vehicle. Therefore, an amount of adjustment of the focus in this case using the device of the present invention is small compared to that in a case using the conventional device where the focus moves from the windshield of the vehicle to the dangerous object outside the vehicle.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The above and other objects, features, and advantages of the present invention will become more apparent from the following detailed description made with reference to the accompanying drawings. In the drawings:
  • FIG. 1 shows a structure of an information providing device for a vehicle;
  • FIG. 2 shows a diagram of units included in a controller in the information providing device;
  • FIG. 3 shows a situation where the information providing device projects visual information of an estimated path for a subject vehicle;
  • FIG. 4 shows a situation where the information providing device projects visual information of danger information towards a dangerous object; and
  • FIG. 5 shows a flowchart of a process which the controller executes.
  • DETAILED DESCRIPTION OF THE INVENTION
  • As shown in FIG. 1, an information providing device 100 mounted in a subject vehicle includes surrounding information detectors 10, driving operation state detectors 20, vehicle running state detectors 30, a controller 40, and a laser projection device 50.
  • The surrounding information detectors 10 include a laser or radar sensor, an image sensor, an ultrasonic sensor, and a navigation device. Using the detectors 10, the controller 40 detects characteristics of an object located outside the vehicle, specifically, frontward, sideward, and rearward of the vehicle. The detected characteristics of the object include, with respect to the object, a kind, a position, a heading direction, or a relative velocity relative to the vehicle. The kind of the object is, for example, another vehicle (further, a passenger automobile, a truck, a two-wheeled motorcycle, or the like), a pedestrian, a traffic control device, a road traffic sign, or a lane marker or a border between lanes on a road (e.g., a white line, a yellow line, or a central reserve).
  • The laser or radar sensor irradiates an area within a predetermined distance outside the vehicle using laser beams or the like and calculates a distance from a vehicle, a velocity relative to the vehicle, and a direction as seen from the vehicle, with respect to the object which reflects the laser beams or the like. Then the laser or radar sensor outputs the calculated values as electric signals to the controller 40.
  • The image sensor includes an optical camera which operates as an imaging unit for photographing an area frontward, sideward, or rearward of the vehicle to take an image. According to the image, the image sensor detects characteristics of an object in the image and outputs the characteristics as electric signals to the controller 40. The characteristics include a kind, a size, and a state (e.g., a lighting color of a traffic light), with respect to the object.
  • The ultrasonic sensor sends ultrasonic sound waves to locations surrounding the vehicle. The locations are about several meters away from the vehicle. Then the ultrasonic sensor detects a distance between the vehicle and the object which reflects the ultrasonic sound waves, and outputs the detected distance as electric signals to the controller 40.
  • The navigation device displays a map around a current position of the vehicle, finds an optimum route to a destination, and shows the route to a driver of the vehicle.
  • In addition, the navigation device has a storage device memorizing map data which includes roads data (e.g., data regarding road traffic signs and traffic control devices) and facilities data. The roads data includes, for example, data regarding locations or kinds of road signs, data regarding locations of traffic control devices relative to intersections or crosswalks, and data regarding kinds or shapes of traffic control devices.
  • Moreover, the navigation device detects a current position of the vehicle and outputs information such as a kind and a distance of a structure on a road within a predetermined distance from the current position. The structure is, for example, a traffic control device and a traffic sign which directs or indicates a vehicle to stop or go.
  • The driving operation state detectors 20 include a throttle sensor, a brake sensor, a steering sensor, a shift state sensor, and a turn signal state sensor, which are not illustrated, and outputs detection signals of each sensor to the controller 40.
  • The throttle sensor detects a throttle position depending on an amount of operation of a throttle pedal. The brake sensor detects a depression force of a brake pedal. The steering sensor detects a direction of operation of a steering wheel and an amount of rotation of the steering wheel relative to the central position. The turn signal state sensor detects a direction (i.e., right or left) indicated by a turn signal.
  • The vehicle running state detectors 30 include a speed sensor, an acceleration sensor, a yew rate sensor, a tire direction sensor, and a wheel loading sensor, which are not illustrated, and outputs detection signals of each sensor to the controller 40.
  • The speed sensor is a wheel speed sensor which detects speed of the vehicle. The acceleration sensor detects acceleration of the vehicle in the anteroposterior direction, the lateral direction, and the vertical direction. The tire direction sensor detects a turn angle of a vehicle wheel around a vertical axis against a reference angle that is an angle of the wheel when the vehicle is going in a straight line. The turn angle to be detected depends on operation of a steering wheel. Therefore, the turn angle can be also detected by the steering sensor. The wheel loading sensor detects a load on each wheel suspension of the vehicle along the vertical direction.
  • The laser projection device 50 radiates laser beams to provide or project visual information by irradiating a target position or a target object outside the vehicle. Here, the irradiated target position or object can be seen through the projected transparent visual information. Thus, the driver can view a surrounding area and the visual information as a part of the surrounding area. The laser projection device 50 is controlled by the controller 50 as described below.
  • The controller 40 is a microcomputer which has a CPU, a RAM, a ROM, and an input/output interface, which are not illustrated.
  • As shown in FIG. 2, the controller 40 includes a surrounding information detection unit 41, a driving operation information generation unit 42, a running information generation unit 43, a driver intention estimation unit 44, a path estimation unit 45, a danger determination unit 46, an object projection calculation unit 47, and a path projection calculation unit 48. Here, the CPU achieves the above units by executing programs corresponding to the units, by reading the programs from the ROM and using the RAM as a work area.
  • The surrounding information detection unit 41 detects surrounding information relating to an object surrounding the vehicle, by comprehensively considering the signals from the surrounding information detectors 10. Here, the object is, for example, another vehicle (a passenger automobile, a truck, a two-wheeled motorcycle, or the like), a pedestrian, a traffic control device, a road traffic sign, or a lane marker or a border between lanes on a road (e.g., a white line, a yellow line, or a central reserve). The surrounding information with respect to the object includes a kind, a position, a heading direction, and a relative velocity to the vehicle. The detected surrounding information is used by the danger determination unit 46 and the object projection calculation unit 47.
  • The driving operation information generation unit 42 generates, based on the detection signals from the driving operation state detectors 20, driving operation information which indicates a state of driving operation by the driver. The driving operation information is used by the driver intention estimation unit 44.
  • The driver intention estimation unit 44 estimates, based on the generated driving operation information, driver's intention regarding driving operation such as starting, accelerating, changing lanes, turning left/right, decelerating, or stopping. Driver intention information which indicates the estimated driver's intention is used by the danger determination unit 46 and the path target calculation unit 48.
  • The running information generation unit 43 generates, based on the detection signals from the vehicle running state detectors 30, running information which indicates a state of movement of the vehicle. The driving operation information is used by the path estimation unit 45.
  • The path estimation unit 45 estimates, based on the generated running information, an estimated path which the vehicle is estimated or expected to follow. In detail, the path estimation unit 45 applies the running information to parameters in a predetermined movement model of the subject vehicle to obtain the estimated path. Information on the estimated path is used by the danger determination unit 46 and the path projection calculation unit 48.
  • The danger determination unit 46 determines whether the detected object is dangerous to advancement of the vehicle and whether the detected structure on a road indicates the vehicle to stop. Such an indication is made, for example, by a stop signal from a traffic control device and a brief stop sign on a road. In addition, the unit 46 determines a danger level (e.g., high or low) of the detected object depending on a distance between the detected object and the vehicle.
  • The determination is executed based on the surrounding information, the driver intention information, and the information on the estimated path. However, the determination can be executed without the driver intention information. The determination can be also executed without the information on the estimated path.
  • When the unit 46 determines that the object is dangerous to advancement of the vehicle or that the structure indicates the vehicle to stop, it generates danger information which includes a kind, a relative position to the vehicle, and a danger level, with respect to the object or the structure (hereafter dangerous object). The danger information is used by the object projection calculation unit 47.
  • The path projection calculation unit 48 calculates path-target positions (e.g., directions and/or distances) to irradiate using laser beams. The path-target positions correspond to an estimated path from a current position of the vehicle to a position which the vehicle is estimated to reach after a period (e.g., 3 seconds) by following the estimated path. The unit 48 outputs the calculated path-target positions to the laser projection device 50.
  • While the vehicle is running, the laser projection device 50 usually provides or projects an estimated path LL, LR (in FIG. 3) of the visual information toward the heading direction of the vehicle as seen from the driver. Therefore, the information providing device 100 can lead the driver's sight line to the path-target positions of the estimated path.
  • In addition, the path projection calculation unit 48 may change the way of the projection, based on the running information (e.g., a speed of the vehicle and an amount of rotation of the steering wheel), by changing at least one of a direction, a length, and a shape, with respect to the projection.
  • The driver generally has difficulty in properly maintaining his/her sight line or focus when his/her arousal level is reduced. In this case, characterless and unchanging visual information possibly decreases his/her arousal level further and causes his/her lack of attention. As a result, the diver possibly overlooks information necessary for driving the vehicle, or fails to react to the information in time.
  • Therefore, the change of the way of the projection based on the running state of the vehicle gives variations to the visual information, and possibly suppresses the reduction of the arousal level.
  • The object projection calculation unit 47 calculates object-target positions (e.g., directions and/or distances) that enable visual information as the danger information to indicate the dangerous object itself and a direction towards a position of the dangerous object, when the unit 46 generates the danger information. The object projection calculation unit 47 moves the estimated path of the visual information from the lines LL, LR in FIG. 3 to lines LT, LT in FIG. 4 that correspond to the calculated object-target positions.
  • The object projection calculation unit 47 also determines the way of projection so that the laser projection device 50 can provide the visual information corresponding to the kind of the dangerous object. For example, when the danger information indicates that the dangerous object is a nearby vehicle or a pedestrian, the unit 47 determines the way of projection so that the projected visual information as the danger information includes a shape of an exclamation mark in addition to the lines LT, LT, as seen in FIG. 4.
  • The controller 40 operates as shown in FIG. 5 as a whole. First, the controller 40 detects the surrounding information using the surrounding information detection unit 41 at Step S10.
  • At Step S20, vehicle states of the driving operation state and the vehicle running state are detected using the driving operation information generation unit 42 and the running information generation unit 43, respectively. At Step S30, vehicle information of the driving operation information and the running information is then generated based on the detections of Step S20 also using the driving operation information generation unit 42 and the running information generation unit 43, respectively.
  • Subsequently, at Step S40, the driver's intention is estimated based on the driving operation information using the driver intention estimation unit 44. Step S50, the estimated path is estimated based on the running information using the path estimation unit 45. At Step S60, the path-target positions are calculated for providing the visual information regarding the estimated path using the path projection calculation unit 48.
  • Then, at Step S70, the detected object or structure is determined to be the dangerous object or not; namely, it is determined whether the detected object is dangerous to advancement of the vehicle and whether the detected structure on a road indicates the vehicle to stop, using the danger determination unit 46. Then, at Step S80, it is determined whether there is a dangerous object outside the vehicle also using the danger determination unit 46.
  • When the determination of Step S80 is negated, Step S110 takes place. Here, the controller 40 causes the laser projection device 50 to project the estimated path of the lines LL, LR towards the heading direction of the vehicle which is calculated at Step S60, as shown in FIG. 3.
  • In contrast, when the determination at Step S80 is affirmed, the danger information is generated at Step S90. At Step S100, the object-target positions and the way of the projection are calculated for providing the visual information of the danger information regarding the dangerous object using the object projection calculation unit 47. After Step S100, Step S110 takes place. Here, the controller 40 causes the laser projection device 50 to move the visual information of the estimated path from the lines LL, LR shown in FIG. 3 to the lines LT, LT including the exclamation mark shown in FIG. 4, based on the object-target positions and way of the projection calculated at Step 100.
  • By executing Steps S10 to S110, the information providing device 100 continuously projects the estimated path as the visual information toward the heading direction of the vehicle while a dangerous object is not around the vehicle. Once the dangerous object appears, the information providing device 100 starts to move the projected estimated path from the positions of the lines LL, LR in FIG. 3 to the positions of the lines LT, LT in FIG. 4 that is approaching the dangerous object. During this movement, the lines continue being projected on a relevant road.
  • Thus, the driver can smoothly follow the movement of the lines of the estimated path as the visual information. Therefore, the delay of movement of the driver's sight line can be suppressed, and the oversight of the dangerous object can be reduced.
  • The present invention should not be limited to the embodiment discussed above and shown in the figures, but may be implemented in various ways without departing from the spirit of the invention.
  • For example, the information providing device 100 may project visual information on a lane maker between two lanes of a road when the vehicle is about to deviate from one of the lanes to the other. Thus, the device 100 can draw the driver's attention to the situation.
  • In addition, the information providing device 100 may have a monitoring unit for monitoring a sight line of the driver. In this case, when the device 100 detects, using the monitoring unit, that the driver is looking aside, it can provide the visual information such as an exclamation mark by irradiating the location which the driver is looking at. Thus, the device 100 can urge the driver to gaze at a proper direction for driving.
  • In addition, the device 100 may retrieve a route, which the vehicle should go along, from the navigation device and provide visual information indicating a direction along the route by irradiating positions in the direction. Thus, the device 100 can provide the information of the route to the driver.
  • In addition, when the vehicle is about to enter a blind intersection, the device 100 may project visual information on an approaching vehicle to notify the approaching vehicle of presence of the vehicle where the device 100 is mounted.
  • In addition, the device 100 may project visual information on a pedestrian who is determined to be not a dangerous object but to be present in front of the vehicle. Thus, the device 100 can notify the pedestrian of presence of the approaching vehicle.
  • In addition, the device 100 may project visual information indicating an intention of changing to an adjacent lane, in conjunction with operation of a turn signal switch and an amount of rotation of the steering wheel. Thus, the device 100 can notify following vehicles in the adjacent lane of the intention of changing lanes of the vehicle where the device 100 is mounted.
  • In addition, the device 100 does not need to always project visual information toward the heading direction while a dangerous object is not present around the vehicle. The device 100 may need to project the visual information toward the heading direction at a certain period just before the projected visual information starts to move to the detected dangerous object.

Claims (8)

1. An information providing device provided in a vehicle, the device comprising:
an object detecting unit for detecting an object outside the vehicle;
a providing unit for providing, outside of the vehicle, visual information which is visually perceptible; and
a controlling unit for causing the providing unit
to provide the visual information to indicate a heading direction of the vehicle and
to move the visual information to indicate a direction towards the detected object when the detected object is determined to include at least one of (i) a danger to advancement of the vehicle and (ii) an indication of stopping the vehicle.
2. The information providing device of claim 1, further comprising:
a running state detecting unit for detecting a running state of the vehicle,
wherein the controlling unit
estimates a path which the vehicle is estimated to follow according to the detected running state and
causes the providing unit to provide the visual information that shows the estimated path to indicate the heading direction of the vehicle.
3. The information providing device of claim 2,
wherein the controlling unit changes, with respect to the shown estimated path, at least one of a direction, a length, and a shape, based on the detected running state.
4. The information providing device of claim 1, further comprising:
a running state detecting unit for detecting a running state of the vehicle,
wherein the controlling unit
estimates a path which the vehicle is estimated to follow based on the detected running state and
determines whether the detected object includes at least one of the danger to advancement of the vehicle and the indication of stopping the vehicle, based on the estimated path.
5. The information providing device of claim 1, further comprising:
a driving operation state detecting unit for detecting a driving operation state of a driver of the vehicle,
wherein the controlling unit
estimates intention of the driver regarding the driving operation based on the detected driving operation state,
determines whether the detected object includes at least one of the danger to advancement of the vehicle and the indication of stopping the vehicle, based on the estimated intention.
6. The information providing device of claim 1,
wherein the object detecting unit detects at least one of another vehicle, a pedestrian, a traffic control device, a road traffic sign, and a boarder between lanes of a road, and
wherein, when the providing unit is caused to move the visual information to indicate the direction towards the detected object, the providing unit provides the visual information which varies depending on a kind of a detected object by the object detecting unit.
7. The information providing device of claim 1,
wherein the providing unit provides the visual information that is formed using laser beams.
8. An information providing method used in a vehicle, the vehicle including:
an object detecting unit for detecting an object outside the vehicle; and
a providing unit for providing, outside of the vehicle, visual information which is visually perceptible,
the information providing method comprising steps of:
causing the providing unit to provide the visual information to indicate a heading direction of the vehicle;
executing a determination of whether or not the detected object includes at least one of (i) a danger to advancement of the vehicle and (ii) an indication of stopping the vehicle; and
causing the providing unit to move the visual information to indicate a direction towards the detected object, when the determination is affirmed.
US11/232,914 2004-10-27 2005-09-23 Information providing device for vehicle Abandoned US20060087416A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US11/287,709 US7394356B2 (en) 2004-10-27 2005-11-28 Information presenting apparatus and method

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2004312849A JP2006127055A (en) 2004-10-27 2004-10-27 Information presentation device for vehicle
JP2004-312849 2004-10-27

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US11/287,709 Continuation-In-Part US7394356B2 (en) 2004-10-27 2005-11-28 Information presenting apparatus and method

Publications (1)

Publication Number Publication Date
US20060087416A1 true US20060087416A1 (en) 2006-04-27

Family

ID=36202056

Family Applications (2)

Application Number Title Priority Date Filing Date
US11/232,914 Abandoned US20060087416A1 (en) 2004-10-27 2005-09-23 Information providing device for vehicle
US11/287,709 Active 2026-05-29 US7394356B2 (en) 2004-10-27 2005-11-28 Information presenting apparatus and method

Family Applications After (1)

Application Number Title Priority Date Filing Date
US11/287,709 Active 2026-05-29 US7394356B2 (en) 2004-10-27 2005-11-28 Information presenting apparatus and method

Country Status (4)

Country Link
US (2) US20060087416A1 (en)
JP (1) JP2006127055A (en)
CN (1) CN100408383C (en)
DE (1) DE102005051049A1 (en)

Cited By (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE102006050548A1 (en) * 2006-10-26 2008-04-30 Bayerische Motoren Werke Ag Motorist warning method for motor vehicle i.e. passenger car , involves determining projection point outside of motor vehicle for production of projection object based on situation information, and selectively producing object in point
WO2008155006A1 (en) * 2007-06-16 2008-12-24 Bayerische Motoren Werke Aktiengesellschaft Method for supporting a driver of a motor vehicle when driving the motor vehicle through a narrow passage and/or for maintaining a safe distance from a preceding vehicle
WO2009066149A3 (en) * 2007-11-20 2009-09-17 Toyota Jidosha Kabushiki Kaisha Collision prediction device and method of controlling the collision prediction device
US8248219B2 (en) 2007-01-19 2012-08-21 Denso Corporation Interior information display apparatus and light irradiation apparatus used for the same
US20140028451A1 (en) * 2011-04-20 2014-01-30 Toyota Jidosha Kabushiki Kaisha Vehicle periphery alert device
DE102012219572A1 (en) * 2012-10-25 2014-04-30 Bayerische Motoren Werke Aktiengesellschaft Device for representing reference on display of vehicle such as passenger car for other road users, has display with spatial arrangement for viewer outside of display to determine reference for display, based on received information
DE102015200131A1 (en) 2015-01-08 2016-07-14 Volkswagen Aktiengesellschaft Method and driver assistance system for assisting a driver of a vehicle
CN106335512A (en) * 2016-11-04 2017-01-18 大连文森特软件科技有限公司 Driving Assistance System Based on AR Augmented Reality and Vehicle Violation Record Query
US9583002B2 (en) 2011-12-14 2017-02-28 Toyota Jidosha Kabushiki Kaisha Vehicle information transmitting device
US10082869B2 (en) * 2017-02-03 2018-09-25 Qualcomm Incorporated Maintaining occupant awareness in vehicles
CN109643490A (en) * 2017-03-29 2019-04-16 松下电器(美国)知识产权公司 Controller of vehicle, control method for vehicle and program
WO2019139959A1 (en) 2018-01-12 2019-07-18 Intelligent Security Systems Corporation Systems and methods for image-based light output
US10408634B2 (en) * 2014-03-25 2019-09-10 Jaguar Land Rover Limited Navigation system
US20190307632A1 (en) * 2016-08-05 2019-10-10 Sony Corporation Information processing device, information processing method, and program
US10522054B2 (en) 2013-09-05 2019-12-31 Crown Equipment Corporation Dynamic operator behavior analyzer
US11120691B2 (en) 2019-12-02 2021-09-14 Denso Corporation Systems and methods for providing warnings to surrounding vehicles to avoid collisions
US11238737B2 (en) 2017-03-09 2022-02-01 Bayerische Motoren Werke Aktiengesellschaft Motor vehicle comprising a lighting module for generating a set of symbols
US20240005795A1 (en) * 2021-03-18 2024-01-04 Huawei Technologies Co., Ltd. Light Projection Apparatus and Method, and Storage Medium

Families Citing this family (42)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1820695A3 (en) * 2006-02-15 2007-11-21 MAGNETI MARELLI SISTEMI ELETTRONICI S.p.A. Control system for automotive display devices
US8965685B1 (en) 2006-04-07 2015-02-24 Here Global B.V. Method and system for enabling precautionary actions in a vehicle
JP4665841B2 (en) * 2006-06-14 2011-04-06 トヨタ自動車株式会社 Warning device for vehicle
EP1892688B1 (en) * 2006-08-24 2010-09-01 Valeo Vision Method for determining the passing of a vehicle in a bottleneck
DE102006050547A1 (en) * 2006-10-26 2008-04-30 Bayerische Motoren Werke Ag Method for displaying information
DE102006050546B4 (en) 2006-10-26 2020-07-23 Bayerische Motoren Werke Aktiengesellschaft Information presentation method
JP5262057B2 (en) * 2006-11-17 2013-08-14 株式会社豊田中央研究所 Irradiation device
US8027029B2 (en) 2007-11-07 2011-09-27 Magna Electronics Inc. Object detection and tracking system
JP2009225322A (en) * 2008-03-18 2009-10-01 Hyundai Motor Co Ltd Vehicle information display system
JP5332356B2 (en) * 2008-07-08 2013-11-06 日産自動車株式会社 Vehicle driving support device
DE102009009473A1 (en) * 2009-02-19 2010-08-26 Volkswagen Ag Driver assisting method for vehicle, involves projecting light signal at road surface by vehicle for alerting road users about collision danger when one road user represents collision danger for vehicle
DE102011081380A1 (en) 2011-08-23 2013-02-28 Robert Bosch Gmbh Method for controlling a light emission of a headlamp of a vehicle
DE202009012589U1 (en) * 2009-09-16 2011-02-03 Sick Ag Optoelectronic sensor
US8502860B2 (en) * 2009-09-29 2013-08-06 Toyota Motor Engineering & Manufacturing North America (Tema) Electronic control system, electronic control unit and associated methodology of adapting 3D panoramic views of vehicle surroundings by predicting driver intent
US10703299B2 (en) * 2010-04-19 2020-07-07 SMR Patents S.à.r.l. Rear view mirror simulation
CN101890947A (en) * 2010-07-05 2010-11-24 邹崇毓 Automobile driving assisting system
US8717192B2 (en) 2010-10-08 2014-05-06 Navteq B.V. Method and system for using intersecting electronic horizons
DE102010063567A1 (en) * 2010-12-20 2012-06-21 Bayerische Motoren Werke Aktiengesellschaft Display system for a motor vehicle
US8781644B2 (en) * 2011-03-21 2014-07-15 Denso Corporation Method and apparatus for recognizing shape of road for vehicles
US8890674B2 (en) * 2011-06-07 2014-11-18 Continental Automotive Systems, Inc. Driver assistance detection system
DE102011081382A1 (en) * 2011-08-23 2013-02-28 Robert Bosch Gmbh Method and device for changing a light emission of at least one headlight of a vehicle
DE102011081371A1 (en) 2011-08-23 2013-02-28 Robert Bosch Gmbh Method and device for controlling a light emission of a headlamp of a vehicle
KR101288069B1 (en) * 2011-11-14 2013-07-18 현대모비스 주식회사 Parking assistance system using projector and method thereof
DE102012002333A1 (en) * 2012-02-07 2013-08-08 Daimler Ag Method for controlling optical axis of beam of headlight distribution of vehicle, involves directing optical axis of beam as warning function to border lane boundary, when lane boundary is exceeded
DE102012003158A1 (en) * 2012-02-17 2013-08-22 Daimler Ag Method for projecting laser light image in environment around vehicle during driving vehicle, involves projecting manipulated contents of captured image to position in predetermined environment by using laser light beams
JP5492962B2 (en) * 2012-09-28 2014-05-14 富士重工業株式会社 Gaze guidance system
DE102012024627A1 (en) * 2012-12-17 2014-06-18 GM Global Technology Operations LLC (n. d. Gesetzen des Staates Delaware) Method for controlling headlamp with adjustable light direction in motor vehicle, involves detecting unlighted road user over camera system, lidar system or radar system
DE102013001282A1 (en) * 2013-01-25 2014-07-31 GM Global Technology Operations LLC (n. d. Ges. d. Staates Delaware) Method for controlling bending light in motor car, involves performing temporary change of color emitted by cornering light, if road users relative to motor car is in direction in which cornering light performs lighting
DE102014000935A1 (en) * 2014-01-22 2015-07-23 Johannes Reusch Method and arrangement for generating a traffic light display for vehicles
DE102014016336A1 (en) * 2014-11-05 2016-05-12 Audi Ag System and method for controlling the light distribution of a laser-based pixel light system of at least one vehicle-side headlamp device
KR102383425B1 (en) * 2014-12-01 2022-04-07 현대자동차주식회사 Electronic apparatus, control method of electronic apparatus, computer program and computer readable recording medium
DE102014019420A1 (en) * 2014-12-22 2016-06-23 GM Global Technology Operations LLC (n. d. Ges. d. Staates Delaware) Motor vehicle headlight system, motor vehicle, method for operating a motor vehicle headlight system and computer program product
DE102015201764B4 (en) * 2015-02-02 2018-04-05 Volkswagen Aktiengesellschaft Method and driver assistance system for generating a light distribution by a vehicle for outputting a driving instruction
DE102015201767B4 (en) 2015-02-02 2021-12-16 Volkswagen Aktiengesellschaft Method for generating a light distribution for the output of driving instructions for another vehicle
DE102015115354B4 (en) * 2015-09-11 2025-02-20 HELLA GmbH & Co. KGaA Method and device for projecting a temporary road marking onto a road
DE102016001692B4 (en) 2016-02-13 2022-01-27 Audi Ag Driver assistance system for controlling light emission from a vehicle headlight device
US10328896B2 (en) * 2017-05-18 2019-06-25 Ford Global Technologies, Llc Vehicle theft avoidance systems and associated methods
CN109887334A (en) * 2017-12-01 2019-06-14 奥迪股份公司 Vehicle drive assist system and method
JP2019220009A (en) * 2018-06-21 2019-12-26 ルネサスエレクトロニクス株式会社 Semiconductor device, message image output method, and program
KR102087359B1 (en) * 2018-10-05 2020-03-10 성균관대학교산학협력단 Methods and apparatuses for attracting driver's eyes based on vehicle path
JP6642681B2 (en) * 2018-11-06 2020-02-12 日本精機株式会社 Vehicle information projection system and vehicle information projection method
DE102020101020B4 (en) 2020-01-17 2024-03-21 Audi Aktiengesellschaft motor vehicle

Family Cites Families (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH02231245A (en) 1989-03-03 1990-09-13 Nariyasu Shiba Advance path forecasting device utilizing light beam for automobile
JPH06255399A (en) 1993-03-04 1994-09-13 Mazda Motor Corp Vehicle display
JP3522317B2 (en) * 1993-12-27 2004-04-26 富士重工業株式会社 Travel guide device for vehicles
US5719568A (en) * 1995-11-28 1998-02-17 Adams; Peter J. Perimeter light system
JP3630252B2 (en) * 1996-02-07 2005-03-16 株式会社エクォス・リサーチ Navigation device
US6272431B1 (en) * 1997-04-29 2001-08-07 Thomas Zamojdo Method for displaying a map in a vehicle en-route guidance system
JP3617307B2 (en) * 1998-04-20 2005-02-02 日産自動車株式会社 Vehicle lighting device
US6198386B1 (en) * 1998-10-02 2001-03-06 White, Ii Locke Vehicle guidance assembly and method
JP3864406B2 (en) * 1999-01-26 2006-12-27 マツダ株式会社 Vehicle display device
CN2379782Y (en) * 1999-02-26 2000-05-24 三水市好帮手电子科技有限公司 Radar type instrument for guiding driver to back a car
US6486798B2 (en) * 2000-05-11 2002-11-26 Rastar Corporation System and method of preventing aircraft wing damage
US6963293B1 (en) * 2000-05-11 2005-11-08 Rastar Corporation System and method of preventing aircraft wingtip ground incursion
JP2002156233A (en) * 2000-11-16 2002-05-31 Fuji Heavy Ind Ltd In-vehicle information display device
US6741186B2 (en) * 2001-05-24 2004-05-25 Phillip N. Ross Infrared road line detector
JP3700614B2 (en) 2001-06-22 2005-09-28 株式会社豊田自動織機 Parking assistance device
JP2003123196A (en) * 2001-10-10 2003-04-25 Denso Corp Device and program for monitoring circumference of vehicle
JP4161584B2 (en) * 2002-02-07 2008-10-08 トヨタ自動車株式会社 Safety device for moving objects
JP2003291688A (en) * 2002-04-03 2003-10-15 Denso Corp Display method, driving support device, program
US6801843B2 (en) * 2002-05-24 2004-10-05 Ford Global Technologies, Llc Vehicle pre-crash sensing based conic target threat assessment system
JP2004185105A (en) * 2002-11-29 2004-07-02 Nissan Motor Co Ltd Obstacle informing device
JP2004220257A (en) * 2003-01-14 2004-08-05 Toyota Motor Corp Vehicle safety assist device and vehicle safety assist system
JP4134785B2 (en) * 2003-03-28 2008-08-20 株式会社デンソー Display device
JP4239771B2 (en) * 2003-09-18 2009-03-18 日産自動車株式会社 VEHICLE DRIVE OPERATION ASSISTANCE DEVICE AND VEHICLE HAVING VEHICLE DRIVE OPERATION ASSISTANCE DEVICE

Cited By (36)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE102006050548B4 (en) * 2006-10-26 2014-05-08 Bayerische Motoren Werke Aktiengesellschaft Procedure for warning other road users
DE102006050548A1 (en) * 2006-10-26 2008-04-30 Bayerische Motoren Werke Ag Motorist warning method for motor vehicle i.e. passenger car , involves determining projection point outside of motor vehicle for production of projection object based on situation information, and selectively producing object in point
US8248219B2 (en) 2007-01-19 2012-08-21 Denso Corporation Interior information display apparatus and light irradiation apparatus used for the same
WO2008155006A1 (en) * 2007-06-16 2008-12-24 Bayerische Motoren Werke Aktiengesellschaft Method for supporting a driver of a motor vehicle when driving the motor vehicle through a narrow passage and/or for maintaining a safe distance from a preceding vehicle
US20100094541A1 (en) * 2007-06-16 2010-04-15 Bayerische Motoren Werke Aktiengesellschaft Method for Assisting a Motor Vehicle Driver When Driving Through a Narrow Passage and/or for Maintaining a Safe Distance from a Vehicle in Front
US10214143B2 (en) 2007-06-16 2019-02-26 Bayerische Motoren Werke Aktiengesellschaft Method for assisting a motor vehicle driver when driving through a narrow passage and/or for maintaining a safe distance from a vehicle in front
WO2009066149A3 (en) * 2007-11-20 2009-09-17 Toyota Jidosha Kabushiki Kaisha Collision prediction device and method of controlling the collision prediction device
US20140028451A1 (en) * 2011-04-20 2014-01-30 Toyota Jidosha Kabushiki Kaisha Vehicle periphery alert device
US10150407B2 (en) 2011-04-20 2018-12-11 Toyota Jidosha Kabushiki Kaisha Vehicle periphery alert device
US9583002B2 (en) 2011-12-14 2017-02-28 Toyota Jidosha Kabushiki Kaisha Vehicle information transmitting device
DE102012219572A1 (en) * 2012-10-25 2014-04-30 Bayerische Motoren Werke Aktiengesellschaft Device for representing reference on display of vehicle such as passenger car for other road users, has display with spatial arrangement for viewer outside of display to determine reference for display, based on received information
US11935426B2 (en) 2013-09-05 2024-03-19 Crown Equipment Corporation Dynamic operator behavior analyzer
US11694572B2 (en) 2013-09-05 2023-07-04 Crown Equipment Corporation Dynamic operator behavior analyzer
EP3671174B1 (en) * 2013-09-05 2023-11-01 Crown Equipment Corporation Dynamic operator behavior analyzer
US10991266B2 (en) 2013-09-05 2021-04-27 Crown Equipment Corporation Dynamic operator behavior analyzer
EP3042174B1 (en) * 2013-09-05 2020-01-15 Crown Equipment Corporation Dynamic operator behavior analyzer
US10522054B2 (en) 2013-09-05 2019-12-31 Crown Equipment Corporation Dynamic operator behavior analyzer
US10408634B2 (en) * 2014-03-25 2019-09-10 Jaguar Land Rover Limited Navigation system
WO2016110365A1 (en) 2015-01-08 2016-07-14 Volkswagen Aktiengesellschaft Method and driver assistance system for assisting a driver of a vehicle
US10220771B2 (en) 2015-01-08 2019-03-05 Volkswagen Aktiengesellschaft Method and driver assistance system for assisting a driver of a vehicle
DE102015200131A1 (en) 2015-01-08 2016-07-14 Volkswagen Aktiengesellschaft Method and driver assistance system for assisting a driver of a vehicle
US11744766B2 (en) 2016-08-05 2023-09-05 Sony Corporation Information processing apparatus and information processing method
US10765588B2 (en) * 2016-08-05 2020-09-08 Sony Corporation Information processing apparatus and information processing method
US20190307632A1 (en) * 2016-08-05 2019-10-10 Sony Corporation Information processing device, information processing method, and program
CN106335512A (en) * 2016-11-04 2017-01-18 大连文森特软件科技有限公司 Driving Assistance System Based on AR Augmented Reality and Vehicle Violation Record Query
KR20190112274A (en) * 2017-02-03 2019-10-04 퀄컴 인코포레이티드 Maintain occupant awareness in the vehicle
EP3577606B1 (en) * 2017-02-03 2022-09-07 Qualcomm Incorporated Maintaining occupant awareness in vehicles
KR102487922B1 (en) 2017-02-03 2023-01-12 퀄컴 인코포레이티드 Maintaining occupant awareness in the vehicle
US10082869B2 (en) * 2017-02-03 2018-09-25 Qualcomm Incorporated Maintaining occupant awareness in vehicles
US11238737B2 (en) 2017-03-09 2022-02-01 Bayerische Motoren Werke Aktiengesellschaft Motor vehicle comprising a lighting module for generating a set of symbols
CN109643490A (en) * 2017-03-29 2019-04-16 松下电器(美国)知识产权公司 Controller of vehicle, control method for vehicle and program
EP3738316A4 (en) * 2018-01-12 2021-12-22 Intelligent Security Systems Corporation SYSTEMS AND PROCESSES FOR IMAGE-BASED LIGHT OUTPUT
WO2019139959A1 (en) 2018-01-12 2019-07-18 Intelligent Security Systems Corporation Systems and methods for image-based light output
US11120691B2 (en) 2019-12-02 2021-09-14 Denso Corporation Systems and methods for providing warnings to surrounding vehicles to avoid collisions
US20240005795A1 (en) * 2021-03-18 2024-01-04 Huawei Technologies Co., Ltd. Light Projection Apparatus and Method, and Storage Medium
US12307899B2 (en) * 2021-03-18 2025-05-20 Shenzhen Yinwang Intelligent Technologies Co., Ltd. Light projection apparatus and method, and storage medium

Also Published As

Publication number Publication date
CN100408383C (en) 2008-08-06
US20060097858A1 (en) 2006-05-11
CN1810547A (en) 2006-08-02
DE102005051049A1 (en) 2006-05-04
US7394356B2 (en) 2008-07-01
JP2006127055A (en) 2006-05-18

Similar Documents

Publication Publication Date Title
US20060087416A1 (en) Information providing device for vehicle
JP6673299B2 (en) Steering support device
US12275425B2 (en) Alert apparatus for host vehicle
JP4044031B2 (en) Vehicle travel support device
US6708087B2 (en) Display system for vehicle
JP4730267B2 (en) Visibility state determination device for vehicle
CN103459225B (en) Driving assist system
US12103520B2 (en) Vehicle driving support system and vehicle driving support method
US20180018939A1 (en) Display system replacing side mirror and method for controlling output brightness thereof
JP2019001183A (en) Lane change support device
WO2019021471A1 (en) Display control method and display control device
US20130058116A1 (en) Method and device for changing a light emission of at least one headlight of a vehicle
CN106364480A (en) Vehicle control device
JP2005078414A (en) Vehicle travel support device
CN110356317B (en) Vehicle reminding method and system and vehicle
JP2008094151A (en) Vehicular light control system
JP2005161977A (en) Vehicular travel supporting device
JP4983564B2 (en) Vehicle tracking device
JP2019003235A (en) Target information acquisition device
JP4492725B2 (en) Light control device
KR20190066115A (en) Vehicle and method for controlling thereof
CN110371110B (en) Vehicle control method and system and vehicle
JP4434864B2 (en) Vehicle lamp control device
KR101395616B1 (en) Device for displaying distance between cars
WO2018199075A1 (en) Vehicle control device and vehicle control method

Legal Events

Date Code Title Description
AS Assignment

Owner name: DENSO CORPORATION, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KUMABE, HAJIME;TSURU, NAOHIKO;NISHIMURA, TAKAO;REEL/FRAME:017027/0499

Effective date: 20050826

STCB Information on status: application discontinuation

Free format text: EXPRESSLY ABANDONED -- DURING EXAMINATION

点击 这是indexloc提供的php浏览器服务,不要输入任何密码和下载