Spatial image cluster for vehicle
11945308 ยท 2024-04-02
Assignee
Inventors
Cpc classification
G02B3/0056
PHYSICS
H04N13/302
ELECTRICITY
B60K2360/143
PERFORMING OPERATIONS; TRANSPORTING
B60K35/211
PERFORMING OPERATIONS; TRANSPORTING
G02B30/40
PHYSICS
B60K35/00
PERFORMING OPERATIONS; TRANSPORTING
G02B3/0062
PHYSICS
B60K2360/347
PERFORMING OPERATIONS; TRANSPORTING
B60K35/10
PERFORMING OPERATIONS; TRANSPORTING
International classification
B60K35/00
PERFORMING OPERATIONS; TRANSPORTING
G02B27/09
PHYSICS
G02B30/40
PHYSICS
G02F1/1335
PHYSICS
G09G3/00
PHYSICS
H04N13/302
ELECTRICITY
Abstract
A cluster according to an embodiment of the disclosure includes a display and a spatial image panel. The display is installed in the vehicle to output predetermined information as a 2D image. The spatial image panel is configured to output a 3D image in a predetermined space in front. The spatial image panel includes a first lens array, a second lens array, and a refractive medium. The first lens array is disposed adjacent to the display and includes a plurality of first lenses arranged on the same plane. The second lens array is disposed in parallel with the first array so that the first lenses and second lenses overlap each other. The refractive medium is disposed between the first lens array and the second lens array.
Claims
1. A spatial image cluster for a vehicle comprising: a display configured to output a 2D image; and a spatial image panel disposed in front of the display and configured to selectively output the 2D image as a 3D image, the spatial image panel comprising: a first lens array disposed adjacent to the display and comprising a plurality of first lenses arranged on a same plane; a second lens array comprising a plurality of second lenses having a same arrangement as the plurality of first lenses of the first lens array such that respective lenses of the plurality of first lenses and the plurality of second lenses are correspondingly aligned; a refractive medium configured to be disposed between the first lens array and the second lens array; and a switching layer, wherein voltages applied to the switching layer are controllable to change refraction directions of light transmitted by the spatial image panel to selectively output different portions of the 2D image as 2D or 3D, wherein the switching layer is embedded in the first lens array and is configured to change a refraction direction of transmitted light according to the applied voltage.
2. The spatial image cluster of claim 1, wherein the plurality of first lenses and the plurality of second lenses are configured to be spaced apart from each other by a same distance.
3. The spatial image cluster of claim 1, wherein a spacing between the spatial image panel and the display is adjustable.
4. The spatial image cluster of claim 3, further comprising a rail and a sliding unit coupled to the spatial image panel and configured to move along the rail to adjust the spacing.
5. The spatial image cluster of claim 4, wherein the sliding unit is coupled to the rail via a screw thread formed on the rail such that the sliding unit is configured to move along a longitudinal direction of the rail in response to rotation of the rail.
6. The spatial image cluster of claim 1, further comprising: an air touch sensor configured to detect approach of a user's input to the 3D image output through the spatial image panel.
7. The spatial image cluster of claim 6, wherein the air touch sensor detects a user's touch to the 3D image output through the spatial image panel.
8. The spatial image cluster of claim 7, wherein a color of an area of the 3D image to which the user's touch is detected is changed to a predetermined color.
9. The spatial image cluster of claim 1, wherein the switching layer is provided between the first lens array and the second lens array.
10. The spatial image cluster of claim 1, wherein the switching layer is provided between the first lens array and the display.
11. A spatial image cluster for a vehicle comprising: a display configured to output a 2D image; and a spatial image panel disposed in front of the display and configured to selectively output the 2D image as a 3D image, the spatial image panel comprising a first lens array and a second lens array arranged on parallel planes, wherein the first lens array is between the display and the second lens array; a refractive medium configured to be disposed between the first lens array and the second lens array and a switching layer, wherein voltages applied to the switching layer are controllable to selectively output different portions of the 2D image as 2D or 3D, wherein the switching layer is embedded in the first lens array and is configured to change a refraction direction of transmitted light according to the applied voltage.
12. The spatial image cluster of claim 11, wherein a distance between the spatial image panel and the display is adjustable.
13. The spatial image cluster of claim 11, wherein the refractive medium has a different refractive index than air.
14. The spatial image cluster of claim 11, wherein the switching layer is provided between the first lens array and the second lens array.
15. The spatial image cluster of claim 11, wherein the switching layer is provided between the first lens array and the display.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1) The above and other aspects, features, and advantages of the present disclosure will become apparent from the detailed description of the following aspects in conjunction with the accompanying drawings, in which:
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12)
DETAILED DESCRIPTION
(13) Hereinafter, in order to describe the disclosure, embodiments according to the disclosure will be described in more detail with reference to the accompanying drawings. Like reference numerals refer to like elements throughout the detailed description.
(14)
(15) As shown in
(16) The cluster 1 according to an embodiment of the disclosure may output a variety of information such as RPM, driving speed, fuel efficiency, fuel amount, and mileage of the vehicle. The cluster 1 according to an embodiment of the disclosure outputs predetermined information as a 3D image in a specific space between the driver and the cluster 1.
(17) That is, the image information output through the cluster 1 may have a form of the 3D image output in the specific space.
(18) An interface of a specific 3D image may be implemented in an appropriate form according to an embodiment to which the disclosure is applied.
(19) In an embodiment of the disclosure, a case in which the cluster 1 is positioned in front of the driver will be described as an example.
(20)
(21) As shown in
(22)
(23) As shown in
(24) The display 100 may be constituted of a screen that outputs a 2D image or a set of light sources. The spatial image panel 200 is disposed in front of the display 100 and transmits an image output through the display 100 to output the transmitted image as a 3D image in a predetermined space.
(25) In an embodiment of the disclosure, the spatial image panel 200 may be constituted of a micro-lens array (MLA).
(26) The spatial image panel 200 includes a first lens array 210, a second lens array 220, and a refractive medium 230. The first lens array 210 is provided by disposing first lenses 212 in a predetermined pattern. The second lens array 220 is provided by disposing second lens 214 in a predetermined pattern. The first lens 212 and the second lens 214 correspond one-to-one, respectively, and are positioned so that light passing through the first lens 212 and the second lens 214 overlaps.
(27) Specifically, the display 100 forms a single plane, and a 2D image is output through one surface of the display 100. In addition, the spatial image panel 200 is also formed as a surface that transmits the 2D image output through one surface of the display 100. The spatial image panel 200 is constituted of the first lens array 210 and the second lens array 220 in which the first lens 212 and the second lens 214 are gathered to form a plane, respectively. The first lens 212 and the second lens 214 included in each of the first lens array 210 and the second lens array 220 are respectively arranged to correspond one-to-one, and light passing through the first lens array 210 and the second lens array 220 overlaps.
(28) The refractive medium 230 is disposed between the first lens array 210 and the second lens array 220. Alternatively, the first lens array 210 and the second lens array 220 may be disposed in a space surrounded by the refractive medium 230. The refractive medium 230 has a refractive index different from that of air. The refraction medium 230 serves to enlarge the viewing angle of light passing between the first lens array 210 and the second lens array 220.
(29) As shown in
(30) Accordingly, the output image 10 is formed as a 3D image in a space between the cluster 1 and the driver and in a predetermined space. The output image 10 may include information such as a vehicle or environment.
(31) More specifically, the first lens array 210 and the second lens array 220 have the first lenses 212 and the second lenses 214 arranged in the same pattern, respectively. The first lens 212 of the first lens array 210 and the second lens 214 of the second lens array 220 overlap, respectively. The first lens array 210 and the second lens array 220 may be symmetrical to each other with the refractive medium 230 interposed therebetween.
(32) That is, a 2D image is output through one surface of the display 100. The spatial image panel 200 is installed at a predetermined interval on one surface on which the image of the display 100 is output. The 2D image output through one surface of the display 100 passes through the spatial image panel 200 and is output as a 3D image in a predetermined space. In the cluster 1 according to an embodiment of the disclosure, the 3D image output through the spatial image panel 200 is formed in a specific space located between the driver and the cluster provided on the vehicle body.
(33)
(34) As shown in
(35) Specifically, a rail 240 is provided on one surface on which the 2D image of the display 100 is output, along a direction in which the image is output from the display 100. At least one rail 240 may be provided, and a thread may be formed on the outer periphery of the rail 240.
(36) The sliding unit 250 that is transferred along the longitudinal direction of the rail 240 as the rail 240 rotates is provided. The first lens array 210, the second lens array 220, and the refractive medium 230 may be fixed to the sliding unit 250. That is, the spatial image panel 200 may be fixed to the sliding unit 250.
(37) Accordingly, as the rail 240 rotates, the spatial image panel 200 moves close to or away from the display 100.
(38) When the spatial image panel 200 moves close to the display 100, a position at which a 3D image is formed is provided close to the cluster 1. Conversely, when the spatial image panel 200 moves away from the display 100, the position at which the 3D image is formed moves away from the cluster 1. Accordingly, when the spatial image panel 200 is operated far from the display 100, the driver can see the output image 10 output as the 3D image more closely.
(39) Accordingly, in an embodiment of the disclosure, the spatial image panel 200 may be operated in conjunction with a variable indicating a vehicle state such as a mileage of the vehicle.
(40) Specifically, when the speed of the vehicle is greater than or equal to a predetermined speed, the spatial image panel 200 may be configured to be closer to the display 100. Alternatively, the spatial image panel 200 may be configured to move away from the display 100 when the vehicle speed is less than or equal to a predetermined speed or when the vehicle is placed under a predetermined situation, such as when entering a child protection area.
(41) Through this, by varying the size of the 3D output image 10 that the driver can identify according to the driving situation or state of the vehicle, there is an effect of further enhancing the user's identification.
(42)
(43) As shown in
(44) The switching layer 400 may be provided in the form of a panel, and the degree of refraction of transmitted light changes according to the voltage of an applied current. The switching layer 400 may be disposed between the first lens array 210 and the second lens array 220. Alternatively, the switching layer 400 may be disposed between the first lens array 210 and the display 100.
(45) The switching layer 400 directly transmits light as shown in
(46) In addition, when the voltage applied to the switching layer 400 is changed as shown in
(47) That is, by adjusting the voltage applied to the switching layer 400, a specific portion or the whole of the cluster 1 may selectively output one of 2D or 3D images.
(48) The output 3D image may be a specific area or specific information of the cluster 1 designated by the user as described above. The user can selectively convert the specific area and specific information of the cluster 1 into a 2D or 3D image.
(49)
(50) As shown in
(51) The air touch sensor 300 detects that the driver's hand enters a predetermined space. In addition, the air touch sensor 300 first collects the position of the 3D output image 10 output through the cluster 1, and determines whether the driver's hand approaches the 3D output image 10.
(52) That is, the air touch sensor 300 is interlocked with a controller and transmits object information detected from the front to the controller. Information of the output 3D output image 10, in particular, image-formed position information is stored in the controller. In addition, information for each portion of the output image 10 to be output is stored.
(53) When the user's hand detected through the air touch sensor 300 reaches a specific position of the 3D output image 10, the controller may operate a specific device according to a predetermined configuration.
(54) Specifically, when the driver touches the 3D output image 10 formed on the space, the air touch sensor 300 detects the touch as an input signal, and the controller interprets the input signal and performs an appropriate output.
(55) For example, the above operation may be an operation such as adjusting the size of the 3D output image 10, adjusting the volume of the 3D output image, or adjusting the brightness of the 3D output image 10. Alternatively, the above operation may be implemented in various forms, such as selecting a type of an image to be output.
(56) In the above, embodiments of the disclosure have been described with drawings, which are exemplary and the disclosure is not limited to the above-described embodiments and drawings. It is apparent that those of ordinary skill in the art can modify the embodiments of the disclosure within the scope of the technical idea of the disclosure. In addition, although an action or effect according to the configuration of the disclosure is not explicitly described while describing an embodiment of the disclosure, it is natural that even the effects predictable by the configuration should be recognized.