INPUT/OUTPUT DEVICE, INPUT/OUTPUT PROGRAM, AND INPUT/OUTPUT METHOD
20170310944 · 2017-10-26
Assignee
Inventors
Cpc classification
G06F3/017
PHYSICS
G06F3/011
PHYSICS
H04N2213/008
ELECTRICITY
G06F1/1637
PHYSICS
G06V10/24
PHYSICS
G06V40/28
PHYSICS
H04N13/239
ELECTRICITY
H04N13/271
ELECTRICITY
H04N13/25
ELECTRICITY
International classification
Abstract
An object of the present invention is to provide an I/O device, an I/O program, and an I/O method which can be used even by a user who has strabismus, etc. Another object of the present invention is to provide an I/O device, an I/O program, and an I/O method for adjusting the strabismus or eyesight, etc. of a user. In addition, a display device can generate a stereoscopic image, a depth level sensor measures a distance to an object, and a control unit performs display on the display device in accordance with the depth level sensor. A display adjustment mechanism adjusts the angle of the display device.
Claims
1. An I/O device comprising: a display device that can generate a stereoscopic image; a depth level sensor that measures a distance to an object; a control unit that performs display on the display device in accordance with the depth level sensor; and a display adjustment mechanism that adjusts the angle of the display device.
2. The I/O device according to claim 1, wherein the display adjustment mechanism can adjust the angle of at least either the vertical axis or perpendicular axis of the display device.
3. The I/O device according to claim 1, wherein the display adjustment mechanism includes a manual adjustment unit through which manual adjustment can be performed.
4. The I/O device according to claim 1, wherein the display adjustment mechanism performs the adjustment depending on determination given by the control unit.
5. The I/O device according to claim 1, wherein the display device includes a plurality of display units, and the display adjustment mechanism can adjust the plurality of display units separately.
6. The I/O device according to claim 1, wherein the display device is a head-mounted display.
7. An I/O program comprising: a display process of enabling generation of a stereoscopic image; a depth level sensor process of measuring a distance to an object; a control process of performing display in the display process in accordance with the depth level sensor process; and a display adjustment process of adjusting the angle in the display process.
8. The I/O program according to claim 7, wherein through the display adjustment process, the angle of at least either the vertical axis or the perpendicular axis in the display process can be adjusted.
9. The I/O program according to claim 7, wherein in the display adjustment process, the adjustment is performed depending on determination given by the control process.
10. The I/O program according to claim 7, wherein the display process includes a plurality of display processes, and the plurality of display processes can be separately adjusted through the display adjustment process.
11. An I/O method comprising: a display step of enabling generation of a stereoscopic image; a depth level sensor step of measuring a distance to an object; a control step of performing display in the display step in accordance with the depth level sensor step; and a display adjustment step of adjusting the angle in the display step.
12. The I/O method according to claim 11, wherein the angle of at least either the vertical axis or the perpendicular axis in the display step can be adjusted in the display adjustment step.
13. The I/O method according to claim 11, wherein at the display adjustment step, the adjustment is performed depending on determination given in the control step.
14. The I/O method according to claim 11, wherein the display step includes a plurality of display steps, and the plurality of display steps can be separately adjusted through the display adjustment step.
Description
BRIEF DESCRIPTION OF DRAWINGS
[0061]
[0062]
[0063]
[0064]
[0065]
[0066]
[0067]
[0068]
[0069]
[0070]
[0071]
[0072]
[0073]
[0074]
[0075]
[0076]
[0077]
[0078]
[0079]
[0080]
[0081]
[0082]
[0083]
[0084]
[0085]
[0086]
[0087]
[0088]
[0089]
[0090]
[0091]
REFERENCE SIGNS LIST
[0092] 100 glasses display device [0093] 220 semi-transmissive display [0094] 2203D virtual image display region (common region) [0095] 300 communication system [0096] 303 camera unit [0097] 410 infrared ray detection unit [0098] 410c manipulation region [0099] 420 gyroscope unit [0100] 430 acceleration detection unit [0101] 4103D three-dimensional space detection region [0102] 450 control unit [0103] 454 anatomy recognition unit [0104] 456 gesture recognition unit [0105] 460 event service unit [0106] 461 calibration service unit [0107] 900 I/O device [0108] H1 hand [0109] RP right shoulder joint [0110] LP left shoulder joint
DESCRIPTION OF EMBODIMENT
[0111] Hereinafter, an embodiment of the present invention is described with reference to the drawings. In the following description, the same reference signs are given to the same components. The names and functions thereof are the same. Accordingly, detailed description thereof is not repeated.
[0112] Moreover, the present invention is not limitatively applied to the following glasses display device, and can also be applied to other wearable devices, other I/O devices, display devices, televisions, monitors, projectors, and the like.
(Configuration Outline of Glasses Display Device)
[0113]
[0114] As illustrated in
[0115] As illustrated in
(Glasses Unit 200)
[0116] As illustrated in
[0117] The pair of semi-transmissive displays 220 is supported by the rim unit 211 of the glasses frame 210. The rim unit 211 is provided with the pair of display adjustment mechanisms 600. The rim unit 211 is also provided with an infrared ray detection unit 410 and a unit adjustment mechanism 500. Details of the unit adjustment mechanism 500 will be described later.
[0118] The pair of display adjustment mechanisms 600 can adjust the angle and the position of the pair of semi-transmissive displays 220 as described later. Details of the pair of display adjustment mechanisms 600 will be described later.
[0119] In the present embodiment, the pair of display adjustment mechanisms 600 of the rim unit 211 of the glasses display device 100 is provided with the pair of semi-transmissive displays 220. Not limited thereto, the pair of display adjustment mechanisms 600 of the rim unit 211 of the glasses display device 100 may be provided with lenses such as normal sunglasses lenses, ultraviolet protection lenses, or glasses lenses, and one semi-transmissive display 220 or the pair of semi-transmissive displays 220 may be separately provided.
[0120] Alternatively, the semi-transmissive display(s) 220 may be provided so as to be embedded in part of the lenses.
[0121] Furthermore, although the pair of display adjustment mechanisms 600 is provided on a side portion of the semi-transmissive displays 220, not limited thereto, the pair of display adjustment mechanisms 600 may be provided around or inside of the semi-transmissive displays 200.
[0122] Further, the present embodiment is not limited to such a glasses type, and can be applied to a hat type and other arbitrary head-mounted display devices as long as the device can be attached to the body of a person and can be arranged within the field of view of the person.
(Communication System 300)
[0123] Next, the communication system 300 is described. The communication system 300 includes a battery unit 301, an antenna module 302, a camera unit 303, a speaker unit 304, a global positioning system (GPS) unit 307, a microphone unit 308, a subscriber identity module card (SIM) unit 309, and a main unit 310.
[0124] Note that the camera unit may be provided with a CCD sensor. The speaker unit 304 may be normal earphones, and may be bone-conduction earphones. The SIM unit 309 includes a near field communication (NFC) unit, another contact-type IC card unit, and a contactless IC card unit.
[0125] As described above, the communication system 300 according to the present embodiment at least has any of the functions of a mobile phone, a smartphone, and a tablet terminal. Specifically, the communication system 300 has a phone function, an Internet function, a browser function, an e-mail function, an image taking function, and the like.
[0126] Accordingly, with the use of the glasses display device 100, the user can use a phone call function similar to that of a mobile phone by means of the communication device, the speaker, and the microphone. Moreover, because the glasses display device 100 is glasses-shaped, the user can make a phone call without using both his/her hands.
(Operation System 400)
[0127] Next, the operation system 400 includes an infrared ray detection unit 410, a gyroscope unit 420, an acceleration detection unit 430, and a control unit 450. The infrared ray detection unit 410 mainly includes an infrared ray emission element 411 and an infrared ray detection camera 412.
(Unit Adjustment Mechanism 500)
[0128] As illustrated in
[0129] The unit adjustment mechanism 500 makes a movement and adjustment in the directions of the arrow V5 and the arrow H5 according to an instruction from the control unit 450.
[0130] For example, when a predetermined gesture is recognized by the control unit 450, the unit adjustment mechanism 500 may be operated at a predetermined angle. In this case, the user can perform a predetermined gesture to adjust the angle of the infrared ray detection unit 410.
[0131] Note that, although the control unit 450 causes the unit adjustment mechanism 500 to operate in the present embodiment, not limited thereto, an adjustment unit 520 of
[0132] Next, a configuration, a processing flow, and a concept of the operation system 400 are described.
[0133] As illustrated in
[0134] Note that the control unit 450 does not need to include all the above-mentioned units, and may include one or more necessary units as appropriate. For example, the gesture data unit 455 and the calibration data unit 457 may be arranged on a cloud service, and the composition processor unit 458 may not be particularly provided.
[0135] Next,
[0136] First, as illustrated in
[0137] Subsequently, on the basis of the structure of a standard human body, an anatomic feature is recognized from the outer shape image data processed in Step S2, by the anatomy recognition unit 454. As a result, an outer shape is recognized (Step S3).
[0138] Further, on the basis of the anatomic feature obtained in Step S3, a gesture is recognized by the gesture recognition unit 456 (Step S4).
[0139] The gesture recognition unit 456 refers to gesture data recorded in the gesture data unit 455, and recognizes the gesture from the outer shape whose anatomic feature has been recognized. Note that, although it is assumed that the gesture recognition unit 456 refers to the gesture data recorded in the gesture data unit 455, not limited thereto, the gesture recognition unit 456 may refer to other arbitrary data, and may perform processing without any reference.
[0140] In such a manner as described above, a gesture of hands is recognized as illustrated in
[0141] Subsequently, the application unit 459 and the event service unit 460 carry out a predetermined event in accordance with the gesture recognized by the gesture recognition unit 456 (Step S5).
[0142] As a result, as illustrated in
[0143] Lastly, the view service unit 462, the calibration service unit 461, the graphics processor unit 463, the display processor unit 464, and the composition processor unit 458 display or virtually display an image on the semi-transmissive displays 220 (Step S6). As a result, skeletons of the hands indicating the gesture are displayed as illustrated in
[0144] Note that the 6-axis sensor driver unit 465 always detects signals from the gyroscope unit 420 and the acceleration detection unit 430, and transmits a posture condition to the display processor unit 464.
[0145] In the case where the user to whom the glasses display device 100 is attached inclines the glasses display device 100, the 6-axis sensor driver unit 465 always receives signals from the gyroscope unit 420 and the acceleration detection unit 430, and controls image display. In this control, the displayed image may be kept horizontal, and may be adjusted in accordance with the inclination.
(One Example of Detection Region and Virtual Display Region)
[0146] Next, a relation between a detection region of the infrared ray detection unit 410 of the operation system 400 and a virtual display region of the pair of semi-transmissive displays 220 is described.
[0147]
[0148] In the following, for convenience of description, a three-dimensional orthogonal coordinate system formed by an x-axis, a y-axis, and a z-axis is defined as illustrated in
[0149] As illustrated in
[0150] The three-dimensional space detection region 4103D is formed by a conical or pyramidal three-dimensional space extending from the infrared ray detection unit 410.
[0151] That is, infrared rays emitted from the infrared ray emission element 411 can be detected by the infrared ray detection camera 412, and hence the infrared ray detection unit 410 can recognize a gesture in the three-dimensional space detection region 4103D.
[0152] Moreover, although one infrared ray detection unit 410 is provided in the present embodiment, not limited thereto, a plurality of the infrared ray detection units 410 may be provided, and one infrared ray emission element 411 and a plurality of the infrared ray detection cameras 412 may be provided.
[0153] Subsequently, as illustrated in
[0154] That is, although images are respectively displayed on the semi-transmissive displays 220 of the glasses display device 100 in actuality, a right-eye image is transmitted through the semi-transmissive display 220 on the right-eye side to be recognized by the user in a three-dimensional space region 2203DR, and a left-eye image is transmitted through the semi-transmissive display 220 on the left-eye side to be recognized by the user in a three-dimensional space region 2203DL. As a result, the two recognized images are composited with each other in the brain of the user, whereby the user can recognize the two images as a virtual image in the virtual image display region 2203D.
[0155] Moreover, the virtual image display region 2203D is displayed using any of a frame sequential method, a polarization method, a linear polarization method, a circular polarization method, a top-and-bottom method, a side-by-side method, an anaglyph method, a lenticular method, a parallax barrier method, a liquid crystal parallax barrier method, a two-parallax method, and a multi-parallax method using three or more parallaxes.
[0156] Moreover, in the present embodiment, the virtual image display region 2203D includes a space region common to the three-dimensional space detection region 4103D. In particular, as illustrated in
[0157] Note that the shape and size of the virtual image display region 2203D can be arbitrarily adjusted by a display method on the pair of semi-transmissive displays 220.
[0158] Moreover, as illustrated in
(Other Examples of Detection Region and Virtual Display Region)
[0159] Next,
[0160] For example, as illustrated in
[0161] As illustrated in
[0162] In this case, the virtual image display region 2203D outputted by the I/O device 900 is generated as a space region common to the three-dimensional space detection region 4103D.
[0163] Moreover, as illustrated in
[0164] Also in this case, the virtual image display region 2203D outputted by the I/O device 900 is generated as a space region common to the three-dimensional space detection region 4103D.
[0165] Then, as illustrated in
[0166] Moreover, although not illustrated, the I/O device 900 may be arranged on the upper side (y-axis positive direction side) of the three-dimensional space detection region 4103D, and the virtual image display region 2203D may be outputted in the vertical downward direction (y-axis negative direction). The virtual image display region 2203D may be outputted in the horizontal direction (x-axis direction). Like a projector or a movie theater, the virtual image display region 2203D may be outputted from the upper back side (the z-axis positive direction and the y-axis positive direction).
(Manipulation Region and Gesture Region)
[0167] Next, a manipulation region and a gesture region in the detection region are described.
[0168] First, as illustrated in
[0169] Moreover, as illustrated in
[0170] That is, as illustrated in
[0171] Then, an overlapping space region of all of: the three-dimensional space detection region 4103D of the infrared ray detection unit 410; a region in which a virtual image display region can exist (in
[0172] Moreover, a portion other than the manipulation region 410c in the three-dimensional space detection region 4103D is set as a gesture region 410g, the portion overlapping with the region obtained by integrating the arm movement region L and the arm movement region R.
[0173] Here, the manipulation region 410c has a stereoscopic shape whose farthest surface in the depth level direction is an arch-like curved surface that is convex in the depth level direction (z-axis positive direction), whereas the virtual image display region 2203D has a stereoscopic shape whose farthest surface in the depth level direction is a planar surface. Due to such a difference in the shape of the farthest surface between the two regions, the user physically feels a sense of discomfort during the manipulation. In order to solve the sense of discomfort, adjustment is performed in a calibration process. Moreover, the details of the calibration process are described below.
(Description of Calibration)
[0174] Next, the calibration process is described.
[0175] As illustrated in
[0176] Moreover, in the calibration process, the finger length, the hand length, and the arm length, which are different for each user, are also adjusted.
[0177] Hereinafter, description is given with reference to
[0178] That is, because the finger length, the hand length, and the arm length are different for each user, the manipulation region 410c is adjusted to suit each user.
[0179] Then, in the glasses display device 100, a display position of the virtual image display region 2203D is determined (Step S12). That is, if the virtual image display region 2203D is arranged outside of the manipulation region 410c, a user's manipulation becomes impossible, and hence the virtual image display region 2203D is arranged inside of the manipulation region 410c.
[0180] Subsequently, the maximum region of the gesture region 410g is set within the three-dimensional space detection region 4103D of the infrared ray detection unit 410 of the glasses display device 100 so as not to overlap with the display position of the virtual image display region 2203D (Step S13).
[0181] Note that it is preferable that the gesture region 410g be arranged so as not to overlap with the virtual image display region 2203D and be provided with a thickness in the depth direction (z-axis positive direction).
[0182] In the present embodiment, the manipulation region 410c, the virtual image display region 2203D, and the gesture region 410g are set in such a manner as described above.
[0183] Next, calibration of the virtual image display region 2203D in the manipulation region 410c is described.
[0184] In the case where it is determined that the finger(s), the hand(s), or the arm(s) of the user exist around the outside of the virtual image display region 2203D in the manipulation region 410c, such rounding as if the finger(s), the hand(s), or the arm(s) of the user existed inside of the virtual image display region 2203D is performed (Step S14).
[0185] As illustrated in
[0186] Hence, if a signal from the infrared ray detection unit 410 is used without being processed, even if the tips of his/her hands go out of the virtual image display region 2203D, the user has difficulty in physically feeling such a state.
[0187] Accordingly, in the process of Step S14 in the present embodiment, the signal from the infrared ray detection unit 410 is processed such that the tips of his/her hands that protrude to the outside of the virtual image display region 2203D are corrected to exist within the virtual image display region 2203D.
[0188] As a result, in the state where the user maximally stretches or slightly bends both his/her arms, a manipulation from the central part to the end part in the planar virtual image display region 2203D with a depth is possible.
[0189] Note that, although the virtual image display region 2203D is formed by a three-dimensional space region whose farthest surface in the depth level direction is a planar surface in the present embodiment, not limited thereto, the virtual image display region 2203D may be formed by a three-dimensional space region that is a curved surface having a shape along the farthest surfaces in the depth level direction of the farthest surface regions L and R in the depth level direction. As a result, in the state where the user maximally stretches or slightly bends both his/her arms, a manipulation from the central part to the end part in the planar virtual image display region 2203D with a depth is possible.
[0190] Further, the semi-transmissive displays 220 display a rectangular image in the virtual image display region 2203D. For example, as illustrated in
[0191] Subsequently, an instruction to the effect that “please surround the displayed image with your fingers” is displayed on the semi-transmissive displays 220 (Step S16). Here, a finger-shaped image may be softly displayed in the vicinity of the image, and a vocal instruction from the speaker may be given to the user instead of such display on the semi-transmissive displays 220.
[0192] According to the instruction, the user places his/her fingers on a portion of the image as illustrated in
[0193] Note that, in the above example, the user defines a rectangular with his/her fingers, and places the rectangular thus defined on the rectangular of the outer edge of the image. For this reason, the visual recognition size and position of the rectangular defined by his/her fingers is made coincident with the visual recognition size and position of the rectangular of the outer edge of the image. However, the method of defining a shape with fingers is not limited thereto, and may be other arbitrary methods such as a method of tracing the outer edge of the displayed image with a finger and a method of pointing to a plurality of points on the outer edge of the displayed image with a finger. Moreover, these methods may be applied to images having a plurality of sizes.
[0194] Note that, although only the case of the glasses display device 100 is taken in the above description of the calibration process, in the case of the I/O device 900, an image may be displayed in the process of Step S11, and a correlation between the displayed image and the infrared ray detection unit 410 may be adjusted in the process of Step S17.
(Finger, Palm, and Arm Recognition)
[0195] Next, finger recognition is described, followed by description of palm recognition and arm recognition in the stated order.
[0196] As illustrated in
[0197] Then, image data is replaced with a distance on a pixel basis by the infrared ray detection unit 410 (Step S23). In this case, the luminance of the infrared ray is inversely proportional to the cube of the distance. A depth map is created using this fact (Step S24).
[0198] Subsequently, an appropriate threshold is set to the created depth map. Then, the image data is binarized (Step S25). That is, noise is removed from the depth map.
[0199] Subsequently, a polygon having about 100 vertexes is created from the binarized image data (Step S26). Then, a new polygon having a larger number of vertexes p, is created using a low-pass filter (LPF) such that the vertexes become smoother, whereby an outer shape OF of the hand illustrated in
[0200] Note that, although the number of vertexes that are extracted from the data binarized in Step S26 in order to create a polygon is about 100 in the present embodiment, not limited thereto, the number of vertexes may be 1,000 or other arbitrary numbers.
(Finger Recognition)
[0201] A convex hull is extracted using Convex Hull from the set of the vertexes p.sub.r of the new polygon created in Step S27 (Step S28).
[0202] After that, a vertex p.sub.0 common between the new polygon created in Step S27 and the convex hull created in Step S28 is extracted (Step S29). The common vertex p.sub.0 itself thus extracted can be used as a tip point of the finger.
[0203] Further, another point calculated on the basis of the position of the vertex p.sub.0 may be used as the tip point of the finger. For example, as illustrated in
[0204] Then, as illustrated in
[0205] A similar process is performed on all the fingers, whereby the skeletons of all the fingers are obtained. As a result, the pose of the hand can be recognized. That is, it can be recognized which of the thumb, the index finger, the middle finger, the ring finger, and the little finger is stretched and which thereof is bent.
[0206] Subsequently, a difference in the pose of the hand is detected in comparison with image data of several frames taken immediately before (Step S32). That is, movement of the hand can be recognized through the comparison with the image data of the several frames taken immediately before.
[0207] Subsequently, the recognized shape of the hand is event-delivered as gesture data to the event service unit 460 (Step S33).
[0208] Subsequently, a behavior according to the event is carried out by the application unit 459 (Step S34).
[0209] Subsequently, drawing in a three-dimensional space is requested by the view service unit 462 (Step S35).
[0210] The graphics processor unit 463 refers to the calibration data unit 457 using the calibration service unit 461, and corrects the displayed image (Step S36).
[0211] Lastly, the resultant image is displayed on the semi-transmissive displays 220 by the display processor unit 464 (Step S37).
[0212] Note that, although the base point of each finger is detected through the process of Step S30 and the process of Step S31 in the present embodiment, the method of detecting the base point is not limited thereto. For example, first, the length of the reference line segment PP.sub.1 is calculated, the reference line segment PP.sub.1 connecting the pair of vertexes p.sub.1 that are adjacent to the vertex p.sub.0 on one side and another side of the vertex p.sub.0, respectively. Then, the length of a line segment connecting the pair of vertexes p.sub.2 on the one side and the another side is calculated. Similarly, the length of each line segment connecting a pair of vertexes on the one side and the another side is calculated in order from vertexes positioned closer to the vertex p.sub.0 to vertexes positioned farther therefrom. Such line segments do not intersect with one another inside of the outer shape OF, and are substantially parallel to one another. In the case where the vertexes at both the ends of such a line segment are in the portion of the finger, the length of the line segment corresponds to the width of the finger, and hence the amount of change thereof is small. Meanwhile, in the case where at least any of the vertexes at both the ends of such a line segment reaches the portion of the valley between the fingers, the amount of change of the length becomes larger. Accordingly, a line segment that has the length whose amount of change does not exceed a predetermined amount and is the farthest from the vertex p.sub.0 is detected, and one point on the detected line segment is extracted, whereby the base point can be determined.
(Palm Recognition)
[0213] Next,
[0214] As illustrated in
[0215] Next,
[0216] As illustrated in
(Arm Recognition)
[0217] Next, the arm recognition is described. In the present embodiment, the arm recognition is carried out after any of the fingers, the palm, and the thumb is recognized. Note that the arm recognition may also be carried out before any of the fingers, the palm, and the thumb is recognized or at the same time as at least any thereof is recognized.
[0218] In the present embodiment, a polygon is extracted from a region larger than the polygon of the shape of the hand of the image data. For example, the processes of Steps S21 to S27 are carried out in a length range of 5 cm or more and 100 cm or less and, more preferably, a length range of 10 cm or more and 40 cm or less, so that an outer shape is extracted.
[0219] After that, a quadrangular frame circumscribed around the extracted outer shape is selected. In the present embodiment, the shape of the quadrangular frame is a parallelogram or a rectangle.
[0220] In this case, because the parallelogram or the rectangle has longer sides opposed to each other, the extending direction of the arm can be recognized from the extending direction of the longer sides, and the direction of the arm can be determined from the direction of the longer sides. Note that, similarly to the process of Step S32, movement of the arm may be detected in comparison with image data of several frames taken immediately before.
[0221] Note that, although the fingers, the palm, the thumb, and the arm are detected from a two-dimensional image in the above description, not limited thereto, the infrared ray detection unit 410 may be further provided, or only the infrared ray detection camera 412 may be further provided, and a three-dimensional image may be recognized from two-dimensional images. As a result, the recognition accuracy can be further enhanced.
(View Example of Semi-Transmissive Display)
[0222] Next,
[0223] As illustrated in
(Description of Unit Adjustment Mechanism 500)
[0224]
[0225]
[0226] As illustrated in
[0227] In this case, as illustrated in
[0228] Here, in the adjustment of the unit adjustment mechanism 500 below the horizontal direction, the unit adjustment mechanism 500 may be operated downward by a gesture, or the operation may be set in advance according to a used application. The adjustment unit 520 may be manually adjusted to direct the unit adjustment mechanism 500 below the horizontal direction.
[0229] As illustrated in
[0230]
[0231] As illustrated in
[0232] In this case, the user can use the hands H1 to operate a keyboard KB displayed in the virtual image display region 2203D to input characters to the image CAVV. More specifically, the keyboard KB not arranged on the desk STA can be virtually displayed, and the keyboard KB can be operated by the hands H1.
[0233] In this case, the user can sit on the chair and put the hands H1 on the desk STA to input characters to the virtual image display region 2203D. Therefore, long-time operation can be easily performed.
[0234] Although the keyboard KB is virtual in the present embodiment, not limited thereto, the keyboard KB may be actually arranged to cause the infrared ray detection unit 410 to perform the detection.
[0235] Although the desk STA and the chair are used as illustrated in
[0236] Note that, although the user puts the hands H1 on the desk STA to perform the operation in the present embodiment, not limited thereto, a user who cannot raise the hands H1 or the like can also perform the operation at a low position, and this is beneficial.
[0237] Next,
[0238] As illustrated in
[0239] Note that, although the manipulation region 410c is narrowed down in the directions of the arrows SM in the case described in
(Description of Display Adjustment Mechanisms 600)
[0240] Next,
[0241] As illustrated in
[0242] As illustrated in
[0243] Note that the display adjustment mechanisms 600 may move one or a plurality of arrows RSL, RSR, RVL, and RVR according to a predetermined gesture of the user detected by the infrared ray detection unit 410. The display adjustment mechanisms 600 may not be automatic, and an adjustment unit 620 (see
[0244] In this case, the pair of semi-transmissive displays 220 can accurately recognize the pair of semi-transmissive displays 220 even if the user has a squint or the like. Furthermore, although the pair of semi-transmissive displays 220 has been described, it is obvious that each one of the pair of semi-transmissive displays 220 may be able to be adjusted.
[0245] Furthermore, the display adjustment mechanism 600 brings the adjustment angle close to zero with the lapse of the use time, on the basis of an instruction from the control unit 450, and thereby, the glasses display device 100 can provide an effect of treating strabismus.
[0246] In the present embodiment, the case where the user has strabismus has been described. However, the present invention is not limited thereto, and even when the user has hyperopia, astigmatism, amblyopia, or color vision abnormality, restoration adjustment therefor can be performed by displaying an image on the pair of semi-transmissive displays 220 such that the angle which has been adjusted is shifted to zero.
[0247] That is, for example, when the user cannot visually recognize an image displayed on the pair of semi-transmissive displays 220 or the image blurs, focus adjustment can be performed on the image displayed on the pair of semi-transmissive displays 220 through a predetermined gesture.
[0248] As a result, restoration of the eyesight of the user can be adjusted.
[0249] As described above, the respective angles of the pair of semi-transmissive displays 220 can be adjusted by the pair of display adjustment mechanisms 600. Accordingly, even when a user has strabismus, etc., adjustment therefor is not required. In addition, correction of strabismus can be achieved by making the angle adjustment gentler, that is, bringing the angle adjustment close to zero with the lapse of the use time of the glasses display device 100. Moreover, restoration of the eyesight can be adjusted by varying the focal point for the pair of semi-transmissive displays 220.
[0250] In addition, the pair of display adjustment mechanisms 600 may adjust at least either the vertical axes or the perpendicular axes of the pair of semi-transmissive displays 220. As a result, the angles of the vertical axes and the perpendicular axes of the pair of semi-transmissive displays 220 become adjustable. Accordingly, even when the user has strabismus, etc., adjustment therefor can be easily performed.
[0251] Moreover, the pair of display adjustment mechanisms 600 can be manually adjusted through the adjustment unit 620. As a result, the angles and positions of the display surfaces of the pair of semi-transmissive displays 220 can be reliably adjusted with ease.
[0252] The pair of display adjustment mechanisms 600 can perform adjustment depending on determination given by the control unit 450. For example, when the control unit 450 determines that the hand H1, which is an object, has performed a predetermined operation, the pair of display adjustment mechanisms 600 may perform adjustment. As a result, display adjustment can be automatically performed.
[0253] In this case, the pair of display adjustment mechanisms 600 can adjust the plurality of semi-transmissive displays 220 separately. As a result, one of the semi-transmissive displays 220 is viewed by one of the eyes and the other semi-transmissive display 220 is viewed by the other eye, and thereby, adjustment corresponding to each eye can be performed.
[0254] In this case, since the glasses display device 100 is compact and wearable, like glasses, for example, the glasses display device 100 can be easily carried about. Further, since the head-mounted display is compact, and the versatility and convenience thereof can be improved.
[0255] In the present invention, the semi-transmissive display 220 corresponds to the “display device”, the hand H1 corresponds to the “object”, the infrared ray detection unit 410 corresponds to the “depth level sensor”, the control unit 450 corresponds to the “control unit”, the three-dimensional space detection region 4103D corresponds to the “region being measured”, the display adjustment mechanism 600 corresponds to the “display adjustment mechanism”, the adjustment unit 620 corresponds to the “manual adjustment unit”, and the glasses display device 100 corresponds to the “I/O device”.
[0256] A preferred embodiment of the present invention has been described hereinabove, but the present invention is not limited to only the embodiment. It should be understood that various other embodiments are possible without departing from the spirit and scope of the present invention. Further, operations and effects produced by the configuration of the present invention are described in the present embodiment, but these operations and effects are given as examples, and are not intended to limit the present invention.