Information processing apparatus, information processing method, and program
11589001 · 2023-02-21
Assignee
Inventors
Cpc classification
H04N7/147
ELECTRICITY
G06V20/52
PHYSICS
H04N7/142
ELECTRICITY
H04N23/90
ELECTRICITY
International classification
G06T19/00
PHYSICS
Abstract
There is provided an information processing apparatus including a display control section that controls a second display device present in a second space. The second display device displays a first captured image obtained by imaging at least one or more first users present in a first space. The display control section superimposes and displays a second user object on the first captured image in accordance with line-of-sight relevant information of each of the first users. The second user object corresponds to each of at least one or more second users present in the second space. This makes it possible to achieve more realistic communication.
Claims
1. An information processing apparatus comprising a display control section configured to control a second display device present in a second space, the second display device displaying a first captured image obtained by imaging at least one or more first users present in a first space, and superimpose and display a second user object on the first captured image in accordance with line-of-sight relevant information of each of the first users, the second user object corresponding to each of at least one or more second users present in the second space, wherein the line-of-sight relevant information includes information regarding a direction from the first user present in the first space to a first display device visually recognized by the first user, the second user object is superimposed and disposed ahead of the line of sight in the direction from the first user to the first display device visually recognized by the first user, and the display control section is implemented via at least one processor.
2. The information processing apparatus according to claim 1, wherein the line-of-sight relevant information includes information regarding a visual distance from the first user present in the first space to the first display device visually recognized by the first user, and the display control section is further configured to control a display position of the second user object on a basis of the information regarding the visual distance, the second user object corresponding to the second user.
3. The information processing apparatus according to claim 2, wherein the display control section is further configured to control a size of the first captured image on the basis of the information regarding the visual distance.
4. The information processing apparatus according to claim 3, further comprising a 3D conversion section configured to generate a virtual 3D space obtained by virtually expressing, in a same space, a positional relationship between the first user, the first display device, the second display device, and the second user, wherein the display control section is further configured to control the display position of the second user object and the size of the first captured image on a basis of a positional relationship between the first user disposed in the virtual 3D space and the first display device disposed in the virtual 3D space and a virtual captured image, the first display device being disposed to be spaced apart by the visual distance from the first user, the virtual captured image being virtually captured from a display starting point disposed at a position on a same side as a side of the first display device in a case where the first user is used as a starting point, the position being farther than a position of the first display device, and the 3D conversion section is implemented via at least one processor.
5. The information processing apparatus according to claim 4, wherein the second display device present in the second space is disposed at a same position as a position of the first user in association with the first user in the virtual 3D space, and the display control section is further configured to control the display position of the second user object and the size of the first captured image by using a positional relationship between the second display device and the first display device in the virtual 3D space.
6. The information processing apparatus according to claim 5, wherein a size of the first captured image is determined on a basis of the visual distance and a field angle of a first imaging device, the first captured image being displayed on the second display device, the first imaging device capturing the first captured image.
7. The information processing apparatus according to claim 1, wherein the display control section is further configured to change a size of the second user object to prevent the first user and the second user object from being superimposed, the first user being displayed in the first captured image.
8. The information processing apparatus according to claim 2, wherein the second user object corresponding to the second user is generated on a basis of an image of the second user, the image being extracted from a second user back view captured image obtained by imaging a back view of the second user.
9. The information processing apparatus according to claim 6, wherein the information processing apparatus includes the first imaging device and the first display device that are integrated with each other.
10. The information processing apparatus according to claim 2, wherein the second user object is created by using information acquired from infrared light, the second user being irradiated with the infrared light, the infrared light being reflected from the second user.
11. The information processing apparatus according to claim 1, wherein the display control section is further configured to control the first display device present in the first space and superimpose and display a first user object on a second captured image in accordance with line-of-sight relevant information of each of at least one or more second users present in the second space, the first display device displaying the second captured image obtained by imaging the second user, the first user object corresponding to each of the at least one or more first users present in the first space.
12. An information processing method comprising: controlling a second display device present in a second space, the second display device displaying a first captured image obtained by imaging at least one or more first users present in a first space; and superimposing and displaying a second user object on the first captured image in accordance with line-of-sight relevant information of each of the first users, the second user object corresponding to each of at least one or more second users present in the second space, wherein the line-of-sight relevant information includes information regarding a direction from the first user present in the first space to a first display device visually recognized by the first user, and the second user object is superimposed and disposed ahead of the line of sight in the direction from the first user to the first display device visually recognized by the first user.
13. A non-transitory computer-readable medium having embodied thereon a program, which when executed by a computer causes the computer to execute an information processing method, the method comprising: controlling a second display device present in a second space, the second display device displaying a first captured image obtained by imaging at least one or more first users present in a first space; and superimposing and displaying a second user object on the first captured image in accordance with line-of-sight relevant information of each of the first users, the second user object corresponding to each of at least one or more second users present in the second space, wherein the line-of-sight relevant information includes information regarding a direction from the first user present in the first space to a first display device visually recognized by the first user, and the second user object is superimposed and disposed ahead of the line of sight in the direction from the first user to the first display device visually recognized by the first user.
Description
BRIEF DESCRIPTION OF DRAWING
(1)
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12)
(13)
(14)
(15)
(16)
(17)
(18)
MODES FOR CARRYING OUT THE INVENTION
(19) The following describes a preferred embodiment of the present disclosure in detail with reference to the accompanying drawings. It is to be noted that, in this description and the drawings, components that have substantially the same functional component are indicated by the same signs and redundant description thereof is thus omitted.
(20) It is to be noted that description is given in the following order.
(21) 1. Regarding Technology of the Present Disclosure
(22) 1.1. Overview of Technology
(23) 1.2. Disposition of Back View
(24) 1.3. Effects
(25) 2. Configuration
(26) 2.1. External Configuration
(27) 2.2. Internal Configuration
(28) 3. Operation Flow
(29) 4. Modification Examples
(30) 4.1. Priority of Visual Distance
(31) 4.2. Use of Distance Measurement Sensor
(32) 4.3. Use of Eyewear Display Device
(33) 5. Hardware Configuration Example
1. Regarding Technology of the Present Disclosure
(34) (1.1. Overview of Technology)
(35) With reference to
(36) In
(37) (1.2. Disposition of Back View)
(38) With reference to
(39) The user A1 present at the A point visually recognizes the display device D.sub.1 present at the A point and communicates with the user B1 at the B point. The camera VC.sub.1 for a chat images A1 within the range of a field angle VCθ.sub.1. The camera VC.sub.1 for a chat is provided to the display device D.sub.1 present at the A point.
(40) Similarly to the user A1 present at the A point, the user B1 present at the B point visually recognizes the display device D.sub.2 present at the B point and communicates with the user A1 at the A point. The camera VC.sub.2 for a chat images B1 within the range of a field angle VCθ.sub.2. The camera VC.sub.2 for a chat is provided to the display device D.sub.2 present at the B point.
(41) The right portion of
(42) In a captured image illustrated in the uppermost row, the user A1 visually recognizes the direction of a line G1.sub.A1 of sight in the right end region of the display device D.sub.1 from the user A1 to the display device D.sub.1. The captured image illustrated in the uppermost row is obtained by imaging the user A1 present at the A point. A back view S.sub.B1 of the user B1 present at the B point is superimposed and disposed ahead of the line G1.sub.A1 of sight in the captured image in the uppermost row.
(43) In a captured image illustrated in the middle row, the user A1 visually recognizes the direction of a line G2.sub.A1 of sight in the middle end region of the display device D.sub.1 from the user A1 to the display device D.sub.1. The captured image illustrated in the middle row is obtained by imaging the user A1 present at the A point. The back view S.sub.B1 of the user B1 present at the B point is superimposed and disposed ahead of the line G2.sub.A1 of sight in the captured image in the middle row.
(44) In a captured image illustrated in the lowest row, the user A1 visually recognizes the direction of a line G3.sub.A1 of sight in the left end region of the display device D.sub.1 from the user A1 to the display device D.sub.1. The captured image illustrated in the lowest row is obtained by imaging the user A1 present at the A point. The back view S.sub.B1 of the user B1 present at the B point is superimposed and disposed ahead of the line G3.sub.A1 of sight in the captured image in the uppermost row.
(45) The same applies to a captured image obtained by performing imaging at the B point. In a captured image illustrated in the uppermost row, the user B1 visually recognizes the direction of a line G1.sub.B1 of sight in the right end region of the display device D.sub.2 from the user B1 to the display device D.sub.2. The captured image illustrated in the uppermost row is obtained by imaging the user B1 present at the B point. A back view S.sub.A1 of the user A1 present at the A point is superimposed and disposed ahead of the line G1.sub.B1 of sight in the captured image in the uppermost row.
(46) In a captured image illustrated in the middle row, the user B1 visually recognizes the direction of a line G2.sub.B1 of sight in the middle end region of the display device D.sub.2 from the user B1 to the display device D.sub.2. The captured image illustrated in the middle row is obtained by imaging the user B1 present at the B point. The back view S.sub.A1 of the user A1 present at the A point is superimposed and disposed ahead of the line G2.sub.B1 of sight in the captured image in the middle row.
(47) In a captured image illustrated in the lowest row, the user B1 visually recognizes the direction of a line G3.sub.B1 of sight in the left end region of the display device D.sub.2 from the user B1 to the display device D.sub.2. The captured image illustrated in the lowest row is obtained by imaging the user B1 present at the B point. The back view S.sub.A1 of the user A1 present at the A point is superimposed and disposed ahead of the line G3.sub.B1 of sight in the captured image in the lowest row.
(48) In this way, a user displayed in a captured image at the A point has a back view of a user at the B point superimposed and displayed ahead of the line of sight in the captured image. This allows the user at the B point to simulate an experience in which the user at the B point communicated in eye contact with the user at the A point who is displayed in the captured image.
(49) (1.3. Effects)
(50) An object and effects of the technology of the present disclosure are specifically described with reference to
(51) In contrast,
(52) The application of the technology of the present disclosure makes it possible to superimpose and dispose a back view of any of the user 5, the user 6, and the user 8 to be visually recognized ahead of the line of sight of the user 7 on the captured image. This allows the user 5, the user 6, and the user 8 to determine which of the user 5, the user 6, and the user 8 is visually recognized by the user 7 and experience a more realistic meeting.
(53) (2. Configuration)
(54) (2.1. External Configuration)
(55) With reference to
(56) According to
(57) At the point A, the display device D.sub.A displays a captured image that is captured by the camera VC.sub.B for a chat at the point B. The user A1 is able to visually recognize the user B1 and the user B2 for communication via the display device D.sub.A. At the point B, the display device D.sub.B displays a captured image that is captured by the camera VC.sub.A for a chat at the point A. The user B1 and the user B2 are able to visually recognize the user A1 for communication via the display device D.sub.B.
(58) The display device D.sub.A superimposes and displays the back view S.sub.A1 of the user A1 on the captured image that is captured by the camera VC.sub.B for a chat at the point B. The back view S.sub.A1 is created from the captured image that is captured by the camera SC.sub.A for a back view. Meanwhile, the display device D.sub.B superimposes and displays the back view S.sub.B1 of the user B1 and a back view S.sub.B2 of the user B2 on the captured image that is captured by the camera VC.sub.A for a chat at the point A. The back view S.sub.B1 and the back view S.sub.B2 are created from the captured images that are captured by the camera SC.sub.B for a back view.
(59) The user A1 is visually recognizing the user B2 displayed on the display device D.sub.A in the direction of a line A1v of sight. The display device D.sub.B displays the back view S.sub.B2 of the user B2 in the direction of the line A1v of sight of the user A1. This allows B1 and B2 present at the point B to recognize that the user A1 is visually recognizing the user B2 for communication.
(60) (2.2. Internal Configuration)
(61) According to
(62) The chat cameras VC.sub.A and VC.sub.B have functions of performing imaging at the respective points. Each of the chat cameras VC.sub.A and VC.sub.B is an imaging device such as an RGB camera or a thermography camera that captures an image (still image or moving image). Each of the chat cameras VC.sub.A and VC.sub.B may include a lens system, a driving system, and an imaging element. Captured images obtained by the chat cameras SC.sub.A and SC.sub.B may be used for image processing or may be used to be displayed on the display devices D.sub.A and D.sub.B.
(63) The back view cameras SC.sub.A and SC.sub.B have functions of imaging back views of users at the respective points. The back view cameras SC.sub.A and SC.sub.B may be imaging devices similar to the chat cameras VC.sub.A and VC.sub.B.
(64) The display devices D.sub.A and D.sub.B have functions of displaying captured images at the respective points and displaying back views of users in accordance with an instruction of the control unit 100. The captured images are captured by the chat cameras VC.sub.A and VC.sub.B in different locations. The back views are imaged by the back view cameras SC.sub.A and SC.sub.B present in the locations of the display devices D.sub.A and D.sub.B. Specifically, each of the display devices D.sub.A and D.sub.B includes a device that is able to visually notify a user of a display image generated by the control unit 100. Example of such a device include display devices such as CRT (Cathode Ray Tube) display device, a liquid crystal display device, a plasma display device, and EL (electroluminescence) display device, a laser projector, an LED projector, and a lamp.
(65) The control unit 100 has functions of acquiring various kinds of data from a variety of devices provided at the A point and the B point and controlling display devices. The control unit includes an acquisition section 110, a person recognition section 120, a 3D conversion section 130, and a display control section 140.
(66) The acquisition section 110 acquires captured images from the back view cameras SC.sub.A and SC.sub.B. In addition, the acquisition section 110 acquires captured images from the chat cameras VC.sub.A and VC.sub.B. The acquisition section 110 acquires a captured image, for example, through communication. Communication may be constantly performed or may be discontinuously performed as appropriate. The acquisition section 110 outputs an acquired captured image to the person recognition section 120.
(67) The person recognition section 120 has a function of identifying the positions of users by processing captured images acquired from the chat cameras VC.sub.A and VC.sub.B. Description is made with reference to
(68) The person recognition section 120 recognizes the position in a three-dimensional space at which the user A1 is present within the horizontal imaging field angle θ.sub.A and the vertical direction imaging field angle φ.sub.A from a captured image that is captured by the chat camera VC.sub.A. The person recognition section 120 may recognize the position of the user A1 with respect to the display device D.sub.A.
(69) Further, the person recognition section 120 has a function of acquiring line-of-sight relevant information of a user from a captured image that is captured by the chat camera VC.sub.A. The line-of-sight relevant information of a user includes pieces of information regarding the user's line of sight and visual distance. A user's line of sight includes the line-of-sight direction and the line-of-sight position of the user with respect to the display device D.sub.A. The visual distance of a user may be the horizontal distance from the eyes of the user to the display device D.sub.A. In a case where description is given with reference to the left portion of
(70) The person recognition section 120 has a function of detecting users from captured images that are captured by the chat cameras VC.sub.A and VC.sub.B. In a case where description is given with reference to
(71) The person recognition section 120 has a function of detecting back views of users from captured images that are captured by the back view cameras SC.sub.A and SC.sub.B in a similar way to detect users from captured images that are captured by the chat cameras VC.sub.A and VC.sub.B. The person recognition section 120 may recognize that the user A1 imaged by the chat camera SC.sub.A at the point A and the user A1 imaged by the back view camera SC.sub.A at the point A refer to the same user and associate the user A1 imaged by the chat camera SC.sub.A at the point A and the user A1 imaged by the back view camera SC.sub.A at the point A. Such association makes it possible to dispose a back view of the user A1 at the position of the user A1 imaged by the chat camera VC.sub.A at the point A in a virtual 3D space that is described below.
(72) The 3D conversion section 130 has functions of generating a virtual captured image by using various kinds of information acquired from the person recognition section 120 and outputting the generated virtual captured image to the display control section 140. The virtual captured image is for controlling a display image that is displayed by a display device at the point B. In a case where description is given with reference to
(73) With reference to
(74) A virtual second display device D.sub.Q is disposed at the position opposed to the virtual first display device D.sub.A. The disposition position of the virtual second display device D.sub.Q is determined on the basis of the positional relationship between the display device D.sub.A and the user A1 at the point A. Specifically, the virtual second display device D.sub.Q is disposed to be spaced apart from the virtual first display device D.sub.A with the same distance in between as the visual distance d.sub.A1 between the display device D.sub.A and the user A1 at the point A.
(75) The virtual first display device D.sub.A displays the users B1 and B2 imaged by the chat camera VC.sub.B as at the point A and the virtual second display device D.sub.Q displays the user A1 imaged by the chat camera VC.sub.A as at the point B as at the point A.
(76) The vertical and horizontal size of the virtual second display device D.sub.Q in a virtual 3D space is determined on the basis of the visual distance and the field angle of the first imaging device that captures the first captured image. In other words, a horizontal size H.sub.Q of the virtual second display device D.sub.Q is determined on the basis of the visual distance d.sub.A1 and the horizontal imaging field angle θ.sub.A of the virtual first display device D.sub.A. It is possible to express the horizontal size H.sub.Q of the virtual second display device D.sub.Q as H.sub.Q=2d.sub.A1 tan(θ.sub.A/2). In contrast, a vertical size V.sub.Q of the virtual second display device D.sub.Q is determined on the basis of the visual distance d.sub.A1 and the vertical imaging field angle φ.sub.A of the virtual first display device D.sub.A. It is possible to express the vertical size V.sub.Q of the virtual second display device D.sub.Q as V.sub.Q=2d.sub.A1 tan(φ.sub.A/2).
(77) The 3D conversion section 130 converts an image by using the disposition of the respective virtual display devices illustrated in
(78) The 3D conversion section 130 further disposes a display starting point in the virtual 3D space described above and generates an image by using a virtual captured image that is captured by a virtual imaging device (not illustrated) from the display starting point. The image is displayed on the display device D.sub.B at the point B.
(79) The display control section 140 controls an image by using a virtual captured image generated by the 3D conversion section 130. The image is displayed on the display device D.sub.B at the point B in a real space.
(80) With reference to
(81) In a case where a back view of the user B1 or B2 is displayed overlapping with the user A1 in the virtual captured image Pv as illustrated in the upper portion of
(82) In addition, in a case where a back view of the user B1 or B2 gets out of the frame in the virtual captured image Pv as illustrated in
(83) Such display control makes it possible to display a display image without causing any strangeness. In addition, it is also possible to achieve display control through a conversion process with reduced processing loads.
(84) Further, in a case where the user B1 or B2 gets out of the frame in the virtual captured image Pv, the display control section 140 may also change the display position of a back view of the user B1 or B2 to cause the virtual captured image Pv to include the back view of the user B1 or B2 who gets out of the frame. The display position may be changed within a range within in which the line of sight of the user A1 and the position of the back view correspond to each other. It is to be noted that the display control section 140 changes a position at which a back view of a user is disposed in a captured image visually recognized by a user at the B point in the present embodiment, but this example is not limitative. The display control section 140 may determine in advance a position at which a back view of a user at the B point is disposed, process the line of sight of a user at the A point at the position at which the back view of the user at the B point is disposed, and process the lines of sight as if the line of sight of the user at the A point matched with the line of sight of the back view of the user at the B point. The respective back views of the users at the B point may be then provided with tags and processed to allow the respective users to be identified. The respective back views of the users at the B point are disposed ahead of the line of sight of the user at the A point who is visually recognized at the B point.
(85) In addition, in a case where a back view of the user B1 or B2 is displayed overlapping with the user A1 in the virtual captured image Pv as illustrated in the upper portion of
(86) (3. Operation Flow)
(87) The operation flow (S100) described so far and performed by the respective devices and the respective components is described with reference to
(88) First, at the point A, the chat camera VC.sub.A shoots an image at the point A (S102). In addition, at the point A, the back view camera SC.sub.A shoots an image of a back view of the user A1 (S104). In contrast, at the point B, the chat camera VC.sub.B shoots an image at the point B (S106). In addition, at the point B, the back view camera SC.sub.B shoots images of back views of the users B1 and B2 (S108).
(89) These operations may concurrently start or sequentially start. The operations may be performed in any order.
(90) Next, the acquisition section acquires captured images from the chat cameras VC.sub.A and VC.sub.B and the back view cameras SC.sub.A and SC.sub.B (S110).
(91) Next, the person recognition section 120 recognizes users from the captured images and detects various kinds of information (S111).
(92) The person recognition section 120 detects the visual distance of the user A1 (S112). The person recognition section 120 then detects and identifies the user A1 from the captured images and detects the visual distance between the chat camera VC.sub.A and the user A1.
(93) The person recognition section 120 identifies the positions of the users in the captured images from the captured images (S113). The person recognition section 120 then detects and identifies the users B1 and B2 from the captured images and identifies the positions in the captured images. It is to be noted that the positions in the captured images may be positions in a real space at the point B.
(94) The person recognition section 120 extracts back views from the captured images of the back view cameras (S114). The person recognition section 120 then detects and identifies the portions of the persons of the users B1 and B2 from the captured image. The person recognition section 120 separates the back views of the users B1 and B2 in the captured images from the background and extracts the back views of the users B1 and B2.
(95) Next, the 3D conversion section disposes the virtual first display device and the virtual second display device in a virtual 3D space on the basis of the positional relationship between a display device and the user A1 at the A point (S116).
(96) Next, the display control section controls an image that is displayed by a display device B at the point B on the basis of the positional relationship between the virtual first display device and the virtual second display device that have been disposed and a virtual captured image that is captured from the display starting point (S120).
(97) The respective components perform the operation flow described above, thereby making it possible to display a display image with no strangeness caused without correcting the face of the user A1, performing image processing on the face of the user A1, or the like. In addition, it is also possible to achieve display control through a 3D conversion process with reduced processing loads.
4. Modification Examples
(98) The information processing apparatus and the information processing method according to the present embodiment have been described so far. In this section, with reference to
(99) (4.1. Priority of Visual Distance)
(100) In the embodiment described above, the example has been described in which the user A1 is present at the point A, the users B1 and B2 are present at the point B. and communication is performed between the point A and the point B. Here, with reference to
(101) The point A illustrated in
(102) In this way, in a case where a plurality of users is present at the point A, the control unit 100 may preferentially perform display control on the visual distance of the user having the smallest visual distance. The direction of a line of sight varies in a wider range as a chat camera is closer. It is therefore possible to suppress the variation range by prioritizing the visual distance of the user having the smallest visual distance as described. In addition, the control unit 100 may identify a speaker from a plurality of users at the point A and preferentially perform display control on the visual distance of the identified speaker.
(103) (4.2. Use of Distance Measurement Sensor)
(104) As an example of another modification example of the embodiment described above, a technology is described with reference to
(105)
(106) The control unit 100 may acquire information regarding a motion of a user by using the distance measurement sensor devices Kn.sub.A and Kn.sub.B and create an avatar or a user object obtained by partially processing the avatar on the basis of the information. In a case where there is a plurality of users, the control unit 100 may associate users identified by the chat cameras VC.sub.A and VC.sub.B and users for whom pieces of information regarding motions of users are acquired by the distance measurement sensor devices Kn.sub.A and Kn.sub.B. The control unit 100 may acquires motions or the like of the skeletal structures of the users in real time and reflect the motions or the like in user objects. In a case where there is a plurality of users, the display control section 140 may reflect the visual characteristics of the respective users in user objects. The user objects correspond to the respective users. This causes the respective user objects to have differences and makes it possible to identify the user objects. The visual characteristics may be, for example, the texture of clothing, hairstyles, or the like.
(107) It is to be noted that the example has been described in the present modification example in which a distance measurement sensor device is used in place of a back view camera A chat camera and a distance measurement sensor device may be, however, the same device. According to this, it is possible to measure a distance by performing image recognition and image processing on a captured image that is captured by a chat camera and suppress the number of installed devices.
(108) (4.3. Use of Eyewear Display Device)
(109) As an example of another modification example of the embodiment described above, a technology is described with reference to
5. Hardware Configuration Example
(110) With reference to
(111) As illustrated in
(112) The CPU 901 functions as an arithmetic processing device and a control device. The CPU 901 controls the overall operation in the information processing apparatus 900 in accordance with a variety of programs. In addition, the CPU 901 may be a microprocessor. The ROM 902 stores a program, an arithmetic parameter, and the like to be used by the CPU 901. The RAM 903 temporarily stores a program used in execution of the CPU 901, a parameter appropriately changed in the execution, and the like. The CPU 901 may form the control unit illustrated in
(113) The CPU 901, the ROM 902, and the RAM 903 are coupled to each other by the host bus 904a including a CPU bus and the like. The host bus 904a is coupled to the external bus 904b such as a PCI (Peripheral Component Interconnect/Interface) bus via the bridge 904. It is to be noted that the host bus 904a, the bridge 904, and the external bus 904b do not necessarily have to be separately included, but the functions thereof may be implemented in one bus.
(114) The input device 906 is achieved by a device through which a user inputs information, such as a mouse, a keyboard, a touch panel, a button, a microphone, a switch, and a lever, for example. In addition, the input device 906 may be, for example, a remote control device using infrared rays or other electric waves. Alternatively, the input device 906 may be an external coupling device such as a mobile phone or PDA supporting an operation of the information processing apparatus 900. Further, the input device 906 may include, for example, an input control circuit or the like that generates an input signal on the basis of information inputted by the user using the input means described above and outputs the generated input signal to the CPU 901. The user of the information processing apparatus 900 is able to input various kinds of data to the information processing apparatus 900 or instruct the information processing apparatus 900 to perform a processing operation by operating this input device 906.
(115) The display device 907 is formed by using a device that is able to visually or aurally notify the user of acquired information. Examples of such a device include display devices such as a CRT display device, a liquid crystal display device, a plasma display device, an EL display device, a laser projector, an LED projector, and a lamp, and sound output devices such as a speaker or a headphone. The display device 907 outputs, for example, results obtained through a variety of processes performed by the information processing apparatus 900. Specifically, the display device 907 visually displays the results obtained through a variety of processes performed by the information processing apparatus 900 in a variety of formats such as text, images, tables, and graphs. Meanwhile, in a case where the sound output device is used, the sound output device converts audio signals including reproduced sound data, acoustic data, or the like into analog signals and aurally outputs the analog signals. The display device 907 is, for example, the display device D.sub.A or the display device D.sub.B illustrated in
(116) The storage device 908 is a device for data storage that is formed as an example of a storage unit of the information processing apparatus 900. For example, the storage device 908 is achieved by a magnetic storage unit device such as HDD, a semiconductor storage device, an optical storage device, a magneto-optical storage device, or the like. The storage device 908 may include a storage medium, a recording device that records data in the storage medium, a reading device that reads data from the storage medium, a deletion device that deletes data recorded in the storage medium, and the like. This storage device 908 stores a program to be executed by the CPU 901, various kinds of data, various kinds of data acquired from the outside, and the like. The storage device 908 stores, for example, a virtual captured image, a variety of parameters, or the like. The virtual captured image is used in a case where the display control section 140 controls the display device D.sub.A and the display device D.sub.B in the control unit 100 illustrated in
(117) The drive 909 is a reader/writer for a storage medium. The drive 909 is built in or externally attached to the information processing apparatus 900. The drive 909 reads out information recorded in a removable storage medium such as a magnetic disk, an optical disc, a magneto-optical disk, or a semiconductor memory that is mounted thereon and outputs the information to the RAM 903. In addition, the drive 909 is also able to write information into the removable storage medium.
(118) The coupling port 911 is an interface coupled to an external apparatus. The coupling port 911 is a coupling port to an external apparatus that is able to transmit data through USB (Universal Serial Bus) and the like, for example.
(119) The communication device 913 is a communication interface including, for example, a communication device and the like for coupling to a network 920. The communication device 913 is, for example, a communication card or the like for wired or wireless LAN (Local Area Network). LTE (Long Term Evolution), Bluetooth (registered trademark), or WUSB (Wireless USB). In addition, the communication device 913 may be a router for optical communication, a router for ADSL (Asymmetric Digital Subscriber Line), a modem for various kinds of communication, or the like. For example, this communication device 913 is able to transmit and receive signals and the like to and from the Internet and another communication apparatus in accordance with a predetermined protocol such as TCP/IP, for example. The communication device 913 is used for the transmission and reception of various kinds of information, for example, between the control unit 100 illustrated in
(120) It is to be noted that the network 920 is a wired or wireless transmission path for information transmitted from an apparatus coupled to the network 920. For example, the network 920 may include a public network such as the Internet, a telephone network, or a satellite communication network, and various LANs (Local Area Networks) including Ethernet (registered trademark), WAN (Wide Area Network), and the like. In addition, the network 920 may include a private network such as IP-VPN (Internet Protocol-Virtual Private Network).
(121) In addition, in the information processing apparatus 900, it is possible to create a computer program for causing the hardware such as CPU. ROM, and RAM built in the information processing system 1000 to exhibit functions equivalent to those of the respective components of the information processing system 1000 according to the present embodiment described above. In addition, a recording medium having the computer program stored thereon may also be included in the scope of the technology according to the present disclosure.
(122) The preferred embodiment of the present disclosure has/have been described above in detail with reference to the accompanying drawings, but the technical scope of the present disclosure is not limited to the embodiment. A person skilled in the art may find various alterations and modifications within the scope of the appended claims, and it should be understood that they will naturally come under the technical scope of the present disclosure.
(123) In addition, the effects described herein are merely illustrative and exemplary, but not limitative. In other words, the technology according to the present disclosure may exert other effects that are apparent to those skilled in the art from the description herein in addition to the above-described effects or in place of the above-described effects.
(124) It is to be noted that the following configurations also fall within the technical scope of the present disclosure.
(125) (1)
(126) An information processing apparatus including
(127) a display control section that controls a second display device present in a second space, the second display device displaying a first captured image obtained by imaging at least one or more first users present in a first space, in which
(128) the display control section superimposes and displays a second user object on the first captured image in accordance with line-of-sight relevant information of each of the first users, the second user object corresponding to each of at least one or more second users present in the second space.
(129) (2)
(130) The information processing apparatus according to (1), in which
(131) the line-of-sight relevant information includes information regarding a visual distance from the first user present in the first space to a first display device visually recognized by the first user, and
(132) the display control section controls a display position of the second user object on the basis of the information regarding the visual distance, the second user object corresponding to the second user.
(133) (3)
(134) The information processing apparatus according to (2), in which the display control section controls a size of the first captured image on the basis of the information regarding the visual distance.
(135) (4)
(136) The information processing apparatus according to (3), further including a 3D conversion section that generates a virtual 3D space obtained by virtually expressing, in a same space, a positional relationship between the first user, the first display device, the second display device, and the second user, in which
(137) the display control section controls the display position of the second user object and the size of the first captured image on the basis of a positional relationship between the first user disposed in the virtual 3D space and the first display device disposed in the virtual 3D space and a virtual captured image, the first display device being disposed to be spaced apart by the visual distance from the first user, the virtual captured image being virtually captured from a display starting point disposed at a position on a same side as a side of the first display device in a case where the first user is used as a starting point, the position being farther than a position of the first display device.
(138) (5)
(139) The information processing apparatus according to (4), in which
(140) the second display device present in the second space is disposed at a same position as a position of the first user in association with the first user in the virtual 3D space, and
(141) the display control section controls the display position of the second user object and the size of the first captured image by using a positional relationship between the second display device and the first display device in the virtual 3D space.
(142) (6)
(143) The information processing apparatus according to any one of (2) to (5), in which a size of the first captured image is determined on the basis of the visual distance and a field angle of a first imaging device, the first captured image being displayed on the second display device, the first imaging device capturing the first captured image.
(144) (7)
(145) The information processing apparatus according to any one of (1) to (6), in which the display control section changes a size of the second user object to prevent the first user and the second user object from being superimposed, the first user being displayed in the first captured image.
(146) (8)
(147) The information processing apparatus according to any one of (2) to (7), in which the second user object corresponding to the second user is generated on the basis of an image of the second user, the image being extracted from a second user back view captured image obtained by imaging a back view of the second user.
(148) (9)
(149) The information processing apparatus according to (6), in which the information processing apparatus includes the first imaging device and the first display device that are integrated with each other.
(150) (10)
(151) The information processing apparatus according to any one of (2) to (9), in which the second user object is created by using information acquired from infrared light, the second user being irradiated with the infrared light, the infrared light being reflected from the second user.
(152) (11)
(153) The information processing apparatus according to any one of (1) to (10), in which the display control section further controls a first display device present in a first space and superimposes and displays a first user object on a second captured image in accordance with line-of-sight relevant information of each of at least one or more second users present in the second space, the first display device displaying the second captured image obtained by imaging the second user, the first user object corresponding to each of the at least one or more first users present in the first space.
(154) (12)
(155) An information processing method including:
(156) controlling a second display device present in a second space, the second display device displaying a first captured image obtained by imaging at least one or more first users present in a first space; and
(157) superimposing and displaying a second user object on the first captured image in accordance with line-of-sight relevant information of each of the first users, the second user object corresponding to each of at least one or more second users present in the second space.
(158) (13)
(159) A program for causing a computer to function to include
(160) a display control section that controls a second display device present in a second space, the second display device displaying a first captured image obtained by imaging at least one or more first users present in a first space, in which
(161) the display control section superimposes and displays a second user object on the first captured image in accordance with line-of-sight relevant information of each of the first users, the second user object corresponding to each of at least one or more second users present in the second space.
REFERENCE SIGNS LIST
(162) D.sub.A, D.sub.B display device VC.sub.A, VC.sub.B camera for a chat SC.sub.A, SC.sub.B camera for a back view 100 control unit 110 acquisition section 120 person recognition section 130 conversion section 140 display control section