Patent classifications
H04N7/15
Intelligent interactive all-in-one machine
The present invention, belonging to the technical field of data transmission, relates to an integrated intelligent interaction machine. The integrated intelligent interaction machine comprises a signal transmission device, a processing device, a upper computer and a network transmission device, wherein the signal transmission device is connected to the processing device, the processing device is connected to the upper computer, and the upper computer is connected to the processing device through the network transmission device. The signal transmission device is used to receive a screen transmission signal and transmit it to the processing device, and the processing device is used to process the screen transmission signal to obtain screen transmission data and transmit the screen transmission data to the upper computer. The upper computer is used to connect to an external network and output a network signal to the network transmission device, and the network transmission device transmits the network signal to the processing device. The upper computer is used to connect the external network and the internal device of the equipment to provide the network signal for the processing device, thereby realizing the communication between the processing device and the external network. The screen transmission data and the network signal are transmitted through different paths, so as to carry out screen transmission communication in the equipment and network sharing, respectively, thereby improving the data transmission convenience of the integrated intelligent interaction machine.
Virtual 3D communications with actual to virtual cameras optical axes compensation
A method for conducting a three dimensional (3D) video conference between multiple participants, the method may include determining, for each participant, updated 3D participant representation information within the virtual 3D video conference environment, that represents participant; wherein the determining comprises compensating for difference between an actual optical axis of a camera that acquires images of the participant and a desired optical axis of a virtual camera; and generating, for at least one participant, an updated representation of virtual 3D video conference environment, the updated representation of virtual 3D video conference environment represents the updated 3D participant representation information for at least some of the multiple participants.
Preventing audio delay-induced miscommunication in audio/video conferences
Embodiments for delay-induced miscommunication reduction are provided. The embodiment may include capturing data streams transmitted between participants in an A/V exchange; translating, on a sender device prior to transmission to a recipient device, an audio stream within the data streams to text; timestamping, on a sender device prior to transmission to the recipient device, each word in the translated audio stream; transmitting the audio stream and the sender-side translated and timestamped audio stream to the recipient device; translating, on the recipient device, the transmitted audio stream to text; timestamping, on the recipient device, each word in the translated audio stream; determining a lag exists in the A/V exchange based on a comparison of each timestamp for corresponding words on the sender-side translated and timestamped audio stream and the recipient-side translated and timestamped audio stream; and generating a true transcript of an intended exchange between the participants based on the comparison.
Preventing audio delay-induced miscommunication in audio/video conferences
Embodiments for delay-induced miscommunication reduction are provided. The embodiment may include capturing data streams transmitted between participants in an A/V exchange; translating, on a sender device prior to transmission to a recipient device, an audio stream within the data streams to text; timestamping, on a sender device prior to transmission to the recipient device, each word in the translated audio stream; transmitting the audio stream and the sender-side translated and timestamped audio stream to the recipient device; translating, on the recipient device, the transmitted audio stream to text; timestamping, on the recipient device, each word in the translated audio stream; determining a lag exists in the A/V exchange based on a comparison of each timestamp for corresponding words on the sender-side translated and timestamped audio stream and the recipient-side translated and timestamped audio stream; and generating a true transcript of an intended exchange between the participants based on the comparison.
Conference System and Transmitter with Low Standby Power Consumption and no Booting Delay Time
A conference system with low standby power consumption includes a transmitter, an image data source, a receiver, and a display device. The transmitter includes a battery for providing power, at least one link port for accessing data, and a processor coupled to the battery and the at least one link port. The image data source is used for transmitting the image data to the transmitter. The receiver is linked to the transmitter for receiving the image data. The display device is linked to the receiver for displaying the image data. When the transmitter and the image data source are electrically coupled, the processor ceases to use the battery of the transmitter and controls the image data source for providing power to the transmitter. When the transmitter and the image data source are separated, the processor uses the battery of the transmitter for driving firmware of the transmitter.
Conference System and Transmitter with Low Standby Power Consumption and no Booting Delay Time
A conference system with low standby power consumption includes a transmitter, an image data source, a receiver, and a display device. The transmitter includes a battery for providing power, at least one link port for accessing data, and a processor coupled to the battery and the at least one link port. The image data source is used for transmitting the image data to the transmitter. The receiver is linked to the transmitter for receiving the image data. The display device is linked to the receiver for displaying the image data. When the transmitter and the image data source are electrically coupled, the processor ceases to use the battery of the transmitter and controls the image data source for providing power to the transmitter. When the transmitter and the image data source are separated, the processor uses the battery of the transmitter for driving firmware of the transmitter.
Video Conferencing Systems Featuring Multiple Spatial Interaction Modes
Systems and methods for multi-attendee video conferencing are described. A system can convert from huddle video conference mode to spatial video conference mode. In particular, by assigning user roles, specific users can have greater control of the video conference as compared to other users. For instance, moderators may have a greater level of control of the video conferencing system. Thus, in example implementations of the present disclosure, specific users can affect transition between two or more video conferencing modes, such as between a huddle video conference mode and a spatial video conference mode.
Video Conferencing Systems Featuring Multiple Spatial Interaction Modes
Systems and methods for multi-attendee video conferencing are described. A system can convert from huddle video conference mode to spatial video conference mode. In particular, by assigning user roles, specific users can have greater control of the video conference as compared to other users. For instance, moderators may have a greater level of control of the video conferencing system. Thus, in example implementations of the present disclosure, specific users can affect transition between two or more video conferencing modes, such as between a huddle video conference mode and a spatial video conference mode.
Apparatus and method for providing 360-degree panoramic background during video call
Disclosed is a method and apparatus for providing a 360-degree panoramic background during a video call. A non-transitory computer-readable recording medium stores instructions that, when executed by a processor, cause the processor to perform operations for providing a background on an electronic device. The operations include separating an object video and an original background video from video data input to a first video call screen of an electronic device during a video call; generating new video data by synthesizing a panoramic video image and the object video such that the new video data includes the object video and includes the panoramic video image as a replacement background video; and displaying, on the first video call screen, the new video data such that a location of the panoramic video image of the new video data moves in response to motion information of a user.
Systems and methods for immersive scenes
One example system for displaying immersive scenes includes a processor and at least one memory device. The memory device includes instructions that are executable by the processor to cause the processor to receive a collection of metadata associated with an immersive scene, identify each of a plurality of properties of the immersive scene based on the collection of metadata, receive a dynamic immersive background, receive a plurality of video streams associated with a video conference, and display each of the plurality of video streams in the immersive scene based at least in part of the plurality of properties of the immersive scene and on the dynamic immersive background.