G02B27/017

SYSTEM FOR END-TO-END ELECTRONIC DATA ENCRYPTION USING AN INTELLIGENT HOMOMORPHIC ENCRYPED PRIVACY SCREEN

A system is provided for end-to-end electronic data encryption using an intelligent homomorphic encrypted privacy screen. In particular, the system may be configured to use homomorphic encryption on sensitive data displayed on a screen of a computing device. An augmented reality (“AR”) device may be used (e.g., a user) with the privacy screen such that the sensitive data that appears on the screen may be selectively decrypted. In this way, an authorized user may be able to view and interact with the encrypted data whereas an unauthorized user may not.

Virtually modeling clothing based on 3D models of customers

Three-dimensional models (or avatars) may be defined based on imaging data captured from a customer. The avatars may be based on a virtual mannequin having one or more dimensions in common with the customer, a body template corresponding to the customer, or imaging data captured from the customer. The avatars are displayed on displays or in user interfaces and used for any purpose, such as to depict how clothing will appear or behave while being worn by a customer alone or with other clothing. Customers may drag-and-drop images of clothing onto the avatars. One or more of the avatars may be displayed on any display, such as a monitor or a virtual reality headset, which may depict the avatars in a static or dynamic mode. Images of avatars and clothing may be used to generate print catalogs depicting the appearance or behavior of the clothing while worn by the customer.

Method and a display device with pixel repartition optimization

A method for presenting an image on a display device (100) includes modifying the image by applying a geometric transformation to the image so that an area of the image on the display device is presented to a viewer with higher density of pixels than that in the rest of the image (S18).

Methods and apparatuses for providing procedure guidance

Apparatuses and methods of operating the same are described. An apparatus including a display, an input device, and a processing device coupled to the display and the input device. The processing device may send an output to the display. The output may include a graphical object associated with a first step of a user-implemented procedure. The processing device may receive an input from the input device. The input may indicate a progress on an execution of the first step by an operator. The processing device may determine whether the input indicates that the operator has completed the first step. The processing device may determine whether the first step is a final step in the user-implemented procedure. The processing device may identify a second step in the user-implemented procedure when the input indicates that the operator has completed the first step and the first step is not a final step.

Image generation apparatus and image generation method using frequency lower than display frame rate

Methods and apparatus provide for generating an image by way of acquiring information relating to at least one of a position and a rotation of a camera operated by a user. The image is generated for display on a display unit viewed by the user. A rate at which the image is generated is at a first frequency, which is lower than a second frequency corresponding to a frame rate of the display unit.

System and method for capturing a spatial orientation of a wearable device

A system and method capture a spatial orientation of a wearable device. The system has at least one capturing unit and at least one processor unit. The at least one capturing unit is designed to capture at least one first position parameter in relation to the wearable device and to capture at least one second position parameter in relation to a body part of a person on which the wearable device is arranged. The at least one processor unit is designed to determine a spatial orientation of the wearable device on the basis of the at least one first position parameter and the at least one second position parameter.

Surface Puck

An image orientation system is provided wherein images (rays of lights) are projected to a user based on the user's field of view or viewing angle. As the rays of light are projected, streams of air can be produced that bend or focus the rays of light toward the user's field of view. The streams of air can be cold air, hot air, or combinations thereof. Further, an image receiver can be utilized to receive the produced image/rays of light directly in line with the user's field of view. The image receiver can be a wearable device, such as a head mounted display.

HEAD-MOUNTED DISPLAY AND INFORMATION DISPLAY APPARATUS

To provide a head-mounted display and an information display apparatus that match the intuition of the user and are excellent in operability. A head-mounted display 1 according to an embodiment of the present technology includes a reception unit 12, an image display element 14, and a display processing unit 132. The reception unit 12 receives an operation signal including information on a relative position of a detection target in contact with an input operation surface 21, which is output from the input device 2. The image display element 14 forms an image V1 presented to a user. The display processing unit 132 causes, based on the operation signal, the image display element 14 to display an operation image V10 with an auxiliary image P indicating a position of the detection target being overlapped on the image V1.

METHOD AND APPARATUS FOR EGO-CENTRIC 3D HUMAN COMPUTER INTERFACE
20180004392 · 2018-01-04 ·

In the method, a processor generates a three dimensional interface with at least one virtual object, defines a stimulus of the interface, and defines a response to the stimulus. The stimulus is an approach to the virtual object with a finger or other end-effector to within a threshold of the virtual object. When the stimulus is sensed, the response is executed. Stimuli may include touch, click, double click, peg, scale, and swipe gestures. The apparatus includes a processor that generates a three dimensional interface with at least one virtual object, and defines a stimulus for the virtual object and a response to the stimulus. A display outputs the interface and object. A camera or other sensor detects the stimulus, e.g. a gesture with a finger or other end-effector, whereupon the processor executes the response. The apparatus may be part of a head mounted display.

METHOD AND APPARATUS FOR MANIPULATING CONTENT IN AN INTERFACE
20180004299 · 2018-01-04 ·

A machine implemented method includes sensing entities in first and second domains. If a first stimulus is present and an entity is in the first domain, the entity is transferred from first to second domain via a bridge. If a second stimulus is present and an entity is in the second domain, the entity is transferred from second first domain via the bridge. At least some of the first domain is outputted. An apparatus includes a processor that defines first and second domains and a bridge that enables transfer of entities between domains, an entity identifier that identifies entities in the domains, a stimulus identifier that identifies stimuli, and a display that outputs at least some of the first domain. The processor transfers entities from first to second domain responsive to a first stimulus, and transfers entities from second to first domain responsive to a second stimulus.