Patent classifications
G06V40/28
Determining how to assemble a meal
In an embodiment, a method includes determining a given material to manipulate to achieve a goal state. The goal state can be one or more deformable or granular materials in a particular arrangement. The method further includes, for the given material, determining, a respective outcome for each of a plurality of candidate actions to manipulate the given material. The determining can be performed with a physics-based model, in one embodiment. The method further can include determining a given action of the candidate actions, where the outcome of the given action reaching the goal state is within at least one tolerance. The method further includes, based on a selected action of the given actions, generating a first motion plan for the selected action.
Omnichannel intelligent negotiation assistant
An omnichannel intelligent negotiation assistant for generating timely, contextual negotiation assistance to a negotiator. The invention includes a semantic term extractor for converting a contract document into a negotiable term sheet. An omnichannel listener captures all negotiation inputs associated with a negotiation event, sequences each negotiation input by time, and analyzes the sentiment of the negotiation inputs in the context of a term sheet. The resulting annotated negotiation input stream is processed by an intervention generator that includes models of the parties and the negotiation itself as well as a referent negotiation model. The intervention generator includes a game theoretic model that, in concert with a trade-off matrix, allows the intervention generator to produce timely contextual interventions to the negotiator that assist in achieving a superior resulting negotiated agreement.
Speech and Computer Vision-Based Control
The present disclosure relates to a method for controlling a digital photography system. The method includes obtaining, by a device, image data and audio data. The method also includes identifying one or more objects in the image data and obtaining a transcription of the audio data. The method also includes controlling a future operation of the device based at least on the one or more objects identified in the image data, and the transcription of the audio data.
ROBOT CONTROL USING GESTURES
A method and a device for operating a robot are provided. According to an example of the method, information of a first gesture is acquired from a group of gestures of an operator, each gesture from the group of gestures corresponding to an operation instruction from a group of operation instructions. A first operation instruction from the group of operation instructions is obtained based on the acquired information of the first gesture, the first operation corresponding to the first gesture. The first operation instruction is executed.
SMART SPORT DEVICE
An Internet of Thing (IoT) sport device includes a body with a processor, a camera and a wireless transceiver coupled to the processor.
BIOMETRIC IDENTIFICATION BY GARMENTS HAVING A PLURLITY OF SENSORS
Biometric identification methods and apparatuses (including devices and systems) for uniquely identifying one an individual based on wearable garments including a plurality of sensors, including but not limited to sensors having multiple sensing modalities (e.g., movement, respiratory movements, heart rate, ECG, EEG, etc.).
Machine-learned model training for pedestrian attribute and gesture detection
Techniques for detecting attributes and/or gestures associated with pedestrians in an environment are described herein. The techniques may include receiving sensor data associated with a pedestrian in an environment of a vehicle and inputting the sensor data into a machine-learned model that is configured to determine a gesture and/or an attribute of the pedestrian. Based on the input data, an output may be received from the machine-learned model that indicates the gesture and/or the attribute of the pedestrian and the vehicle may be controlled based at least in part on the gesture and/or the attribute of the pedestrian. The techniques may also include training the machine-learned model to detect the attribute and/or the gesture of the pedestrian.
Local physical environment modeling in extended reality environments
Disclosed herein is an environmental scanning tool that generates a digital model representing the surroundings of a user of an extended reality head-mounted display device. The environment is imaged in both a depth map and in visible light for some select objects of interest. The selected objects exist within the digital model at higher fidelity and resolution than the remaining portions of the model in order to manage the storage size of the digital model. In some cases, the objects of interest are selected, or their higher fidelity scans are directed, by a remote user. The digital model further includes time stamped updates of the environment such that users can view a state of the environment according to various timestamps.
SAFETY FOR WEARABLE VIRTUAL REALITY DEVICES VIA OBJECT DETECTION AND TRACKING
The technology disclosed can provide improved safety by detecting potential unsafe conditions (e.g., collisions, loss of situational awareness, etc.) confronting the user of a wearable (or portable) sensor configured to capture motion and/or determining the path of an object based on imaging, acoustic or vibrational waves. Implementations can enable improved safety to users of virtual reality for machine control and/or machine communications applications using wearable (or portable) devices, e.g., head mounted displays (HMDs), wearable goggles, watch computers, smartphones, and so forth, or mobile devices, e.g., autonomous and semi-autonomous robots, factory floor material handling systems, autonomous mass-transit vehicles, automobiles (human or machine driven), and so forth, equipped with suitable sensors and processors employing optical, audio or vibrational detection.
Electronic apparatus for providing a virtual keyboard and controlling method thereof
Disclosed are an electronic apparatus and a method of controlling the same. The electronic apparatus includes a camera, a display, a memory, and a processor configured to execute at least one instruction to: detect a plurality of fingers in a plurality of first image frames obtained through the camera and, in response to an identification that a pose of the plurality of detected fingers corresponds to a trigger pose, enter a character input mode, detect a first motion of a finger among the plurality of fingers in a plurality of second image frames obtained through in the character input mode, identify a key corresponding to the first motion, from among a plurality of keys mapped to the finger, based on a position of the finger by the first motion and a reference point set to the finger, and control the display to display information corresponding to the identified key.