Patent classifications
H04S2420/01
ACOUSTIC REPRODUCTION METHOD, RECORDING MEDIUM, AND ACOUSTIC REPRODUCTION SYSTEM
An acoustic reproduction method is an acoustic reproduction method for causing a user to perceive a first sound as a sound arriving from a first position in a three-dimensional sound field and a second sound as a sound arriving from a second position different from the first position in the three-dimensional sound field. The acoustic reproduction method includes: obtaining a movement speed of a head of the user; and generating an output sound signal for causing the user to perceive sounds that arrive from predetermined positions in the three-dimensional sound field. In the generating, when the movement speed obtained is greater than a first threshold, the output sound signal for causing the user to perceive the first sound and the second sound as a sound arriving from a third position between the first position and the second position is generated.
COMPUTER SYSTEM FOR PROCESSING AUDIO CONTENT AND METHOD THEREOF
A computer system for processing audio content may receive content that includes metadata on spatial features about a plurality of objects, convert a format set according to a production environment of the content to a format according to a playback environment in an electronic apparatus, and transmit the content in the converted format to the electronic apparatus. The computer system may support content produced in various production environments and various playback environments.
Method and device for processing audio signal, using metadata
Disclosed is a device for processing an audio signal, which renders an audio signal. The device for processing an audio signal includes a processor. The processor receives metadata including an audio signal and first element reference distance information and renders a first element signal on the basis of the first element reference distance information, wherein the first element reference distance information indicates the reference distance of an element signal. The audio signal is capable of including a second element signal which may be simultaneously rendered with the first element signal, and the metadata is capable of including second element distance information indicating the distance of the second element signal. The number of bits required for representing the first element reference distance information is smaller than the number of bits required for representing the second element distance information.
System and method for an audio reproduction device
System and method for enhancing audio reproduced by an audio reproduction device is described. A plurality of convolution coefficients are generated for a predefined space. A digital audio signal is modified based on the generated convolved digital audio signal to generate a convolved digital audio signal. The convolved digital audio signal is converted to a convolved analog audio signal. The convolved analog audio signal is fed to the audio reproduction device.
HEAD TRACKING SYSTEM
A head tracking system for use in a headrest, includes at least one optical circuit having a plurality of sensors. Each sensor arranged to detect light reflected from at least portion of an illuminated scene. The portion of the scene contains at least a part of a head to be tracked, and to provides an output signal corresponding to an amount of reflected light detected by the sensor. An acquisition circuit is arranged to receive a plurality of output signals from the plurality of sensors and to derive a frame from said output signals. A processing unit is arranged to receive the frame and to derive an indication of a change in position and/or orientation of the head by comparing the received frame with one or more frames obtained at different moments in time.
Head-mounted display apparatus, sound image output system, and method of outputting sound image
An HMD includes a display unit mounted on a head of a user, and configured to display an image such that a real object located in a real space is visually recognizable, a right earphone and a left earphone configured to output a sound, a position specification unit configured to specify a position of the real object and a virtual object, and a sound output control unit configured to generate a synthesized sound with the position of the virtual object as a sound source position, and output the synthesized sound from the right earphone and the left earphone. The sound output control unit adjusts the synthesized sound so that the synthesized sound becomes an audible sound bypassing the real object when the position of the real object is located between the position of the virtual object and a position of the display unit.
Method and system for generating an HRTF for a user
A method of obtaining a head-related transfer function for a user is provided. The method comprises generating an audio signal for output by a handheld device and outputting the generated audio signal at a plurality of locations by moving the handheld device to those locations. The audio output by the handheld device is detected at left-ear and right-ear microphones. A pose of the handheld device relative to the user's head is determined for at least some of the locations. One or more personalised HRTF features are then determined based on the detected audio and corresponding determined poses of the handheld device. The one or more personalised HRTF features are then mapped to a higher-quality HRTF for the user, wherein the higher-quality HRTF corresponds to an HRTF measured in an anechoic environment. This mapping may be learned using machine learning, for example. A corresponding system is also provided.
Wearer identification based on personalized acoustic transfer functions
A wearable device includes an audio system. In one embodiment, the audio system includes a sensor array that includes a plurality of acoustic sensors. When a user wears the wearable device, the audio system determines an acoustic transfer function for the user based upon detected sounds within a local area surrounding the sensor array. Because the acoustic transfer function is based upon the size, shape, and density of the user's body (e.g., the user's head), different acoustic transfer functions will be determined for different users. The determined acoustic transfer functions are compared with stored acoustic transfer functions of known users in order to authenticate the user of the wearable device.
SOUND EFFECT OPTIMIZATION METHOD, ELECTRONIC DEVICE, AND STORAGE MEDIUM
A sound effect optimizing method, an electronic device, and a non-transitory computer computer-readable storage medium are provided. The method includes controlling the speaker to play an audio signal emitted by a virtual sound source; receiving a sound source identifying result, the sound source identifying result including a first position relationship, and the first position relationship being a position relationship between the virtual sound source and a user and determined by the audio signal; and adjusting a sound effect parameter until the first position relationship is consistent with a second position relationship in response to the first position relationship being inconsistent with the second position relationship, the second position relationship being an actual position relationship between the virtual sound source and the user.
SPATIAL AUDIO CONTROLLER
A method performed a local device that is communicatively coupled with several remote devices, the method includes: receiving, from each remote device with which the local device is engaged in a communication session, an input audio stream; receiving, for each remote device, a set parameters; determining, for each input audio stream, whether the input audio stream is to be 1) rendered individually or 2) rendered as a mix of input audio streams based on the set of parameters; for each input audio stream that is determined to be rendered individually, spatial rendering the input audio stream as an individual virtual sound source that contains only that input audio stream; and for input audio streams that are determined to be rendered as the mix of input audio streams, spatial rendering the mix of input audio streams as a single virtual sound source that contains the mix of input audio streams.