FRAME SYNCHRONOUS PACKET SWITCHING FOR HIGH-DEFINITION MULTIMEDIA INTERFACE (HDMI) VIDEO TRANSITIONS
20200275052 ยท 2020-08-27
Assignee
Inventors
- Jiong Huang (San Jose, CA, US)
- Laurence A. Thompson (Morgan Hill, CA, US)
- Le Yuan (Shenzhen, CN)
- HUA LONG (SHENZHEN, CN)
- Yong Su (Shenzhen, CN)
- Zhigui WEI (Shenzhen, CN)
- Feng Wang (Shenzhen, CN)
Cpc classification
H04N7/0881
ELECTRICITY
G09G5/005
PHYSICS
G09G2370/04
PHYSICS
H04N19/46
ELECTRICITY
G09G2320/0673
PHYSICS
G09G2340/0407
PHYSICS
H04N7/0884
ELECTRICITY
H04N21/435
ELECTRICITY
G09G2370/12
PHYSICS
International classification
Abstract
An apparatus for use in a high-definition media interface (HDMI) source device includes an HDMI interface for transmitting video data and metadata to a sink device. The apparatus is configured to encode the metadata in an auxiliary video information (AVI) information frame (InfoFrame). The apparatus is further configured to transmit the AVI InfoFrame during a frame synchronous transmission window (FSTW) of the video data, wherein the FSTW begins during a video blanking interval (VBI) of the video data, on a first video blank pixel that immediately follows a last active video pixel of a preceding video frame or video field and ends a predetermined number of video lines after a start of the VBI.
Claims
1. (canceled)
2. An apparatus for use in a source device for transmitting data using a high definition media interface (HDMI), the apparatus comprising: an HDMI interface for transmitting data to and receiving data from a sink device; a memory holding executable code; and a processor, coupled to the memory and to the HDMI interface, the processor configured by the executable code to perform operations including: obtaining video data and metadata for transmission to the sink device; and transmitting the metadata during a frame-synchronous transmission window (FSTW) of the video data, wherein the FSTW begins during a video blanking interval (VBI) of the video data.
3. The apparatus of claim 2, wherein the FSTW begins on a first video blank pixel of a vertical blanking interval of a video field or video frame of the video data that immediately follows a last active video pixel of a preceding video frame or video field and ends a predetermined number of video lines after a start of the VBI.
4. The apparatus of claim 2, wherein the obtained metadata includes metadata for a static high dynamic range (S-HDR) video sequence and the operations further include: encoding the metadata for an S-HDR video field or S-HDR video frame of the S-HDR video sequence in an auxiliary video information (AVI) information frame and in a DRange and Mastering (DRAM) information frame; and transmitting the AVI information frame and the DRAM information frame during the FSTW of the S-HDR video field or S-HDR video frame.
5. The apparatus of claim 2, wherein the obtained metadata includes metadata for a dynamic high dynamic range (HDR) video sequence and the operations further include: encoding the metadata for a dynamic HDR video frame or dynamic HDR video field of the dynamic HDR video sequence in an auxiliary video information (AVI) information frame and in a HDR dynamic metadata extended (HDR DME) information frame; and transmitting the AVI information frame and the HDR DME information frame during the FSTW of the dynamic HDR video field or dynamic HDR video frame.
6. The apparatus of claim 2, wherein the operation of transmitting the metadata via the HDMI interface further includes encoding the metadata in an auxiliary video information (AVI) information frame.
7. The apparatus of claim 2, wherein: the operation of obtaining the video data and the metadata for transmission to the sink device further includes obtaining video data and metadata for a static high dynamic range (S-HDR) video sequence and obtaining video data and metadata for a dynamic high dynamic range (HDR) video sequence; and the operation of transmitting the metadata via the HDMI interface further includes: transmitting the metadata for the S-HDR video sequence during the FSTW of a first S-HDR video field or S-HDR video frame of the S-HDR video sequence; and transmitting the metadata for the dynamic HDR video sequence during the FSTW a first dynamic HDR video field or dynamic HDR video frame of the dynamic HDR video sequence.
8. The apparatus of claim 7, wherein: the operation of obtaining the video data and the metadata for transmission to the sink device further includes obtaining video data and metadata for standard dynamic range (SDR) video sequence; and the operation of transmitting the metadata via the HDMI interface further includes: transmitting the metadata for the SDR video sequence during the FSTW of a first SDR video field or SDR video frame of the SDR video sequence.
9. The apparatus of claim 2, wherein obtained video data includes standard dynamic range (SDR) video data and the metadata includes metadata for the SDR video data.
10. An apparatus for use in a sink device for receiving data using a high-definition media interface (HDMI), the apparatus comprising: an HDMI interface for receiving data from a source device; a memory holding executable code; and a processor, coupled to the memory and to the HDMI interface, the processor configured by the executable code to perform operations including: receiving a video sequence from the source device via the HDMI interface, the video sequence including a plurality of video fields or video frames, each video field or video frame including an active video interval and one or more video blanking intervals (VBIs); extracting metadata for the video sequence from a frame-synchronous transmission window (FSTW) one video field or video frame of the video sequence, wherein the FSTW begins during one VBI of the one or more VBIs, on a first video blank pixel that immediately follows a last active video pixel of a preceding video frame or video field and ends a predetermined number of video lines after a start of the one VBI; and applying the extracted metadata to video data in the active video interval of the one video field or video frame.
11. The apparatus of claim 10, wherein the received video sequence includes a static high dynamic range (S-HDR) video sequence including a plurality of S-HDR video fields or S-HDR video frames and the operations further include: extracting a DRange and Mastering (DRAM) information frame from at least one FSTW of at least one S-HDR video field or S-HDR video frame of the plurality of S-HDR video fields or S-HDR video frames: extracting the metadata from the DRAM information frame; and applying the extracted metadata to the video data in the active video interval of the one S-HDR video field or S-HDR video frame.
12. The apparatus of claim 10, wherein the received video sequence includes a dynamic high dynamic range (HDR) video sequence and the operations further include: extracting a dynamic HDR dynamic metadata extended (HDR DME) information frame from at least one FSTW of one dynamic HDR video field or dynamic HDR video frame of the HDR video sequence; extracting the metadata from the HDR DME information frame; and applying the extracted metadata to the video data in the active video interval of the one dynamic HDR video field or dynamic HDR video frame.
13. The apparatus of claim 10, wherein: the operation of receiving the video sequence includes receiving a static high dynamic range (S-HDR) video sequence and receiving a dynamic high dynamic range (HDR) video sequence; and the operation of extracting the metadata from the FSTW further includes: extracting the metadata for the S-HDR video sequence from the FSTW of a first S-HDR video field or S-HDR video frame of the S-HDR video sequence; and extracting the metadata for the dynamic HDR video sequence from the FSTW of a first dynamic HDR video field or dynamic HDR video frame of the dynamic HDR video sequence.
14. the apparatus of claim 10, wherein the operation of extracting the metadata from the video sequence further includes extracting the metadata from an auxiliary video information (AVI) information frame of the video sequence.
15. The apparatus of claim 10, wherein video sequence includes a standard dynamic range (SDR) video sequence and the operation of extracting the metadata from the video sequence further includes extracting metadata for the SDR video sequence.
16. A method for processing data from a high-definition media interface (HDMI), the method comprising: obtaining a video sequence via an HDMI interface, the video sequence including a plurality of video fields or video frames, each video field or video frame including an active video interval and one or more video blanking intervals (VBIs); extracting metadata for the video sequence from a frame-synchronous transmission window (FSTW) of one video field or video frame of the video sequence, wherein the FSTW begins during one VBI of the one or more VBIs of the video data, on a first video blank pixel that immediately follows a last active video pixel of a preceding video frame or video field and ends a predetermined number of video lines after a start of the one VBI; and applying the extracted metadata to video data in the active video interval of the one video field or video frame.
17. The method of claim 16, wherein the video sequence includes a static high dynamic range (S-HDR) video sequence including a plurality of S-HDR video fields or S-HDR video frames and the method further comprises: extracting a DRange and Mastering (DRAM) information frame from at least one FSTW of at least one S-HDR video field or S-HDR video frame of the plurality of S-HDR video fields or S-HDR video frames; extracting the metadata from the DRAM information frame; and applying the extracted metadata to the video data in the active video interval of the one S-HDR video field or S-HDR video frame.
18. The method of claim 16, wherein the video sequence includes a dynamic high dynamic range (HDR) video sequence and the method further comprises: extracting a dynamic HDR dynamic metadata extended (HDR DME) information frame from at least FSTW of at least one dynamic HDR video field or dynamic HDR video frame of the HDR video sequence; extracting the metadata from the HDR DME information frame; and applying the extracted metadata to the video data in the active video interval of the one dynamic HDR video field or dynamic HDR video frame.
19. The method of claim 16, wherein: the obtaining of the video sequence includes obtaining a static high dynamic range (S-HDR) video sequence and obtaining a dynamic high dynamic range (HDR) video sequence; and the extracting of the metadata from the FSTW further includes: extracting the metadata for the S-HDR video sequence from the FSTW of a first S-HDR video field or S-HDR video frame of the S-HDR video sequence; and extracting the metadata for the dynamic HDR video sequence from the FSTW of a first dynamic HDR video field or dynamic HDR video frame of the dynamic HDR video sequence.
20. the method of claim 16, wherein the extracting of the metadata from the video sequence further includes extracting the metadata from an auxiliary video information (AVI) information frame of the video sequence.
21. The method of claim 16, wherein video sequence includes a standard dynamic range (SDR) video sequence and the extracting of the metadata from the video sequence further includes extracting metadata for the SDR video sequence.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0028]
[0029]
[0030]
[0031]
[0032]
[0033]
[0034]
[0035]
[0036]
[0037]
[0038]
[0039]
DETAILED DESCRIPTION
[0040] In the following description, reference is made to the accompanying drawings that form a part hereof, and in which is shown by way of illustration specific embodiments which may be practiced. These embodiments are described in sufficient detail to enable those skilled in the art to practice the subject matter, and it is to be understood that other embodiments may be utilized and that structural, logical, and electrical changes may be made without departing from the scope of the present subject matter. The following description of example embodiments is, therefore, not to be taken in a limited sense, and the scope of the present subject matter is defined by the appended claims.
[0041] The functions or algorithms described herein may be implemented in software in one embodiment. The software may consist of computer-executable instructions stored on computer-readable media or computer-readable storage device such as one or more non-transitory memories or other type of hardware based storage devices, either local or networked. Further, such functions correspond to modules, which may be software, hardware, firmware, or any combination thereof. Multiple functions may be performed in one or more modules as desired, and the embodiments described are merely examples. The software may be executed on processing circuitry that may include a single core microprocessor, multi-core microprocessor, digital signal processor, application specific integrated circuit (ASIC), field programmable gate array (FPGA), or other type of data processing circuitry operating on a computer system, such as a personal computer, server or other computer system, turning such computer system into a specifically programmed machine.
[0042] In many existing systems, video information originates from a single source such as a digital versatile disk (DVD) player or a television tuner. These sources typically provide video data with a uniform dynamic range and may provide either SDR data or S-HDR data. To display video data from these sources, the HDMI interface provides for S-HDR metadata signaling (e.g., AVI InfoFrames and DRAM InfoFrames) and SDR signaling (e.g., AVI InfoFrames).
[0043] S-HDR signaling works well when the video data changes between HDR and SDR infrequently (e.g., when an S-HDR disk is inserted in the DVD player). Increasingly, however, video data is provided in a streaming format in which disparate video segments are stitched together into a single stream. Some segments may be SDR segments while others are HDR segments. As described below with reference to
[0044] More recently, different types of HDR video data may be provided in a single scene or for a single frame. For example, in a relatively dark scene, the range of luminance values may be significantly less than the full range of the HDR signal. For example, a 10-bit luminance signal may have values bounded by 0-255, the range of an 8-bit video signal. In this instance, an opto-electric transfer function (OETF) and corresponding electro-optical transfer function (EOTF) may be applied so that the image data in the scene may be mapped into the 10-bit range of the luminance signal, reducing quantization distortion in the reproduced image. These signals are dynamic HDR signals that may use HDR DME InfoFrames to send the EOTF to the sink device.
[0045] Because the dynamic HDR video signals having HDR DME may change on a frame-by-frame basis, the HDR DME InfoFrames are processed with frame-synchronous timing to ensure proper display of the HDR video data. The embodiments described below also send AVI InfoFrames and DRAM InfoFrames in a frame-synchronous transmission window (FSTW). The FSTW, which has the same timing as FAPA with location start 0 (FAPA0), starts on the first video blank pixel that immediately follows the last active video pixel of a video frame/field and ends FAPA_end lines prior to the start of the next active region (as described in section 10.10.1.1 of the High-Definition Multimedia Interface Specification Version 2.1). Briefly, FAPA_end may be one-half the number of lines in the VBI or less, depending on the number of lines in the VBI. The FSTW is used by sink devices compatible with dynamic HDR video and has timing that corresponds to the FAPA. Sending the AVI InfoFrames and DRAM InfoFrames as well as HDR DME InfoFrames during the FSTW reduces image distortion that may occur on switching among SDR, S-HDR and dynamic HDR video formats. As used herein, FSTW is identical to FAPA0.
[0046]
[0047] The TMDS channels 120, 122, and 124 allow the source device to transmit video and audio data 154, 156 to the sink device at rates up to 6 gigabits per second (Gbps) using differential signals synchronized by the clock signal transmitted through the TMDS clock channel 126. The audio data 156 may be encoded in data islands, described below, that are transmitted in the vertical and horizontal blanking intervals of the transmitted video data 154.
[0048] The DDC 128 is a serial channel that includes a serial data (SDA) conductor (not separately shown) and a serial clock (SCL) conductor (not separately shown). The DDC 128 is used to send/receive control data between the sending unit 110 and the receiving unit 150. For example, the sending unit 110 may use the DDC 128 to read enhanced extended display identification data (E-EDID), such as a vendor-specific data block (VSDB) from the receiving unit 150. For this operation, the receiving unit 150 may include a read only memory (ROM) (not shown) that stores the E-EDID of the HDMI receiving unit 150.
[0049] The sending unit 110 uses the HPD line to sense that the sink device is coupled to the cable 140 and is powered on. Responsive to the HPD line having a positive DC bias potential, the sending unit 110 reads the E-EDID data via the DDC 128 to determine the capabilities of the receiving unit 150. The CEC channel 130 allows users to control devices connected by the HDMI cable 140 using a single remote control device (not shown). As described below, the E-EDID may include information about the HDR capabilities of the sink device, for example, whether the sink device supports S-HDR and/or dynamic HDR.
[0050]
[0051]
[0052] The processor 202 controls the operation of other components of the HDMI source device 200. The memory 204 holds data and instructions for the processor 202. The processor 202 may operate the display controller 206 to control a display panel (not shown) used to control the operation of the HDMI source device 200. The display controller 206 may also interface with an input device such as a touchscreen and/or keypad (not shown) to allow a user to input data for controlling the HDMI source device 200. The processor 202 may also control the network interface 208 to allow the source device 200 to access media content from a network (e.g., the Internet) via a browser or a video streaming application. As described above, this media content may be streaming video including SDR segments, S-HDR segments, and/or dynamic HDR segments. The communication interface 212 of the HDMI sending unit 210 is controlled by the processor 202 to communicate with the sink device (described below with reference to
[0053] In the example source device 200, compressed video and audio data from the DVD interface 220 and/or the network interface 208 are provided to the audio video decoder 214. The decoder 214 may include a motion picture experts group (MPEG) decoder such as an H.222/H.262 (MPEG2), H.264 advanced video coding (AVC), and/or H.265 high efficiency video coding (HEVC) decoder. The decoder 214 generates baseband video and audio data from the encoded data provided by the network interface 208, DVD interface 220, or provided directly to the AV decoder 214 as indicated in
[0054] When the encoded video stream includes high dynamic range video data, the audio/video decoder 214 extracts the HDR metadata (e.g., DRAM and/or HDR DME) from the encoded video data and provides it to the HDMI sending unit 210 to be included in data islands to be transmitted inside or outside of frame synchronous transmission windows (FSTWs) of the video data sent to the HDMI receiving unit. For video data provided directly to the audio video decoder 214, any associated HDR metadata may be provided to the metadata acquisition circuitry 218. This metadata may be provided to the InfoFrame processing circuitry 216 to be included in the data islands transmitted by the HDMI transmitter 211.
[0055] If the sink device 300 (
[0056]
[0057] The processor 302 controls the operation of other components of the HDMI sink device 300. The memory 304 holds data and instructions for the processor 302. The processor 302 may operate the display controller 306 to control a display panel (not shown) used to control the operation of the HDMI sink device 300. The controller 306 may also interface with an input device such as a touchscreen and/or keypad (not shown) to allow a user to input data for controlling the HDMI sink device 300. The sink device 300 receives audio and video data via the TMDS channels 120, 122, 124 and 126 or FRL lanes 166, 168, 170 and 172, described above with reference to
[0058] The HDMI receiving unit 310 extracts audio data from the data islands in the horizontal and vertical blanking intervals of the video signal outside of the FSTW and provides the audio data to the audio processing circuitry 318. The audio data generated by the audio processing circuitry 318 and the video data generated by the video processing circuitry 316 are provided to a presentation device including a monitor (not shown) and a sound system (not shown).
[0059] Each of the memories 204 and 304 may include volatile memory and/or non-volatile memory. The non-volatile memory may include removable storage and non-removable storage. Computer storage includes random access memory (RAM), read-only memory (ROM), erasable programmable read-only memory (EPROM) and electrically erasable programmable read-only memory (EEPROM), flash memory or other memory technologies, compact disc read-only memory (CD ROM), Digital Versatile Disks (DVD) or other optical disk storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium capable of storing computer-readable instructions.
[0060] The various processing devices and circuits shown in
[0061] As described below with reference to
[0062] The communication interface 312 of the HDMI receiving unit 310 is controlled by the processor 302 to communicate with the source device 200 via the DDC/SDA/SCL channel of the HDMI interface. The processor 202 uses this interface to receive commands and data from, and to transmit commands and data to, the source device 200 via the communication interface 312. For example, the sink device 300 may provide to the source device 200 information (e.g., a vendor-specific data block (VSDB)) indicating the capabilities of the sink device 300 Similarly, the sink device 300 may obtain information about the source device 200 via the DDC/SDA/SCL channel of the HDMI interface.
[0063]
[0064] In sink devices that support frame-synchronous processing, control information in the HDR DME is applied to the immediately following active video data so dynamic HDR video data in the active video area 420 is properly displayed. Sink devices supporting frame synchronous processing identify the HDR DME and copy metadata data to appropriate control registers and memory elements in the sink device 300. This may include, for example, copying EOTF data to implement a particular EOTF to be used for displaying the dynamic HDR video data or configuring the sink device 300 to handle the pixel depth (e.g., the number of bits in each pixel) or a particular color space configuration indicated by the HDR DME.
[0065] The example sink device 300 includes a vendor-specific data block (VSDB) (not shown), for example in the E-EDID, containing information on the capabilities of the sink device 300. The VSDB may indicate that the sink device 300 supports only SDR video data; SDR and S-HDR video data; or SDR, S-HDR, and dynamic HDR video data. As described above, when the sink device 300 does not support either dynamic HDR video data or S-HDR data, the source device may convert the dynamic HDR data to S-HDR data compatible with the AVI InfoFrames, and may convert the S-HDR data to SDR data compatible with the AVI InfoFrames before sending the converted video data to the sink device 300. The example embodiments send the AVI InfoFrames, DRAM InfoFrames, and HDR DME InfoFrames during the region of the vertical blanking interval beginning at the first blank pixel that immediately follows the last active video pixel of a video frame/field and ending FAPA_end lines prior to the start of the next active region. This region corresponds to the FSTW 414 described above.
[0066]
[0067]
[0068] Metadata for the SDR and S-HDR video data is contained in AVI InfoFrames. Although
[0069] The metadata for the S-HDR video sequence 604 is contained in an AVI InfoFrame 614 and in DRAM InfoFrame 622, which are transmitted by the source device 200 at field/frame time T100 but do not become active until field/frame time T101. Similarly, at field/frame time T500, the source device 200 sends the second S-HDR metadata in AVI InfoFrame 618 and DRAM InfoFrame 624. The metadata in these two InfoFrames 618, 624 becomes active at field/frame time T501 and remains active until time T601, when the metadata in the AVI InfoFrame 620 for the third SDR sequence 610 becomes active.
[0070] In
[0071] As shown in
[0072]
[0073] The metadata for the SDR video is contained in AVI InfoFrames. The AVI InfoFrame 716 containing metadata or the first SDR video sequence 702 is received in data islands during the non-FAPA area of the VBI or during the HBI of the field/frame starting at field/frame time T0. As shown in
[0074] At field/frame time T200, the sink receives the first dynamic HDR video sequence 706 and accompanying metadata including AVI InfoFrame 720 and HDR DME 734. The AVI InfoFrame 720 is received outside of the FAPA interval of the VBI while the HDR DME is received during the FAPA interval (FAPA0 or FAPA1) of the VBI. As shown in
[0075] At field/frame time T300, the sink receives the second SDR video sequence 708 and the AVI InfoFrame 722 containing the metadata for the second SDR sequence 708. Because the AVI InfoFrame 722 is received outside of the FAPA area of the VBI, it does not become active until field/frame time T301 and remains active until field/frame time T400.
[0076] At field/frame time T400, the sink receives the second dynamic HDR video sequence 710 and accompanying metadata including AVI InfoFrame 724 and HDR DME 738. As shown in
[0077] The sink receives the second S-HDR video sequence 712 and accompanying metadata at field/frame time T500. The S-HDR metadata includes AVI InfoFrame 726 and DRAM InfoFrame 740. Both of these frames are received outside of the FAPA area of the VBI and, thus, do not become active until field/frame time T501. The metadata for the second S-HDR video sequence 712 remains active between field/frame times T501 and T601.
[0078] At time T601, the sink receives the third SDR video sequence 714 and its accompanying metadata, AVI InfoFrame 728. Because the AVI InfoFrame 728 is received outside of the FAPA area, it does not become active until field/frame time T601.
[0079] The actual flow includes several instances of mismatch between the displayed video data and the dynamic range metadata used to process the video data. For example, the display begins with displayed SDR video sequence 742 at field/frame time T100 followed by a mismatch interval 744 between field/frame times T100 and T101. This mismatch occurs because the first S-HDR video sequence 704 is processed using the SDR metadata because the metadata in the AVI InfoFrame 718 and DRAM InfoFrame 730 for the S-HDR video sequence 704 have not been transferred to the InfoFrame processing circuitry 314 (e.g., have not become active) until field/frame time T101. From field/frame time T101 to T200, the S-HDR video data 746 is properly displayed using the first S-HDR metadata. Even though the DRAM InfoFrame 730 metadata is active until field/frame time 201, there is no mismatch at the transition beginning at field/frame time T200 because the HDR DME 734 metadata overrides the DRAM InfoFrame 730 metadata. Because it is received during the FAPA0 interval, the first HDR DME 734 metadata is processed in a frame-synchronous manner and is transferred to the InfoFrame processing circuitry 314 so that the metadata may be passed to the video processing circuitry 316 in time to process the video data at field/frame time T200. The displayed dynamic HDR video sequence 748 continues to field/frame time T300 at which there is another mismatch 750. At field/frame time T300, the first HDR DME metadata 734 is no longer active; however, the second SDR metadata has not yet become active. The mismatch 750 occurs because the SDR video information in the field/frame starting at time T300 is processed using the AVI InfoFrame 720 metadata. Once SDR metadata in AVI InfoFrame 722 becomes active at field/frame time T301, the system properly displays the SDR video data 752 until field/frame time T400. At T400, again due to the frame-synchronous processing, the system properly displays the dynamic HDR video data 754 using the second HDR DME 738 metadata and AVI InfoFrame 724. A mismatch 756 occurs, however, in the field/frame starting at T500 because the second S-HDR metadata in DRAM InfoFrame 740 has not become active, so that the corresponding S-HDR video data is processed using the metadata in the AVI InfoFrame 724 for the second dynamic HDR video sequence. Once the metadata in the AVI InfoFrame 726 and the DRAM InfoFrame 740 become active at T501, the second S-HDR video data 758 is displayed properly. The actual flow continues at field/frame time T600 with another mismatch 760, when the third SDR video sequence 714 is processed using the second S-HDR metadata contained in the InfoFrames 726 and 740. The SDR video data 762 displays properly after field/frame time T601.
[0080] Although the examples in
[0081] The visual artifacts that occur on switching to SDR or S-HDR from dynamic HDR may be more noticeable than those which occur on switching between SDR and S-HDR because, due to the dynamic nature of dynamic HDR metadata, the changes may be less predictable, unlike legacy HDMI in which the changes are static or pseudo-static. The HDMI 2.1 Specification implements frame-accuracy for switching on HDR DME processing but not for switching off HDR DME processing. The visual artifacts experienced during the mismatch intervals may include reduced contrast, for mismatch interval 744, when S-HDR video is incorrectly interpreted as SDR video, or incorrect dimming with missing shadow details for mismatch 760, when SDR video is incorrectly interpreted as S-HDR video. The artifacts may also include incorrect color. The occurrence of these artifacts may be increased in systems operating according to the HDMI 2.1 standard due to the addition of dynamic HDR sequences, since the dynamic HDR sequences may be stitched with S-HDR or SDR in a linear stream before delivery, resulting in more frequent and more visible artifacts.
[0082]
[0083]
[0084] As shown in
[0085] In the example shown in
[0086] To minimize visual artifacts in sinks that do not support frame-accuracy, the source device 200 sends the data to the sink device 300 according to the legacy HDMI standards so that all video packets are accurately processed within a set amount of time, for example, one to four fields/frames times after each video transition.
[0087]
[0088] When, at block 904, the source device 200 determines that the sink device 300 can process dynamic HDR video sequences, the source device 200, at block 908, formats the video data so that all of the metadata in the AVI InfoFrames, DRAM InfoFrames, and HDR DME InfoFrames is sent during the FSTW.
[0089]
[0090] The metadata describes how the video data sent during the active video interval is to be displayed. For example, the metadata may include: information on color remapping; a color volume transform to be applied; maximum, minimum and average luminance values in a scene and target maximum, minimum and average luminance values; data describing a transfer function (e.g. an EOTF) to be applied to the luminance data; and/or data specific to an application running on the source device. The content and format of the metadata is described in a standard issued by the Consumer Technology Association, entitled A DTV Profile for Uncompressed High Speed Digital Interfaces CTA-861-G (November 2016).
[0091] With reference to
[0092]
[0093] As described above, with reference to
[0094] As shown in
[0095] Although the examples described above concern metadata transitions related to the changing dynamic range of the video signals, it is contemplated that other metadata transitions in video or audio signals may be implemented as frame-synchronous transitions. For example, object-oriented audio and video data such as may be used in virtual-reality and augmented-reality applications may be transmitted through the HDMI interface. In this instance, frame-synchronous processing may be desirable to coordinate the video and audio data to motions and/or gestures of the user.