Method and apparatus for composition of subtitles
09749576 ยท 2017-08-29
Assignee
Inventors
- Dirk Gandolph (Ronnenberg, DE)
- Jobst Hoerentrup (Wenningsen, DE)
- Ralf Ostermann (Hannover, DE)
- Hartmut Peters (Barsinsinghausen, DE)
- Harald Schiller (Hannover, DE)
Cpc classification
G11B27/32
PHYSICS
G11B27/10
PHYSICS
H04N21/4312
ELECTRICITY
International classification
H04N9/80
ELECTRICITY
H04N9/82
ELECTRICITY
H04N5/92
ELECTRICITY
Abstract
Embodiments of the invention include a subtitling format encompassing elements of enhanced syntax and semantic to provide improved animation capabilities. The disclosed elements improve subtitle performance without stressing the available subtitle bitrate. This will become essential for authoring content of high-end HDTV subtitles in pre-recorded format, which can be broadcast or stored on high capacity optical media, e.g. the Blue-ray Disc. Embodiments of the invention include abilities for improved authoring possibilities for the content production to animate subtitles. For subtitles that are separate from AV material, a method includes using one or more superimposed subtitle layers, and displaying only a selected part of the transferred subtitles at a time. Further, colors of a selected part of the displayed subtitles may be modified, e.g. highlighted.
Claims
1. Apparatus for composition of subtitles for audio/video presentations, wherein data streams containing video, audio and subtitle information are transferred from a storage medium, the apparatus comprising: a processor for retrieving from said storage medium subtitle layer data, wherein the subtitle layer data contain graphic subtitle elements and data for a first color look-up table; the processor extracting from said retrieved subtitle layer data subtitle display region information for defining a subtitle display region and subtitle display region cropping information, for cropping the display region, the subtitle display region cropping information comprising a parameter for horizontal position of the cropped subtitle display region, a parameter for vertical position of the cropped subtitle display region, a parameter for width of the cropped subtitle display region, and a parameter for height of the cropped subtitle display region; and a renderer for enabling automatic cropping of rectangular parts of the subtitle elements to be displayed in the cropped subtitle display region; wherein the cropped parts of the subtitle elements to be displayed are defined by said subtitle display region cropping information; and wherein said first color look-up table is to be used within said cropped parts of the subtitle elements.
Description
DRAWINGS
(1) Exemplary embodiments of the invention are described with reference to the accompanying drawings and tables, which show:
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
EXEMPLARY EMBODIMENTS OF THE INVENTION
(11) The invention can preferably be embodied based on the syntax and semantic of the DVB subtitle specification (DVB-ST). To provide improved capabilities for the manipulation of graphic subtitle elements, the semantics of DVB-ST's page composition segment (PCS) and region composition segment (RCS) are expanded.
(12) DVB_ST uses page composition segments (PCS) to describe the positions of one or more rectangular regions on the display screen. The region composition segments (RCS) are used to define the size of any such rectangular area and identifies the color-lookup-table (CLUT) used within.
(13) Embodiments of the proposed invention keeps backward compatibility with DVB-ST by using different segment_types for the enhanced PCS and RCS elements, as listed in
(14)
(15) The enhanced PCS shown in
(16) The enhanced RCS shown in
(17) The enhanced PCS and enhanced RCS elements provide the ability that subtitles can be manipulated independent from the encoding method i.e. independent from whether they are encoded as character data or pixel data.
(18) The enhanced PCS and RCS can be used to perform many different animation effects for subtitles. Those could be wiping boxes, blinds, scrolling, wipes, checker boxes, etc. The following figures show an application example for karaoke.
(19)
(20) The region sub-CLUT location shown in the lower part of
(21) Picking up all parameters defined with the previous figures results in the displayed subtitle as depicted in
(22) As the enhanced PCS are sent within MPEG packet elementary stream (PES) packets labeled by presentation time stamps (PTS), any effect can be synchronized to the AV.
(23) Another idea of the invention is the superseding of subtitle animation parameters by the user. This offers a way to realize interactive subtitles. The enhanced PCS parameters are transferred as a default, and the user may change them via a remote control for example. Thus the user is able to move, crop or highlight the subtitle.
(24) This could be an advantage for a user defined repositioning of a subtitling text, so that the user can subjectively minimize the annoyance by the subtitle text placement on top of the motion video. Also the color of the subtitles could be set according to users preferences.
(25) Another application for overriding subtitle animation parameters like position, cropping rectangle, CLUTs and sub-CLUTs is the realization of some very basic sort of interactive gaming. The subtitle may carry pixel data of an animated character. This character is subsequently moved on the display screen driven by either user interaction, programmatic control or both.
(26) The overriding of subtitle animation parameters can be implemented in at least two ways. The first option is that the overriding parameters SD replace the parameters DD send in the bitstream. The second option is that the overriding parameters SD are used as an offset that is added to or subtracted from the subtitle animation parameters DD send in the bitstream.
(27) The enhanced PCS and RCS provide a lot more of animation capabilities not explained. Following is a non-exhaustive list of examples: wiping boxes, blinds, scrolling, wipes, checker boxes in details.
(28) Exemplary video and graphics planes are shown in
(29)
(30) The apparatus contains a still picture decoder SPDec and an MPEG-2 video decoder MVDec, but since only one of them is used at a time, a switch s1 can select which data shall be used for further processing. Moreover, two identical decoders AVSGDec1,AVSGDec2 are used for decoding subtitle and animation data. The outputs of these two decoders AVSGDec1, AVSGDec2 may be switched by independent switches s2,s3 to either a mixer MX, or for preprocessing to a mixer and scaler MXS, which outputs its resulting data to said mixer MX. These two units MX, MXS are used to perform the superimposing of its various input data, thus controlling the display order of the layers. The mixer MX has inputs for a front layer f2, a middle front layer mf, a middle back layer mb and a background layer b2. The front layer f2 may be unused, if the corresponding switch s3 is in a position to connect the second AV sync graphics decoder AVSGDec2 to the mixer and scaler MXS. This unit MXS has inputs for front layer f1, middle layer m and background layer b. It superimposes these data correspondingly and sends the resulting picture data to the background input b2 of the mixer MX. Thus, these data represent e.g. a frame comprising up to three layers of picture and subtitles, which can be scaled and moved together within the final picture. The background input b1 of the mixer and scaler MXS is connected to the switch s1 mentioned above, so that the background can be generated from a still picture or an MPEG-2 video.
(31) The output of the first AV sync graphics decoder AVSGDec1 is connected to a second switch s2, which may switch it to the middle layer input m of the mixer and scaler MXS or to the middle back layer input mb of the mixer MX. The output of the second AV sync graphics decoder AVSGDec2 is connected to a third switch s3, which may switch it to the front layer input f1 of the mixer and scaler MXS or to the middle front layer input mf of the mixer MX.
(32) Depending on the positions of the second and third switch s2,s3, either the output of the first or the second AV sync graphics decoder AVSGDec1,AVSGD2 may have priority over the other, as described above. For having the data from the first decoder AVSGDec1 in the foreground, the second switch s2 may route the subtitle data to the middle back input mb of the mixer MX, while the third switch s3 routes the animation graphics data to the front input f1 of the mixer and scaler MXS, so that it ends up at the background input b2 of the mixer MX. Otherwise, for having the data from the second decoder AVSGDec2 in the foreground, the switches s2,s3 may route their outputs to the same unit, either the mixer and scaler MXS or the mixer MX, as shown in