System and method for presenting a video via transcode
11197058 · 2021-12-07
Assignee
Inventors
Cpc classification
H04N21/41407
ELECTRICITY
H04N23/57
ELECTRICITY
H04N7/18
ELECTRICITY
H04N21/440281
ELECTRICITY
H04N21/440263
ELECTRICITY
H04N21/440236
ELECTRICITY
International classification
H04N7/18
ELECTRICITY
H04N21/4402
ELECTRICITY
H04N21/414
ELECTRICITY
Abstract
A method for handling video includes extracting video content and context information from a video file. The context information is associated with the video content. The method further includes transmitting the video content via a first communication path and transmitting the context information via a second communication path separate from the first communication path.
Claims
1. A method for handling video comprising: using an imaging device of an unmanned aerial vehicle (UAV) including one or more propulsion units providing lifting power and a storage medium storing a video file to perform: extracting video content and context information from a video file, the context information being associated with the video content; transmitting the video content via a first communication path; decoding the context information to generate a first context bitstream; downscaling the first context bitstream to generate a second context bitstream; encoding the second context bitstream to generate an auxiliary context bitstream; and transmitting the auxiliary context bitstream via a second communication path separate from the first communication path.
2. The method of claim 1, wherein extracting the context information includes extracting at least one of audio data, caption data, environmental data, or configuration data from the video.
3. The method of claim 1, wherein downscaling the first context bitstream includes reducing a bitrate or a bit per second rate of the first context bitstream.
4. The method of claim 3, wherein reducing the bitrate includes reducing a sampling rate of the first context bitstream.
5. The method of claim 1, wherein encoding the second context bitstream includes encoding the second context bitstream with a protocol that compresses the second context bitstream.
6. The method of claim 5, wherein the protocol is recognized by an application (“APP”) of a remote device receiving the auxiliary context bitstream.
7. The method of claim 1, wherein transmitting the video content further includes transcoding the video content to generate an encoded video bitstream.
8. The method of claim 7, further comprising: merging the encoded video bitstream with the auxiliary context bitstream to generate a composite bitstream.
9. The method of claim 8, wherein merging the encoded video bitstream with the auxiliary context bitstream includes synchronizing the encoded video bitstream and the auxiliary context bitstream to generate the composite bitstream.
10. The method of claim 9, wherein synchronizing the encoded video bitstream and the auxiliary context bitstream includes matching header information of data packets of the encoded video bitstream with header information of data packets of the auxiliary context bitstream.
11. The method of claim 10, wherein matching the header information of the data packets of the encoded video bitstream with the header information of the data packets of the auxiliary context bitstream includes performing at least one of: matching a timestamp of the data packets of the encoded video bitstream with a timestamp of the data packets of the auxiliary context bitstream; or matching a serial number of the data packets of the encoded video bitstream with a serial number of the data packets of the auxiliary context bitstream.
12. The method of claim 8, further comprising: presenting the composite bitstream via a display.
13. The method of claim 12, wherein presenting the composite bitstream includes presenting the composite bitstream with an application of a remote device.
14. The method of claim 12, wherein presenting the composite bitstream includes presenting the composite bitstream in a fast-forward mode or a fast-rewind mode.
15. The method of claim 1, wherein the video file is captured by an imaging device on an unmanned aerial vehicle (“UAV”) and stored in a storage medium on the UAV; the method further comprising, before extracting the video content and the context information: acquiring the video file from the storage medium.
16. An Unmanned Aerial Vehicle (UAV) comprising: one or more propulsion units providing lifting power; a storage medium storing a video file; and an imaging device including an imaging processor for: extracting video content and context information from the video file, the context information being associated with the video content; transmitting the video content via a first communication path; decoding the context information to generate a first context bitstream; downscaling the first context bitstream to generate a second context bitstream; encoding the second context bitstream to generate an auxiliary context bitstream; and transmitting the auxiliary context bitstream via a second communication path separate from the first communication path.
17. An imaging system for handling video comprising: a storage medium storing a video file; and one or more processors, individually or collectively: extracting video content and context information from the video file, the context information being associated with the video content; transmitting the video content via a first communication path; decoding the context information to generate a first context bitstream; downscaling the first context bitstream to generate a second context bitstream; encoding the second context bitstream to generate an auxiliary context bitstream; and transmitting the auxiliary context bitstream via a second communication path separate from the first communication path; wherein the imaging system is mounted on an unmanned aerial vehicle (UAV) including one or more propulsion units providing lifting power.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1)
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12)
(13)
(14)
(15)
(16)
(17)
(18)
(19) It should be noted that the figures are not drawn to scale and that elements of similar structures or functions are generally represented by like reference numerals for illustrative purposes throughout the figures. It also should be noted that the figures are only intended to facilitate the description of the embodiments. The figures do not illustrate every aspect of the described embodiments and do not limit the scope of the present disclosure.
DETAILED DESCRIPTION OF THE EMBODIMENTS
(20) In currently available presenting systems, such as aerial imaging with an Unmanned Aerial Vehicle (“UAV”) systems and other remotely controlled camera systems, wireless transmission is normally applied for transmitting the video. Such systems however are susceptible to transmission bottlenecks due to insufficient transmission bandwidth. Although some systems attempt to alleviate the bottleneck issue by implementing bitrate reduction via HDMI or BT 656 ports before transmitting the video, such systems require an extra central processing unit (“CPU”), resulting in greater complexity and much higher implementation costs.
(21) Since currently-available remote presenting solutions require costly hardware and can cause communication bottlenecks, a system and method for presenting a video clip with assistance of available resources of an imaging device advantageously can alleviate the communication bottleneck and provide a basis for providing remote presenting function for mobile platforms, such as Unmanned Aerial Vehicles (“UAVs”). This result can be achieved, according to one embodiment, by an exemplary system 200 illustrated in
(22) Turning to
(23) Although shown and described as reading the video content from the storage medium 600 for purposes of illustration only, the ISP 320 can read the video content from any other suitable sources or acquire the video content by extracting the video content from the video clip.
(24) The ISP 320 can include a transcoding module 830 for transcoding the video content. The transcoding module 830 can be a software program that is executed by the ISP 320 or can include hardware, software, firmware or a combination thereof. The transcoding module 830 can transcode the video content via the steps shown and described with reference to
(25) In some embodiments, the ISP 320 can read context information of the video clip from the storage medium 600, at 821. The ISP 320 can read the context information of the video clip from the storage medium 600 via any conventional manners. Alternatively and/or additionally, the ISP 320 can read the context information from any other suitable sources or acquire the context information by extracting the context information from the video clip.
(26) The ISP 320 can include a merger 820 that can be a software program that is executed by the ISP 320 or can include hardware, software, firmware or a combination thereof. The merger 820 can merge the transcoded video content with the context information to generate a combined video bitstream. Additional detail regarding the merger 820 and the combined video bitstream will be provided below with reference to
(27) The combined video bitstream can be transmitted, via a transmission medium 700, to a mobile device 900 for presenting the video clip. Although shown and described as transcoding the video content for purposes of illustration only, the ISP 320 can optionally transcode the context information. Because of reduced bitrate of the video content and/or the reduced bitrate of the context information, the system 200 can advantageously alleviate a communication bottleneck with the transmission medium 700.
(28) Turning to
(29) The video clip can be contained a video file that is captured by the imaging device 800 and/or obtained from any other sources, e.g., captured by another imaging device 800. The video file can reside in a storage device, such as a storage medium 600 (shown in
(30) For purposes of presenting the video clip via a remote device 900, the video clip can be converted to a format compatible with the remote device 900, and the video clip can be transmitted to the remote device 900. In some embodiments, the ISP 320 has powerful image processing capacities and can utilize those capacities to convert the video clip. For example, if the ISP 320 includes powerful encoding and/or decoding capacities, the ISP 320 advantageously can convert the format of the video clip to the format compatible to the remote device 900 via the decoding and/or encoding capacities. The ISP 320 can also reduce a size of the video clip to facilitate a transfer of the video clip while decoding and encoding the video clip.
(31) As illustrated in
(32) The transcoded video content and/or the context information can be presented via the remote device 900. The remote device 900 can be any type of device that has an application (“APP”) installed for presenting the video content and/or the context information. Such devices can include, but are not limited to, any remote device, such as a smartphone, an iPad®, a personal digital assistant (or PDA), a notepad or any other type of computing device that can execute an APP and that can present the video content and/or the context information. The APP can provide a user with an optimized user interface that can present the user with improved operation experience. In some embodiments, the remote device 900 can present the video content and/or the context information in one of a plurality of presenting modes, including, but not limited to, a fast-forward mode and/or a fast-rewind mode.
(33) Although shown and described as transcoding only the video content for purposes of illustration only, the context information can be transcoded for reducing the size of the video clip to facilitate the transmission.
(34)
(35) At 124, the video content can be extracted from the video clip. The extracting of the video content from the video clip can be conducted while the video file is being read. Alternatively and/or additionally, the extracting of the video content can be conducted when the video clip is read into the ISP 320, in part and/or in its entirety. The extracting of the video content can be performed in any suitable manner. The extracted video content can be in an encoded format that is recognizable to the ISP 320.
(36) Although shown and described as being loaded from the storage medium 600 for purposes of illustration only, the video clip can be acquired from other sources, e.g., from another imaging device 800 for capturing the video clip.
(37)
(38) In
(39) The first video bitstream 211 can have a high bitrate. For example, a high definition 1080p video with a protocol H.264 can have a bitrate of 5000 kbps, a standard definition 480p video can have a bitrate of 2500 kbps and a low definition 360p can have a bitrate of 1200 kbps. A wireless connection can be a bottleneck for transmitting a video clip with such high bitrates. Therefore, in some embodiments, the bitrate of the video clip can be reduced while the video clip is decoded.
(40) At 154, after the video content of the video clip is being decoded to generate the first video bitstream 211, the first video bitstream 211 can be downscaled to reduce the bitrate or a data size of the first video bitstream 211 to generate a second video bitstream 212. In some embodiments, the first video bitstream 211 can be downscaled by, e.g., reducing a resolution or reducing a frame per second rate of the video content. A bitrate of the second video bitstream 212, for example, can be several times less than the bitrate of the first video bitstream 211. As an exemplary example, a reduction from high definition 1080p video to low definition 360p video can reduce the bitrate from 5000 kbps to 1200 kbps. Similarly, a reduction of frame per second rate can result in linear reduction of the bitrate of the video clip.
(41) Although shown and described as downscaling the video content after the encoding of the video content for purposes of illustration only, the downscaling step 154 can be performed at any suitable moment, e.g., while decoding step 152 is being performed.
(42) In some embodiments, the bitrate reduction can be configurable via certain transcoding parameters, such as a resolution reduction ratio and/or a reduction ratio for frame per second rate. Quality and/or continuity of a video presentation result can be adjusted via such transcoding parameters. In some other embodiments, the transcoding parameters can be adjusted when the video is presented in a fast-forward mode or a fast-rewind mode.
(43) The downscaled second video bitstream 212 can be encoded, at 156, to generate a third video bitstream 213 that can further facilitate a transmission of the video clip. The second video bitstream 213 can be encoded in accordance with a protocol that is recognizable by a video viewing APP of the remote device 900 (shown in
(44)
(45) At 131, the bitrate of the first video bitstream 211 can be reduced by lowering the resolution of the first video bitstream 211. The resolution (also known as presenting resolution) for a digital video clip is a number of distinct pixels of each frame contained in the video clip. The resolution can be set forth as a predetermined number of pixels in each dimension that can be presented in the frame, which can be represented as width by height, with units of pixels. As an exemplary example, a high definition video can have a resolution of 1024×768, meaning the width of the frame can be 1024 pixels and the height of the fame can be 768 pixels. For purposes of presenting the video clip, the resolution can be reduced to, e.g., 320×240 with certain expenses of image quality.
(46) The resolution reduction can be performed in any conventional manner. Reduction of a resolution of the first video bitstream 211 can result in a reduction of the bitrate of the first video bitstream 211. In an alternative embodiment, the resolution reduction can be adjusted with a set of resolution parameters as part of transcoding parameters. Alternatively and/or additionally, reduction rates in either and/or both directions can be adjusted in a manual or automatic manner.
(47) At 132, the bitrate of the first video bitstream 211 can be reduced by lowering the frame per second rate. The frame per second rate (also known as frame rate or frame frequency) is a frequency at which an imaging device can present consecutive frames. For example, the video clip can have an exemplary frame per second rate of 20 FPS (frame per second) to 90 FPS that can be reduced to, e.g., 10 FPS to 30 FPS.
(48) Lowering the frame per second rate can be performed in any conventional manner. Like resolution reduction, reduction of the frame per second rate of the first video bitstream 211 can result in a reduction of the bitrate of the first video bitstream 211. In an alternative embodiment, the reduction of the frame per second rate can be adjusted, as part of the transcoding parameters, with a set of frame per second parameters in a manual or automatic manner.
(49) Although shown and described as reducing the bitrate with either lowering the resolution or the frame per second rate for illustrative purposes only, the bitrate of the first video bitstream 211 can also be reduced by a combination of lowering the resolution and the frame per second rate. Alternatively and/or additionally, any other approach for reducing the bitrate of the first video bitstream 211 can be applied, e.g., any applicable compression approaches.
(50)
(51) As illustrated in
(52) In alternative embodiments, the third video bitstream 213 and the context information can be transmitted from the ISP 320 to the remote device 900 in separated communication channels. In the alternative embodiments, the third video bitstream 213 and the context information can be merged with a selected device (not shown) other than the ISP 320.
(53) The third video bitstream 213 can be transmitted in accordance with a public protocol or a proprietary protocol. Such protocols can include, but are not limited to, a Real Time Messaging Protocol (“RTMP”) protocol or a Real Time Streaming Protocol (“RTSP”) protocol, or any suitable proprietary transfer protocols.
(54)
(55) At 130, the auxiliary context bitstream 313 can be transmitted for merging with the third video bitstream 213. The auxiliary context bitstream 313 can represent context information of the video clip and can be acquired via transcoding the context information or via directly transmitting the context information. At 170, a composite bitstream 411 is created by merging any applicable bitstream tool, such as a Transport Stream Packet Editor (“TSPE”) and the like.
(56) The merging, at 170, can be triggered manually or automatically. In a configuration where an imaging device 800 is arranged on a UAV 901 (collectively shown in
(57)
(58) At 163, the first context bitstream 311 can be transcoded to create a second context bitstream 312. The first context bitstream 311 can be transcoded, in a similar manner as shown and described with reference to
(59) At 165, an auxiliary context bitstream 313 can be generated by encoding the second context bitstream 312. The auxiliary context bitstream 313 can be transmitted to a location where the auxiliary context bitstream 313 can meet the third video bitstream 213 (shown in
(60) Although shown and described as applying similar approaches to reduce the bitrate of the first context bitstream 311 for purposes of illustration only, the bitrate of the first context bitstream 311 can be reduced by other applicable approaches, e.g., stripping off unnecessary information from the context information. Although shown and described as transcoding the first context bitstream 311 for purposes of illustration only, in some embodiments, the context information can be transmitted for merging with the third video bitstream 213 directly, making transcoding the first context bitstream 311, at 163, optional.
(61)
(62) The timestamps of the third video bitstream 213 can be a sequence of characters representing the times, e.g., at which the video clip was recorded. The timestamps can be distributed evenly or unevenly along the third video bitstream 213 to ensure satisfactory matching with the auxiliary context bitstream 313. For a purpose of this disclosure, the timestamps can either be absolute timestamps or relative timestamps. Additionally and/or alternatively, the timestamps can be distributed along the auxiliary context bitstream 313 in a same manner or a different manner as the timestamps of the third video bitstream 213 to ensure satisfactory matching with the third video bitstream 213.
(63) The third video bitstream 213 and the auxiliary context bitstream 313 can be synchronized, at 172, for example, by matching the timestamps of the third video bitstream 213 with the timestamps of the auxiliary context bitstream 313. At 175, matched portions of the third video bitstream 213 and the auxiliary context bitstream 313 can be combined (or merged) to form a composite bitstream 411 in a similar manner shown and described with reference to the composite bitstream 411 of
(64) Although shown and described as using timestamps for purposes of illustration only, the third video bitstream 213 and the auxiliary context bitstream 313 can be synchronized with any other suitable form of matching marks, such as packet labels.
(65)
(66) At 180, the composite bitstream 411 can be transmitted to the remote device 900 via any transmission medium 700 (shown in
(67) The remote device 900 can run a video application (“APP”) that can present the composite bitstream 411. The APP can provide a user with an optimized user interface that can present the user with improved operation experience. In some embodiments, the remote device 900 can present the composite bitstream 411 in a fast-forward mode and/or a fast-rewind mode. The remote device 900 can be a smartphone, an iPad®, a notepad and the like. Such APPs can comprise any application including but not limited to, a BSPlayer, an MXPlayer, a DicePlayer and the like. The composite bitstream 411 can be presented to a user (not known) who can update configuration settings of an ISP 320 for transcoding, at 210. In some embodiments, the configuration settings can be updated according to operations of the remote device 900.
(68) Although shown and described as presenting the composite bitstream 411 for updating the configurations for purposes of illustration only, the configuration can be updated based only on the context information, and/or the process can be performed in an automatic manner based on the context information.
(69)
(70) The configuration for transcoding can include, but is not limited to, a resolution reduction rate and/or a reduction rate for frame per second rate. The resolution reduction rate can be adjusted, at 221, based on a resolution quality of the presented information. For example, when the resolution quality is unsatisfied, the resolution reduction rate can be decreased, and vice versa. Similarly, the reduction rate for frame per second rate can be adjusted, at 222, based on a continuity of the presented information and/or a real-time presentation operation, e.g., a fast-forward operation or a fast-rewind operation. For example, when the continuity of the presented information is unsatisfied, the reduction rate for frame per second rate can be decreased, and vice versa.
(71) The configuration can be updated via another APP installed on the remote device 900 or via another device, such as a controller that can control the imaging device 800 and/or the UAV 901 (shown in
(72)
(73) The imaging device 800 is shown as consisting of a sensor 310 and the ISP 320. The sensor 310 can be adapted for capturing images that can form the video clips. The sensor 310 can capture the images and transmit the images as raw video data 315 to the ISP 320. The ISP 320 can include an encoder 330 for encoding the raw video data 315 into bitstream 335. The encoder 330 can be separate from the ISP 320 and/or, as shown in
(74) The video clips captured by the imaging device 800 can usually be in a high resolution. Therefore, the encoder 330 can encode the video clips in accordance with a protocol capable for handling high resolution image data, such as a MPEG-4 Part 10, Advanced Video Coding (“H.264”) or any protocols under a High Efficiency Video Coding (HEVC) standard, including but not limited to, MPEG-H Part 2 and ITU-T H.265 and the like. As shown and described herein, in some embodiments, the ISP 320 can have powerful capacities in encoding and/or decoding video clips.
(75) The encoded bitstream 335 can be stored into the storage medium 600. The storage medium 600 can be attached to the imaging device 800 or otherwise associated with the imaging device 800 in any manner. The storage medium 600 can be of any type of storage media that can be accessed by the imaging device 800, including but not limited to, a secure digital card (“SD”), an embedded multimedia card (“eMMC”), a solid state drive (“SSD”), a USB disk, or any other type of mobile disk, and the like. The storage medium 600 can have enough storage volume and/or enough speed for storing the video clips and for accessing the video clips by the imaging device 800.
(76) Although shown and described as comprising an encoder for purposes of illustration only, the encoder 330 can comprise a software module and/or a software program running in the ISP 320, or can include hardware, software, firmware or a combination thereof.
(77)
(78) The video clip, being either captured by the same imaging device 800 or stored in the storage medium 600 from another source, can be retrieved by the ISP 320 associated with the imaging device 800 as encoded video content 702. The ISP 320 can include a decoder 420 that can decode the content of the video clip. In other words, the decoder 420 can be at least partially integrated with the ISP 320 of the imaging device 800. The decoded content can be passed, as a decoded video bitstream 422, to another processor 430 for further processing. The processor 430 that can be a dedicated image processor and is illustrated as being external from the imaging device 800. The imaging device 800 can include a special port (not shown) for passing the decoded content to the processor 430. The special port can include, but is not limited to, a High Definition Multimedia Interface (“HDMI”) port or a BT.656 (also known as ITU656) interface and the like.
(79) The processor 430 can be associated with an encoder 435 that can encodes the decoded content to generate an encoded video bitstream 436. The encoder 435 can be separate from the processor 430 and/or, as shown in
(80) The encoded video bitstream 436 can be transmitted to the remote device 900 via a transmission medium 700. The transmission medium 700 can be any type of wired and/or wireless communication connection or any type of network connections. The APP installed on the remote device 900 can be used to present the video clip via the encoded video bitstream 436 while the encoded video bitstream 436 is being transmitted. In one embodiment, the remote device 900 can include a video display (not shown) for presenting the video clip.
(81) Although shown and described as being separate from the imaging device 800 for purposes of illustration only, the encoder 435 can be at least partially integrated with the ISP 320 associated with the imaging device 800.
(82)
(83) The video clip can comprise one or more images captured by the imaging device 800 and can contain both video content and context information, such as audio content and/or other metadata. The metadata can include, but is not limited to, automatically collected video metadata, such as aperture, shutter speed, location coordinates, GPS coordinates and the like, and manual written information, such as transcripts of dialogue or conversations and other text descriptions.
(84) In the manner described above with reference to the storage medium 600 of
(85) When transmitting a selected video clip, the ISP 320 can retrieve the selected video clip from the storage medium 600 by read 702. The video clip can be read into the ISP 320 as a document, partially or in its entirety, or as a bitstream. The ISP 320 can execute a software program, such as an extract module 810, to separate video content of the video clip from context information of the video clip. The video content of the video clip can be in a form of video content 811 that can be passed to a decoding module 815. The video content 811 can be in a format of a bitstream or a file.
(86) The decoding module 815 can be a software program that is executed by the ISP 320 or can include hardware, software, firmware or a combination thereof. The decoding module 815 can decode the video content 811 according to a predetermined protocol. In some embodiments, the predetermined protocol can be the same protocol that was used to encode the video content. The decoding module 815 can decode the video content 811 to generate a first video bitstream 211. The first video bitstream 211 can be passed to a downscaling module 816.
(87) The downscaling module 816 can be a software program that is executed by the ISP 320 or can include hardware, software, firmware or a combination thereof. The downscaling module 816 can downscale the first video bitstream 211 to generate a second video bitstream 212. The downscaling module 816 can downscale the first video bitstream 211 by reducing a bitrate and/or bit per second rate of the first video bitstream 211. The bitrate of the first video bitstream 211 can be reduced by either reducing a resolution of the first video bitstream 211 and/or by reducing a sampling rate or a frame per second rate of the first video bitstream 211. The second video bitstream 212 can be passed to an encoding module 817.
(88) The encoding module 817 can be a software program that is executed by the ISP 320 or can include hardware, software, firmware or a combination thereof. The encoding module 817 can encode the second video bitstream 212 to generate a third video bitstream 213. The encoding module 817 can encode the second video bitstream 212 in accordance with a protocol that is compatible to a remote device 900 and/or an APP running on the remote device 900. Such protocols can include, but are not limited to, a 3rd Generation Partnership Project (“3GPP”), Motion Picture Experts Group (“MPEG4”), Real Time Streaming Protocol (“RTSP”), Flash Lite, Mobiclip and the like. The third video bitstream 213 can have a much lower bit per second rate than the first video bitstream 211 due to the downscaling. The third video bitstream 213 can be passed to a merger 820.
(89) The merger 820 can be a software program that is executed by the ISP 320 or can include hardware, software, firmware or a combination thereof. The merger 820 can merge the third video bitstream 213 with context information in a form of an auxiliary context bitstream 313, of the video clip to generate a composite bitstream 411. When merging the third video bitstream 213 with the auxiliary context bitstream 313, the merger 820 can synchronize the third video bitstream 213 with the auxiliary context bitstream 313 by matching timestamps distributed along both bitstreams 213, 313 in a manner shown and described with reference to
(90) Although shown and described as using timestamps for purposes of illustration only, the third video bitstream 213 and the auxiliary context bitstream 313 can be synchronized with any other form of matching marks, such as a packet label. Although shown and described as being a software program being executed by the ISP 320 for purposes of illustration only, the merger 820 can be a separate processing module that is independent from the imaging device 800.
(91) The composite bitstream 411 can be transmitted to the remote device 900 via a transmission medium 700. The transmission medium 700 can be any type of wired and/or wireless communication connection, or any type of network connections. The APP installed on the remote device 900 can be used to present the video clip by presenting the composite bitstream 411 while the encoded video bitstream 411 is being transmitted.
(92) Although shown and described as presenting the composite bitstream 411 with the APP running on the remote device 900 for purposes of illustration only, the composite bitstream 411 can be presented with any suitable viewers, such as any applicable video viewers running with any hardware, firmware, or a combination of hardware and firmware.
(93)
(94) The context content 821 can be passed to the decoding module 833. The decoding module 833 can be a software program that is executed by the ISP 320 or can include hardware, software, firmware or a combination thereof. The decoding module 833 can decode the context content 821 according to a predetermined protocol. In some embodiments, the predetermined protocol can be the same with the protocol used to encode the context information. The decoding module 833 can decode the context content 821 to generate first context bitstream 311. The first context bitstream 311 can be passed to the downscaling module 835.
(95) The downscaling module 835 can be a software program that is executed by the ISP 320 or can include hardware, software, firmware or a combination thereof. The downscaling module 835 can downscale the first context bitstream 311 to generate a second context bitstream 312. The downscaling module 835 can downscale the first context bitstream 311, for example, by reducing a bit per second rate of the first context bitstream 311. The second context bitstream 312 can be passed to the encoding module 837.
(96) The encoding module 837 can be a software program that is executed by the ISP 320 or can include hardware, software, firmware or a combination thereof. The encoding module 837 can encode the second context bitstream 312 to generate an auxiliary context bitstream 313. The encoding module 837 can encode the second context bitstream 312, for example, in accordance with a protocol that is compatible to combine with the third video bitstream 213 (shown in
(97)
(98) In some embodiments, video content can comprise more volume than the context information. Therefore, a size reduction of the video content can be very helpful to facilitate a transmission via a communication connection, particularly a wireless communication connection. In contrast, a size reduction of the context information can be unnecessary. For that purpose, a transcoding operation of the context information processing module 823 can be unnecessary and thus can be bypassed via the bypass 831.
(99)
(100) In
(101) The context information of the video clip can be contained as a file that is separated from the file containing the video content and that is stored in the storage medium 600. The ISP 320 can read the context information directly from the storage medium 600 as context content 821. The ISP 320 can execute a context information processing module 823 to process the context content 821 and to generate an auxiliary context bitstream 313 that can be transmitted to the merger 820 for merging with the third video bitstream 213. The context information processing module 823 can either transcode the context content 821 in a manner shown and described with reference to
(102) Although shown and described as being contained in a file stored in the storage medium 600 for purposes of illustration only, the context information can be stored in a separate storage medium or captured from a separate source, e.g., audio content from a microphone (not shown) or caption information from a keyboard (not shown), etc.
(103) As shown in
(104) If the merger 820 is separated from the remote device 900 as shown in
(105) Although shown and described as both the video content and the context information are transmitted via the ISP 320 to the merger 820 for purposes of illustration only, the context information can be transmitted to the merger 820 from other sources, e.g., from a microphone or a keyboard etc.
(106)
(107) The mobile platform 901 can comprise a plurality of propulsion units, e.g., propellers 911, for providing lifting power to the UAV. The mobile platform 901 can have a fuselage 910 that can couple with the imaging device 800. As shown and described herein, a storage medium 600 (shown in
(108) As shown and described herein, a merger 820 can be either a program executed by the ISP 320 of the imaging device 800 or be executed by a processor 430 (shown in
(109) The composite bitstream 411 can be presented via the remote device 900. The presented content, particularly the context information, can be used as basis to adjust configuration settings of the ISP 320, such as parameters used for transcoding the video content and/or the context information. The adjustment can be performed manually by an operator or be performed automatically by the remote device 900. The remote device 900 is shown and described for purposes of illustration only. In some embodiments, the composite bitstream 411 can be presented and/or the configuration settings can be adjusted via any suitable devices, such as a notepad, a laptop computer, a desktop computer and the like.
(110) Although shown and described as using a merger to merge the two bitstreams 814, 822 for purposes of illustration only, the remote device 900 can adjust the configurations for transcoding based solely on the context information.
(111) The described embodiments are susceptible to various modifications and alternative forms, and specific examples thereof have been shown by way of example in the drawings and are herein described in detail. It should be understood, however, that the described embodiments are not to be limited to the particular forms or methods disclosed, but to the contrary, the present disclosure is to cover all modifications, equivalents, and alternatives.