Embedded system for video processing with hardware means

10237603 · 2019-03-19

Assignee

Inventors

Cpc classification

International classification

Abstract

A process is provided for video processing and distribution of video data within a betting agency. The process includes receiving data streams of dynamic live betting odds and game statistics and receiving parallel TV live broadcasts of sports events as video data streams. The process further includes compressing the data streams and the video data streams yielding combined data streams, outputting each of the combined data streams via a protocol based network, receiving a combined data stream from the protocol based network pertaining to the betting on a supply of receiver servers, decompressing the combined data stream, and outputting the combined data stream displayed in any order on one or several monitors in real time. The combined data stream being continuous from allocated embedded systems, which operate as a client, wherein the clients individually provide the combined data stream from the supply of the receiver servers.

Claims

1. A process for video processing and distribution of video data within a betting agency comprising: receiving of data streams pertaining to betting, the data streams being dynamic live betting odds and game statistics; receiving parallel TV live broadcast of several sports events as video data streams; compressing of the data streams and the video data streams yielding in each case a combined data stream; outputting the combined data stream via a protocol based network; receiving the combined data stream from the protocol based network on a supply of receiver servers; decompressing of the at least one combined data stream; outputting the combined data stream displayed in any order on one or several monitors in real time, the at least one combined data stream being continuous from allocated embedded systems, which are operating as a client, wherein the clients individually provide the at least one combined data stream from the supply of the receiver servers to adjust the graphical appearance across the one or several monitors in real time to include 2 to 6 video data streams with selectable size and positioning that are displayed in at least one of the modes of: (a) one of the 2 to 6 simultaneous video data streams is displayed cross-faded over a second of 2 to 6 simultaneous video data streams or (b) four different data streams of the 2 to 6 simultaneous video data streams are displayed simultaneously, or (c) three streams of the 2 to 6 simultaneous video data streams are displayable simultaneously on a 90 rotated screen, or (d) six streams of the 2 to 6 simultaneous video data streams are displayable on one monitor, the at least one mode being selectable via a graphical user interface to the several monitors not only with respect to their adjustment, but also with respect to the contents to be displayed, by a respective allocation of the 2 to 6 simultaneous video data streams wherein scaling and/or rotating and/or merging and/or cross-fading of more than one decompressed image and audio data stream occurs; outputting of one or more combined data streams occurs on at least two monitors arranged as a video wall; the compressing and the decompressing of the continuous combined data stream generates a continuous image- and audio-data stream with at least 60 frames per second and a resolution of at least 19201080 pixels without data jam; intermediate storing and repeated output of the at least one combined data stream occurs.

2. The process according to claim 1 further comprising accepting more than one combined data stream via the protocol based network.

3. The process according to claim 1 further comprising displaying on a single monitor more than one decompressed image and audio data stream.

4. The process according to claim 1 wherein the compressing and the decompressing of the continuous combined data stream are conducted with an individual activation code.

5. The process according to claim 4 further comprising emitting an activation code warning signal or an activation code approval signal, which conduct a monitoring and comparison of the individual activation code.

6. The process according to claim 1 wherein the compressing is carried out according to the H.264 or the H.265 standard.

7. The process according to claim 1 further comprising encrypting and decrypting the at least one combined data stream.

8. The process according to claim 1 further comprising generating an output of control signals for remote control of the one or several monitors.

9. The process according to claim 1 further comprising interactively controlling of the at least one combined data stream by an observer.

10. The process according to claim 1 wherein the compressing and the decompressing of the continuous combined data stream are conducted with a Field Programmable Gate Array (FGPA).

11. The process according to claim 10 wherein the FPGA is individually encrypted.

Description

(1) Hereafter, the invention is illustrated further based on examples referring to the respective Figures.

(2) It is shown in

(3) FIG. 1 the schematic set-up of an embodiments of the video distribution system according to the present invention for application in a betting agency;

(4) FIG. 2 the set-up of a simple embodiment of the video distribution system according to the present invention;

(5) FIG. 3 a more complex embodiment of the video distribution system according to the present invention;

(6) FIG. 4 the schematic set-up of an embedded system according to the present invention;

(7) FIG. 5 the schematic set-up of an embedded system according to the present invention with a server functionality;

(8) FIG. 6 the schematic set-up of an embedded system according to the present invention with a client functionality;

(9) FIG. 7 an illustration of an embedded system according to the present invention as a module;

(10) FIG. 8 a schematic representation of the video data processing of an embedded system according to the present invention running as a server;

(11) FIG. 9 a schematic representation of the video data processing of an embedded system according to the present invention operating as a client; and

(12) FIGS. 10A-10K a selection of display modes attainable with an embedded system according to the present invention operating as a client; and

(13) FIG. 11 a software-technological set-up of a branch; and

(14) FIG. 12 a schematic representation of the network security module; and

(15) FIG. 13 a schematic representation of the driver decoupling of a VideoWizard module in a PC operating system environment; and

(16) FIG. 14 an application of the system according to the present invention as a roulette transmission within a casino; and

(17) FIG. 15 an application of the system as a live game transmission outside of a casino; and

(18) FIG. 16 a schematic representation of the security of the network transmission; and

(19) FIG. 17 a schematic overview of the communication structure of the software units, and

(20) FIG. 18 a schematic representation of a user interface of a display device; and

(21) FIG. 19 a schematic partial representation of a single unit XIX according to FIG. 19 in an alternative embodiment; and

(22) FIG. 20 a schematic partial representation of a single unit XIX according to FIG. 19 in a further alternative embodiment.

(23) The invention comprises a modern video distribution system on the basis of network based multiple H.264 data streams. Thereby, different video sources are converted into H.265 data streams by embedded systems operating as a server and decoded by embedded systems operating as a client and displayed on monitors.

(24) The application of the video distribution system is intended for example for betting agencies, where the image sources are normally provided by satellite receivers and are individually distributed to the available TV monitors by shop masters/users.

(25) The typical set-up of a video distribution system according to the present invention in a betting agency is depicted in FIG. 1 and consists of the four functional units receiving, distribution, display, and control. The receiving function is realised by a variable number of satellite receivers 100, which possess allocated embedded systems 110, which are operating as a server. The encoded video streams of the satellite receivers 100 are distributed via a network 120 (with a switch/distributer, VLAN). The display occurs via a variable number of monitors 130, which receive their images from allocated embedded systems 140, which are operating as a client, wherein the clients 140 individually provide from the supply of the receiver servers 110 selected image data streams.

(26) Thereby, the selection of the image data streams occurs at local control PCs 150, at which the shop master/user can dynamically create the allocation between satellite program and monitor 130 (control). To this end, the TV monitors 130 and the satellite receivers 100 are visualised graphically on a display device 155, which is allocated to the control PC 150, as shown for example in FIG. 18.

(27) Each embedded system 110 or 140, which is connected to a Sat-receiver 100 or TV monitor 130, additionally possesses an infrared control function, which allows to remotely operating the allocated device. Every satellite receiver 100 is also optionally connected to a network connection, via which the satellite channel listings can be updated and changed.

(28) The control of the embedded systems 110 or 140 and, therefore, the allocated devices 100 or 130 occurs physically via network 120 and structurally via the operator PC software. If necessary, several such operator stations 150 allow the control of the devices 100, 130.

(29) According to a preferred embodiment, the system according to the present invention constitutes Flexible hardware based multi-function video gaming and visualization system for digital television and digital media content, combination of a freely selectable global hardware based video stream network with game visualisations and interactive live content display.

(30) According to a preferred embodiment, the system according to the present invention is arbitrary scalable by modular set-up (see also FIG. 2 or FIG. 3). A client can access several hundred server streams. The only limitation is the capacity of the used network.

(31) According to a preferred embodiment, a division in clients and servers occurs. Servers receive image data and feed them into the network. Clients receive the network streams and display them in very different manners. Interactive control via network (for example PC, IPad, network compatible device). Mixed use as client and server simultaneously is possible.

(32) According to a preferred embodiment, the system according to the present invention processes different sources, which can be fed via satellite or network locally or remotely (globally worldwide!) (live cams, digital and analogue TV sources, hard disk, magnetic disk data, media player etc.).

(33) According to a preferred embodiment, the system according to the present invention is designed for receiving of the image source and real time encoding according to H.264-methods in hardware (data reduction for sending via network).

(34) According to a preferred embodiment, the system according to the present invention is designed of several H.264 source data streams per client module.

(35) According to a preferred embodiment, the system according to the present invention is able to provide different visualisation modes by the client (see also FIG. 10): a) 1 stream; b) 2 simultaneous streams; c) 4 simultaneous streams per TV; d) 3 rotated streams; e) 6 streams etc.; video wall mode, wherein one or several streams can be distributed onto a wall consisting of many monitors; g) cross-fading mode of different streams (blue box); h) Fading/cross-fading of live streams in game application; i) interactive control of the mode by the user; j) replay function of video streams on demand.

(36) According to a preferred embodiment, encryption and decryption of data streams occurs in real time in hardware.

(37) According to a preferred embodiment, the system provides an interactive possibility to show video contents anew or to change data streams and display modes interactively.

(38) According to a preferred embodiment, transfer via network and distribution via switch system (i.e. gigabit, Ethernet, UDP protocol etc.) occurs.

(39) According to a preferred embodiment, the system is set-up using combined FPGA and DSP/special chip technology (real time hardware).

(40) The basic concept of the video distribution system and the embedded systems for video processing is depicted schematically in FIG. 2 and is based on two functional units in combination: the H.264 video functions for compression and decompression and the functions for network security with hardware firewall and real time data encryption.

(41) H.264 Video Unit

(42) The video unit of the embedded system consists of one or several H.264 encoders, which can compress image data in real time using the state of the art H.264 method, and one or several H.264-decoders, which can decode and display several compressed image data streams simultaneously. Image and sound data are received via a camera system or a HDMI/PAL source with up to 60 frames per second and transmitted to an embedded system, which is operating as a server, and which processes, records, and compresses these data to a H.264 data format. The image data streams can then either be transferred uncompressed to a PC or a monitor or can be transmitted via a network connection. This is depicted schematically in FIG. 9.

(43) GigE vision compatible cameras 210, satellite receivers with HDMI output, DVD players or other devices adapted for HDMI output serve as image data source.

(44) The embedded system additionally possesses the opposite functionality (client). It is possible to feed H.264 as well as MPEG2 data streams, which are decompressed in this embedded system in real time and can either be recoded or directly outputted via HDMI. This is depicted schematically in FIG. 10. Thereby, it is possible to play videos in progressive and interlaced data format (interlaced scanning) and to deinterlace them as required.

(45) FIGS. 10A and 10C show a single video data stream, which is displayed scaled and positioned on a display (LCD, LED, OLED, TV etc. monitor) and underlayed with a background colour.

(46) FIGS. 10B and 10D show two single video data streams, which are scaled and positioned and underlayed with a background colour. One of the two videos is scaled smaller and is in the foreground.

(47) FIG. 10E shows two separate video data streams, which are both scaled and positioned and underplayed with a background colour. One of the two videos is scaled smaller and shown in a blending with the other video (blending).

(48) FIG. 10F shows four separate video data streams, all four of which are scaled and in a fixed position (non-overlapping).

(49) FIG. 10G shows three separate video data streams, all three of which are scaled and rotated by 90, and in a fixed position (non-overlapping).

(50) FIG. 10H shows six separate video data streams, all six of which are scaled and have a fixed position (non-overlapping).

(51) FIG. 10I shows a video data stream on a Video Wall realised with 4 monitors, wherein each of the 4 monitors displays a respective section of the complete image to be shown.

(52) FIG. 10J shows three data streams, wherein video data streams as well as the HTML browser data streams are displayed mixed in different windows. The windows are to be arranged scalable, positionable, rotating (rotatable), and overlapping with each other.

(53) FIG. 10K shows two video data streams in different windows, wherein a larger window displays an image, i.e. from a satellite receiver, and a smaller window shows an information string, for example a so-called news ticker.

(54) The whole system is constructed in such a way that only a single type of embedded system (VideoWizard module) is required physically. This embedded system is designed for security with respect to the data streams, but also with respect to product piracy.

(55) Thereby, the video distribution system is devised in such a way that systems with over 100 servers and clients can be realised in a single gigabit network stream (this is dependent on the resolution of the data streams). Alternatively the data stream can be cascaded in a tree like fashion. A system with many components is schematically depicted in FIG. 3.

(56) The second simultaneous functionality of the embedded system consists of a network security functionality, which massively increases data transmission security via a hardware implemented firewall and AES real time data encryption. Thereby, the core functions in the present case are a 256 bit AES encryption of the network traffic as well as a hardware firewall with port activation and logging function.

(57) General Set-Up of the Embedded System:

(58) Conceptually, the embedded system 400 consists of a single component or a module (VideoWizard hardware), which is supplemented by a motherboard 710, which contains the additional functions for standalone operation.

(59) Functionally, the module is able to realise the server functions, the client functions, and the network security functions. These are described in more detail below.

(60) By setup, the embedded system (see FIG. 5 and FIG. 8) consists of a combination of FPGA and embedded processor, GPU and special hardware technology (all three functions in the utilised DSP). Thereby, both units share different functions. The functionality of the FPGA 410 is necessary for real time computing and processing of the GigE data of the camera 210, the processing of HDMI input data, and particularly PC connection and security aspects, the DSP realises the DVI output and scaling as well as the H.264 data processing as an encoder/decoder.

(61) Advantageously, a driver decoupling of the VideoWizard modules (for example for different operating systems such as Windows XP Embedded and Linux) is possible by means of the embedded system, wherein the VideoWizard module, which is designed as a PC plugin card, is faded into the operating system environment as a network card. Thereby, the connection is a network driver, which allows the use of all standardised network programmes of the operation system. The second access interface is realised via a low level user library, which provides the necessary control, image generating and DMA functions. The FPGA can provide network functions in combination with the TI co-processor. For example, an image data and control driver 133 is provided in conjunction with a so-called application-programming interface API 134 as well as a separate network driver 135 (compare FIG. 13 and FIG. 12).

(62) The display device 155, which is allocated to the control PC 150 according to FIG. 18, depicts a graphical user interface for visualisation of the available monitors 156, and/or windows 157, and/or a remote control unit 158. By means of selection menus 154 the monitors 156 can be switched on or off, the data streams to be displayed on the on monitors 156 or in the windows 157, which represent screen sections, can be selected, a display format can be selected, and the display size can be defined. The data streams to be selected individually may comprise video data, information data, for example in the form of a so-called news ticker, browser URL, that is Internet pages, and/or image data. By means of the graphical user interface it is also possible to save and/or load standard settings and/or user settings.

(63) According to FIG. 19 a large-scale image A, which is visualised on the display device 155, is displayed on several monitors, which are assembled to a video wall 159, which is partially overlapping with two windows 157, which display the images B and C, wherein the window 157 with the image C partially overlaps with the window 157 with the image B. In order to make adjustments concerning the window 157 with the image C, the window 157 is selected by means of an input device, for example a mouse, which is allocated to the control PC 150 and the display device 155, as depicted by the grey background, and the selection menu 154, which is allocated to the window 157, with different menu items 153, is expanded visually perceptible. After the selection of the menu item 153, depicted in grey, the allocated remote control unit 158 appears, with whose switches adjustments are to be made, or sent to the respective monitors or clients by means of especially adjusted commands. Obviously, the display device 155 may be designed as a touch screen, whereby the operation by a user can be very easily realised.

(64) The representation of configurations, particularly the visualised assembly and sizes of the monitors 156 or windows 157 on the display device 155, corresponds to the actual assembly of monitors or is defined by means of a specified image. Previously saved configurations can be selected under a respective menu item 153, which can also be displayed in a title bar on the display device 155. By means of the input device, a selection menu with the representation of all available clients can be displayed.

(65) The following types of contents can be defined in the visualisation of the display device 155: image: An image defined by the user is displayed. Images stored on an image server in a specific directory can be selected. info display: All available channels in the info display system can be selected. TV channels: Modules recognised as a server can be selected as TV channels. Browser URL: Any URL can be selected in this menu, however browser support on the modules can be restricted, i.e. no flash contents displayed.

(66) With the menu item 153 distribution in XML format defined layouts may be allocated, which are available dynamically, meaning that possible layouts are only displayed for the present monitor or group of monitors.

(67) Description of a XML Configuration:

(68) Layout <Layout id=xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx xmlns=urn:admirababsolutevision:light:layout:v1>

(69) xxx . . . corresponds to a GUID, which is unique in the directory. The file name of the configuration should also comply with this GUID.

(70) GUID can be created by a GUID generator. i.e.: http://www.guidgenerator.com/online-guid-generator.aspx

(71) GUID (or UUID) is an acronym for Globally Unique Identifier (or Universal Unique Identifier). It can be a 128-Bit integer for identifying resources.

(72) The name entry is used by the system, therefore a name, which describes the distribution, should be entered.

(73) <Name>Split 2/PictureInPicture</Name>

(74) Under translations the name of the distribution is entered, which is displayed in the GUI. The translation according to the Windows system language is displayed. If there is no entry for the selected system language, the translation en is displayed.

(75) <Translations>

(76) <Translation language=de>2 Anzeigen/Bild in Bild</Translation>

(77) <Translation language=en>2 Views/Picture in Picture</Translation>

(78) </Translations>

(79) <Screens> describes how the monitors or screens are arranged in the coordinate system, their resolution, and the rotation.

(80) <Screens>

(81) <Screen x=0 y=0 width=1920 height=1080 rotation=0></Screen>

(82) <Screen x=0 y=1080 width=1920 height=1080 rotation=0></Screen>

(83) </Screens>

(84) Thereby, all monitors or screens, which are comprised in the distribution, should be entered. x=0 y=0 describe the position of the monitor/screen in the coordinate system of the video wall. Thereby, for instance the upper left point of the display can be used as a reference for positioning.

(85) For example width=1920 height=1080 describe the size of the display. rotation=0 describes the orientation, namely rotation of the display. For instance, only the values 0, 90, 180 und 270 can be supported. Currently, only 0 und 90 make sense, because only horizontal and vertical orientations can be selected in device settings.

(86) <Contents> defines the viewing areas in the coordinate system. These areas can span several monitors.

(87) <Contents>

(88) <Content x=0 y=0 width=1920 height=1080></Content>

(89) <Content x=200 y=200 width=640 height=480></Content>

(90) </Contents>

(91) x=0 y=0 correspond to the starting coordinates of the content area in the coordinate system. For example, the upper left point can be used as a reference for positioning

(92) width=1920 height=1080 describe the size of the viewing area. If a content, which does not correspond to the indicated aspect ratio, is allocated to the content window, the content is adapted to the content area and scaled.

(93) Example according to FIG. 20:

(94) A video wall 159 consists of six FullHD TVs (32) and three viewing areas E, F, G:

(95) Viewing area E: Content over the whole wall of 32 TVs

(96) Viewing area F: Content over a partial area of 4 TVs. The display overlaps with viewing area E

(97) Viewing area G: Content over a partial area of 4 TVs. The display overlaps with viewing area E and a part of viewing area F.

(98) An accordingly constructed XML configuration file can for instance be transmitted to the client modules via an API interface.

(99) Server Functionality

(100) The server functionality of the embedded system is depicted schematically in FIG. 5 and partially in FIG. 12 (left of the network) and consists of the following detail functions:

(101) receiving of HDMI data streams up to 19201080 p60 (progressive, 60 Hz) or camera data streams; compression of the image- and audio-data streams in H.264 format; packaging of control and audio data into the H.264 data stream; standalone capability with access (via server PC 510) via Gigabit-Ethernet to internal registers of the embedded system; individual FPGA security against product piracy; AES encryption of the data stream; sending the compressed data via GigE network

(102) Client Functionality

(103) The client functionality of the embedded system is depicted schematically in FIG. 6 as well as partially in FIG. 12 (right of the network) and consists of the following detail functions:

(104) receiving of the compressed data or a data stream, respectively, via GigE network; AES decryption of the data stream; decompression of the image- and audio-data from the H.264 format or extraction of the control data from the data stream, respectively; decompression of the MPEG2 image data streams where applicable; deinterlace function with high quality for realisation of progressive image material; image scaling; restoration of an image data stream with 60 frames per second from the H.264 data streams; standalone capability with access (via client PC 610) via Gigabit-Ethernet to internal registers of the embedded system; individual FPGA security against product piracy; output of video audio and control data;

(105) The embedded system (the VideoWizard) can preferably be utilised in combined modes. For example, a mix of client and server (with a selection of partial functions) or a recoding of several input data stream to new output streams is possible.

(106) Network Function (Firewall and Data Encryption)

(107) The network functionality of the embedded system consists of a hardware implemented firewall and AES real time data encryption, whereby the data transmission security is massively increased. Thereby, the core functions in the present case are a 256 bit AES encryption of the network traffic and a hardware firewall 131 with port activation and logging function. In detail the FPGA 410 predominantly realises the network functions (FIG. 13). Thereby, the FPGA contains a 256 bit AES real time encryption unit 132, which encrypts or decrypts, respectively, the incoming and outgoing network data. It is possible to choose between different H.264 data streams, which are transmitted into/out of the embedded system via network. The firewall 131 allows filtering of IP, port, and Mac addresses via a rule table. An adjustable logging function of the firewall 131 allows recording of the valid/invalid access attempts onto the embedded system. Optionally unlocked ports/IPs/Mac addresses can be blocked automatically after a certain amount of failed attempts or an alarm can be triggered if too many unallowed accesses to the embedded system occur per second. Regarding this matter, failed attempt monitoring means, which monitor the occurrence of failed attempts exceeding a defined threshold, are provided in the system.

(108) Security Against Product Piracy

(109) By means of an individual hardware serial number saved in the embedded system, it can be ensured that the embedded system can only be operated with an individual activation code. Therefore, a copy of the module for duplication is pointless. In the system, activation code monitoring means 161 can be provided, which are able to perform a monitoring or comparison of activation codes with a database and, if necessary, can emit an activation code warning signal if no match of the activation code or codes occurs or can emit an activation code approval signal if a match of the activation code or codes occurs.

(110) Security During Image and Data Transmission

(111) Optionally, the image data streams can be secured via the network with an AES encryption method, whereby the data is protected from an attacker in the transmission path. This concerns data transmission via Internet as well as internal transmission, for example in a casino.

(112) High stability and safeguard against failure during image data transmission and display. By the utilisation of two independent real time capable computing components a high operating safety with high output reserves is realised.

(113) Stability by Independence

(114) By means of the independence of the embedded system from a PC control, the stability of the overall application is increased. In case of a failure of a part of the overall system, the embedded systems continue to function independently.

(115) Network Security by a Hardware Firewall

(116) By use of a hardware firewall the unauthorised access from outside via not permitted ports, not activated IP addresses, or not authorised Mac addresses can be prevented. Moreover, the logging allows a recording of permitted and non-permitted accesses. By utilisation of a hardware firewall, the firewall cannot let accidentally, in case of capacity overload, not permitted packages pass, as is the case for software-implemented firewalls.

(117) Numerous modifications and developments of the described embodiments are realisable. For instance, an embedded system according to the present invention can also be used without a network, wherein it provides the server and client functionality simultaneously. Furthermore, different coding and encryption standards or methods can be utilised, and also according to the desired area of application, different interfaces, network protocols and network architectures can be used without changing the essence of the invention.

Glossary

(118) Image data stream, audio data stream, video data stream, combined data stream We consider these terms to mean data streams with a respective channel of the respective content (images, audio, video). A combined data stream contains image as well as audio data. Therefore, a video data stream can be an image data stream or a combined data stream.

(119) Client/Server

(120) We consider the terms client and server to mean so-called streaming clients or streaming servers. The term streaming client designates a special client for streaming media, which can be either software or hardware. Typical streaming clients support special streaming protocols such as RTP, RTCP, and/or RSVP. The term streaming server designates a decided server for distribution of streaming media data via a network. Typical streaming servers support special streaming protocols such as RTP, RTSP, RTCP, and RSVP (auxiliary protocol for QoS-method IntServ). Thereby, the term streaming media designates from a computer network received and simultaneously played audio and video data. The process of data transmission itself is termed streaming, and transmitted (streamed) programmes are termed live stream or shortly stream.
(According to http://de.wikipedia.org/wiki/Streaming-Client,
http://de.wikipedia.org/wiki/Streaming-Server and
http://de.wikipedia.org/wiki/Streaming_Media)

(121) Embedded System

(122) The term embedded system (Eng. embedded system) describes an electronic data processor or computer, which is embedded (embedded) into a technical context. Thereby, the data processor assumes either monitoring, control, or regulatory functions or is responsible for a form of data or signal processing, for example during encryption or decryption, encoding or decoding, or filtering.
(Cited according to http://de.wikipedia.org/wiki/Eingebettetes_System)

(123) Hardware Means

(124) We consider the term hardware means to mean electronic components, which are optimised for certain purposes and execute their function mainly in real time.

(125) Typically, DSP, programmable hardware such as FPGAs, special ASICs (application-specific integrated circuit), or combinations thereof are used for such purposes. These can also contain embedded microprocessors (i.e. ARM processors). In the context of this application the term hardware means does not comprise multi-purpose microprocessors, which are not optimised for the purposes named herein. Multi-purpose microprocessors (i.e. by Intel or AMD) are typically found in personal computer as central processing units.

(126) DSP

(127) A digital signal processor or DSP serves the purpose of continuous processing of digital signals (i.e. audio or video signals). For processing of analogue signals, the DSP is used in conjunction with analogue-digital converters and digital-analogue converters. Compared to multi-purpose microprocessors DSPs contain a processor, which is speed-optimised for frequently required mathematical operations. Some DSPs already contain the necessary A/D and D/A converters at their input and output.
(Cited according to http://de.wikipedia.org/wiki/Digitaler_Signalprozessor)

(128) A field programmable gate array or FPGA is an integrated circuit, into which a logic circuit can be programmed.

(129) However, the term programming in this context is to be distinguished from the term describing the design of software for a conventional processor: In a FPGA circuit structures are generated by means of hardware description languages or in the form of wiring diagrams and these data are transmitted into the component for configuration. Thereby, certain switch positions are activated or deactivated, resulting in a concretely implemented digital circuit. Therefore, instead of the term programming the term configuration of a FPGA is also used. By the specific configuration of internal structures different circuits can be realised in a FPGA, culminating in highly complex structures, such as i.e. microprocessors.

(130) The configuration typically takes place once before each use, whereby the FPGA is configured for a specific function, which it loses again upon switching off of the operating voltage. Therefore, a non-volatile memory, which stores the configuration, whose content on itself is also updatable, is allocated to the FPGA in most cases.

(131) (according to http://de.wikipedia.org/wiki/Fpga)

(132) Ethernet, GigE

(133) Ethernet is a technology, which specifies software (protocols etc.) and hardware (cables, splitters, network cards etc.) for tethered data networks. It facilitates data exchange in the form of data packages between the devices connected to a local network (LAN).

(134) Firewall

(135) A firewall is a security system, which protects a network or a single computer from unwanted network accesses. The firewall serves to restrict the network access based on sender or destination address and used services. It monitors data traffic and decides according to specified rules, where certain network packages can pass through or not. The simple filtering of data packages according to network addresses is the basic function of all firewalls.

(136) HD

(137) This abbreviation stands for high definition, (Eng.) wherein in our case image resolutions of 1280720 or 19201080 pixels are meant, wherein the higher resolution is also termed full HD.

(138) Video resolution comprises the same parameters as image resolution (lines and columns or pixel number, aspect ratio) and extends these by the temporal aspect of frame rate. Thereby, it has to be differentiated between the repetition of partial (mostly half images with interlaced scanning, interlaced), and full images (frames, progressive scan). Common frame repetition rates are 24, 25, 50, or 60 Hz. In the HDTV area 720p and 1080i are common. One speaks of full HD from 1080p25, which means 19201080 pixels, progressive, 25 Hz. (According to http://de.wikipedia.org/wiki/Videoauflsung, http://de.wikipedia.org/wiki/Full_HD and http://de.wikipedia.org/wiki/High_Definition_Television)

(139) Video

(140) We consider video to mean moved images or image streams, with or without sound.

(141) Interlaced Scanning, Progressive Scan

(142) Interlaced scanning serves to reduce flicker in television engineering. It was developed with the intention to display signals flicker free with a minimal bandwidth. Thereby, a complete image (frame) is constructed from two different half images.

(143) The progressive scan (Eng. progressive scan) designates a technique in image construction, wherein the output devicein contrast to interlaced scanningdoes not receive line interlaced half images, but real full images.

(144) (Cited according to http://de.wikipedia.org/wiki/Zeilensprungverfahren and http://de.wikipedia.org/wiki/Vollbildverfahren)