Patent classifications
G06K9/34
Image processing apparatus, image processing method, and image processing program for clipping images included in a large image
The apparatus includes an image data obtainer, a candidate region extractor, a candidate line extractor, an overlap degree determiner, and a clip image region extractor. The candidate region extractor extracts, as a candidate region, a region containing an object detectable from the image data. The candidate line extractor extracts, as a candidate line, a line that is at least either a line segment or an arc included in the image data. The overlap degree determiner determines whether the degree of overlap between a closed line forming the outline of the candidate region extracted and the candidate line extracted is greater than or equal to a preset predetermined first percentage value. If the overlap degree determiner determines that the degree of overlap is greater than or equal to the first percentage value, the clip image region extractor 19 extracts the candidate region as a clip image.
Controlling a route based on priority levels associated with delivery action or surveillance action
An unmanned aerial vehicle (UAV) may perform a surveillance action at a property of an authorized party. The property may be defined by a geo-fence, which may be a virtual perimeter or boundary around a real-world geographic area. The UAV may image the property to generate surveillance images, and the surveillance images may include image data of objects inside the geo-fence and image data of objects outside the geo-fence. While gathering surveillance images, or after the surveillance images have been gathered, the geo-fence information may be used to obscure or remove image data referring to objects outside the geo-fence. Geo-clipped surveillance images may be generated by physically constraining a sensor of the UAV, by performing pre-image capture processing, or post-image capture processing. Geo-clipped surveillance images may be limited to authorized property, so privacy is ensured for private persons and property.
IMAGING APPARATUS, IMAGE DATA PROCESSING METHOD OF IMAGING APPARATUS, AND PROGRAM
An imaging apparatus includes a storage portion that stores captured image data obtained by imaging a subject by an imaging element and is incorporated in the imaging element, an output portion that is incorporated in the imaging element, and a plurality of signal processing portions that are disposed outside the imaging element, in which the output portion includes a plurality of output lines each disposed in correspondence with each of the plurality of signal processing portions and outputs each of a plurality of pieces of image data into which the captured image data stored in the storage portion is divided, to a corresponding signal processing portion among the plurality of signal processing portions from the plurality of output lines, and any of the plurality of signal processing portions combines the plurality of pieces of image data.
Robust, adaptive and efficient object detection, classification and tracking
Embodiments of a method and system described herein enable capture of video data streams from multiple, different video data source devices and the processing of the video data streams. The video data streams are merged such that various data protocols can all be processed with the same worker processors on different types of operating systems, which are typically distributed. In an embodiment the multiple video data sources comprises at least one mobile device executing a video sensing application that produces a video data stream for processing by video analysis worker processes. The processes include automatically detecting moving objects in a video data stream, and further tracking and analyzing the moving objects.
Video processing for embedded information card localization and content extraction
Metadata for one or more highlights of a video stream may be extracted from one or more card images embedded in the video stream. The highlights may be segments of the video stream, such as a broadcast of a sporting event, that are of particular interest. According to one method, video frames of the video stream are stored. One or more information cards embedded in a decoded video frame may be detected by analyzing one or more predetermined video frame regions. Image segmentation, edge detection, and/or closed contour identification may then be performed on identified video frame region(s). Further processing may include obtaining a minimum rectangular perimeter area enclosing all remaining segments, which may then be further processed to determine precise boundaries of information card(s). The card image(s) may be analyzed to obtain metadata, which may be stored in association with at least one of the video frames.
APPARATUS, METHOD, AND STORAGE MEDIUM
An apparatus includes an extract unit configured to extract features of a first image based on an electromagnetic wave in a first frequency band, an acquire unit configured to acquire motion information about the features, a classify unit configured to classify the features into a first group and a second group based on the motion information, and a remove unit configured to remove, from the first image, a signal corresponding to the feature belonging to the first group.
Character line division apparatus and method, and storage medium
The character line division apparatus has character line contact determination unit configured to determine whether a character line of interest comes into contact with another character line in the height direction, count value derivation unit configured to derive a count value of the number of black pixels in the width direction in the character line of interest in the character line of interest determined to come into contact with another character line, and character line division unit configured to perform character line division for the character line of interest determined to come into contact with another character line on the periphery of a reference point derived from the character line information and the character height information by taking a first position at which the count value becomes a minimum in the height direction of the character line of interest as a character line division position.
Region proposal networks for automated bounding box detection and text segmentation
Arbitrary image data may be transformed into data suitable for optical character recognition (OCR) processing. A processor may generate a plurality of intermediate feature layers of an image using convolutional neural network (CNN) processing. For each intermediate feature layer, the processor may generate at least one text proposal using a region proposal network (RPN). The at least one text proposal may comprise a portion of the intermediate feature layer that is predicted to contain text. The processor may merge the text proposals with one another to form a patch of the image that is predicted to contain text. The processor may determine outer coordinates of the patch. The outer coordinates may comprise at least leftmost, rightmost, topmost, and bottommost coordinates. The processor may generate a quadrilateral of the image that is a smallest quadrilateral including the leftmost, rightmost, topmost, and bottommost coordinates.
Systems and methods for modifying labeled content
Systems and methods are disclosed for modifying labeled target content for a capture device. A computer-implemented method may use a computer system that includes non-transient electronic storage, a graphical user interface, and one or more physical computer processors. The computer-implemented method may include: obtaining labeled target content, the labeled target content including one or more facial features that have been labeled; modifying the labeled target content to match dynamically captured content from a first capture device to generate modified target content; and storing the modified target content. The dynamically captured content may include the one or more facial features.
Distributable descriptive recipe for intelligent image and video processing systems
This disclosure relates to a method for rendering images. First, a user request is received from a user interface to access an image effect renderer recipe, comprising conditional logic and non-visual image data, from an effect repository. Next, at least one image signal is received. Objects are identified within the image signal(s). The image effect renderer recipe is processed via an effect renderer recipe interpreter to generate image processing steps and image processing prioritizations. The image processing steps are then ordered in accordance with the image processing prioritizations. Next, an image processor applies the image processing steps to the identified objects of the image signal(s) to generate at least one processed image signal. The processed image signal(s) are then displayed on a display device.