G06F16/535

INFORMATION PROCESSING UNIT, INFORMATION PROCESSING METHOD, AND PROGRAM

An information processing unit includes: a diagnostic image input section that inputs the diagnostic image; an operation information obtaining section that obtains display operation history information representing an operation history of a user who controls displaying of the diagnostic image; a query image generation section that extracts a predetermined region of the input diagnostic image to generate a query image; a diagnosed image obtaining section that supplies the generated query image and the display operation history information to a diagnosed image search unit and obtains the diagnosed image obtained as a search result by the diagnosed image search unit; and a display control section that displays the diagnostic image and the obtained diagnosed image for comparison.

System and Method for Automatically Selecting Images to Accompany Text

A system for selecting an image to accompany text from a user in connection with a social media post. The system is capable of receiving text from the user, identifying one or more search terms based on the text, identifying candidate images from images in one or more image databases using the search terms, presenting one or more candidate images to the user, receiving from the user a selected image from the one or more candidate images, generating the social media post comprising the selected image and the user-submitted text, and transmitting the social media post for display.

Determining fine-grain visual style similarities for digital images by extracting style embeddings disentangled from image content

The present disclosure relates to systems, methods, and non-transitory computer readable media for accurately and flexibly identifying digital images with similar style to a query digital image using fine-grain style determination via weakly supervised style extraction neural networks. For example, the disclosed systems can extract a style embedding from a query digital image using a style extraction neural network such as a novel two-branch autoencoder architecture or a weakly supervised discriminative neural network. The disclosed systems can generate a combined style embedding by combining complementary style embeddings from different style extraction neural networks. Moreover, the disclosed systems can search a repository of digital images to identify digital images with similar style to the query digital image. The disclosed systems can also learn parameters for one or more style extraction neural network through weakly supervised training without a specifically labeled style ontology for sample digital images.

Determining fine-grain visual style similarities for digital images by extracting style embeddings disentangled from image content

The present disclosure relates to systems, methods, and non-transitory computer readable media for accurately and flexibly identifying digital images with similar style to a query digital image using fine-grain style determination via weakly supervised style extraction neural networks. For example, the disclosed systems can extract a style embedding from a query digital image using a style extraction neural network such as a novel two-branch autoencoder architecture or a weakly supervised discriminative neural network. The disclosed systems can generate a combined style embedding by combining complementary style embeddings from different style extraction neural networks. Moreover, the disclosed systems can search a repository of digital images to identify digital images with similar style to the query digital image. The disclosed systems can also learn parameters for one or more style extraction neural network through weakly supervised training without a specifically labeled style ontology for sample digital images.

System and Method of Identifying Visual Objects

A system and method of identifying objects is provided. In one aspect, the system and method includes a hand-held device with a display, camera and processor. As the camera captures images and displays them on the display, the processor compares the information retrieved in connection with one image with information retrieved in connection with subsequent images. The processor uses the result of such comparison to determine the object that is likely to be of greatest interest to the user. The display simultaneously displays the images the images as they are captured, the location of the object in an image, and information retrieved for the object.

MULTI-DOMAIN CONVOLUTIONAL NEURAL NETWORK

In one embodiment, an apparatus comprises a memory and a processor. The memory is to store visual data associated with a visual representation captured by one or more sensors. The processor is to: obtain the visual data associated with the visual representation captured by the one or more sensors, wherein the visual data comprises uncompressed visual data or compressed visual data; process the visual data using a convolutional neural network (CNN), wherein the CNN comprises a plurality of layers, wherein the plurality of layers comprises a plurality of filters, and wherein the plurality of filters comprises one or more pixel-domain filters to perform processing associated with uncompressed data and one or more compressed-domain filters to perform processing associated with compressed data; and classify the visual data based on an output of the CNN.

MULTI-DOMAIN CONVOLUTIONAL NEURAL NETWORK

In one embodiment, an apparatus comprises a memory and a processor. The memory is to store visual data associated with a visual representation captured by one or more sensors. The processor is to: obtain the visual data associated with the visual representation captured by the one or more sensors, wherein the visual data comprises uncompressed visual data or compressed visual data; process the visual data using a convolutional neural network (CNN), wherein the CNN comprises a plurality of layers, wherein the plurality of layers comprises a plurality of filters, and wherein the plurality of filters comprises one or more pixel-domain filters to perform processing associated with uncompressed data and one or more compressed-domain filters to perform processing associated with compressed data; and classify the visual data based on an output of the CNN.

AUTOMATICALLY DETECTING USER-REQUESTED OBJECTS IN DIGITAL IMAGES
20230237088 · 2023-07-27 ·

The present disclosure relates to an object selection system that accurately detects and optionally automatically selects user-requested objects (e.g., query objects) in digital images. For example, the object selection system builds and utilizes an object selection pipeline to determine which object detection neural network to utilize to detect a query object based on analyzing the object class of a query object. In particular, the object selection system can identify both known object classes as well as objects corresponding to unknown object classes.

AUTOMATICALLY DETECTING USER-REQUESTED OBJECTS IN DIGITAL IMAGES
20230237088 · 2023-07-27 ·

The present disclosure relates to an object selection system that accurately detects and optionally automatically selects user-requested objects (e.g., query objects) in digital images. For example, the object selection system builds and utilizes an object selection pipeline to determine which object detection neural network to utilize to detect a query object based on analyzing the object class of a query object. In particular, the object selection system can identify both known object classes as well as objects corresponding to unknown object classes.

System and Method for Automatically Selecting Images to Accompany Text

A system for selecting an image to accompany text from a user in connection with a social media post. The system is capable of receiving text from the user, identifying one or more search terms based on the text, identifying candidate images from images in one or more image databases using the search terms, presenting one or more candidate images to the user, receiving from the user a selected image from the one or more candidate images, generating the social media post comprising the selected image and the user-submitted text, and transmitting the social media post for display.