METHOD & APPARATUS FOR PRODUCING A VIDEO IMAGE STREAM
20230123905 · 2023-04-20
Inventors
Cpc classification
G06Q30/0643
PHYSICS
G06T19/20
PHYSICS
International classification
G06T19/20
PHYSICS
G06T7/246
PHYSICS
Abstract
Methods and apparatus are described for producing a computer-generated image by monitoring one or more kinds of activity of a user to identify at least one item of interest; and creating a computer-generated image including a representation of the item of interest overlaid onto an image of a model, environment or contextually relevant scenario. The process begins at 102 with monitoring a user's online activity. Software tracks and stores which purchasable items on the website/mobile application the user appears to be most interested. In a particular example, a user's abandoned basket feed is tracked and stored. In another example, cookies information is stored. Another indicator of items that a user is most interested in is length of time browsing a webpage associated with a particular item. Dwell time and cursor tracking software may also be used to identify which items the user viewed most. At 114, a model simulation is created. The object to be modelled relates to the items of interest. In the example where the items of interest are wearable items such as clothing garments, the model to be simulated is a human person said human person resembling the user as closely as possible based on the available personal data.
Claims
1. A method of producing a computer-generated image, the method comprising: monitoring one or more kinds of activity of a user to identify at least one item of interest; and creating a computer-generated image including a representation of the item of interest overlaid onto an image of a model.
2. The method of claim 1, further comprising one or more of the following steps: collecting and storing data of the at least one item of interest; retrieving personal and/or contextual data about the user from at least one of an online source and a dataset; selecting a three-dimensional model simulation from a database of three-dimensional model simulation files; selecting an action simulation for the model simulation to perform from a database of stock action files, the selected action being at least partly based on contextual data of the user; retrieving an item image file of the at least one item of interest from a database of item image files based on the stored data of the at least one item of interest; applying an item-specific dynamics model and a motion tracking model to the item image file; compiling the selected model simulation, action simulation and image file to create the single computer-generated video image stream.
3. The method of claim 2, further comprising applying an item-specific texture simulation to the item image file before compiling a video image stream.
4. The method of claim 1, wherein the user's online activity includes items viewed on an eCommerce site.
5. The method of claim 4, wherein the ecommerce site is a retail site and the at least one item of interest is a garment.
6. The method of claim 4, wherein the model to be simulated is a human model.
7. The method of claim 2, wherein: the model to be simulated is a human model; and the human model is created using a three-dimensional person simulator at least partly based on the personal data.
8. The method of claim 7, wherein the personal data comprises at least one of user age, height, clothes measurements, race, body type, and gender.
9. The method of claim 2, wherein: the user's online activity includes items viewed on an eCommerce site; and the item image file comprises a fabric pattern image of the garment.
10. The method of claim 9, wherein the item-specific dynamics model is a fabric dynamic model.
11. The method of claim 3, wherein: the user's online activity includes items viewed on an eCommerce site; the item image file comprises a fabric pattern image of the garment; and the item-specific texture simulation is a fabric texture simulation.
12. The method of any of claim 2, wherein the stock action files comprise three-dimensional pre-recorded motion tracking data.
13. The method of claim 2, wherein the model simulation files comprises pre-recorded lighting data.
14. The method of claim 1, further comprising rendering a background to the video image stream, the background selected from a database of pre-recorded background images based on contextual data on the user.
15. The method of claim 14, wherein the contextual data on the user comprises at least one of location, time of day, time of year, local weather conditions and address.
16. The method of claim 1, further comprising retrieving situational data about the user from at least one of an online source and a dataset.
17. The method of claim 16, wherein situational data comprises life events that the user is experiencing.
18. The method of claim 1, further comprising rendering the video image stream in a cloud environment.
19. The method of claim 18, further comprising delivering the rendered video through to a video service for serving of the video in an inline frame to the user.
20. An apparatus for creating a computer-generated image, the apparatus comprising: a monitoring unit for monitoring one or more kinds of user activity to identify at least one item of interest; and a compilation module for creating a computer-generated image including a representation of the item of interest overlaid onto an image of a model.
21. The apparatus of claim 20, further comprising one or more of the following: a first data storage unit for storing the collected data relating to at least one item of interest; a data retrieval unit for retrieving personal and/or contextual data about the user from at least one of an online source and a dataset; a second storage unit for storing a user's personal and/or contextual data; a first database of three-dimensional model simulations and a first processing unit for selecting a model simulation selected from the first database of three-dimensional model simulation; a second database of stock action files of actions for the model simulation to perform and a second processing unit for selecting an action file from the second database, the action selected at least partly on contextual data; a third database of item image files and a third processing unit for retrieving an item image file of a computer image of the at least one item of interest from the third database; a fourth processing unit for applying an item-specific dynamics model and a motion tracking model to the item image file; and wherein the compilation module compiles the selected model simulation, action file and item image file into the single computer-generated video image stream.
22. The apparatus of claim 20, further comprising a cloud environment module for rendering a video image file.
23. The apparatus of claim 20, further comprising a video delivery module for delivering a rendered video through to a video service for serving of the video in an inline frame to the user.
24. A program comprising instructions that, when executed on a processor, performs a method of claim 1.
25. A computer-generated video image stream created using a method of claim 1.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0041] A preferred embodiment of the present invention will now be described, by way of example only, with reference to the accompanying diagrammatic drawings, in which:
[0042]
[0043]
[0044]
[0045]
[0046]
[0047]
DETAILED DESCRIPTION OF PREFERRED EMBODIMENT
[0048] Referring to the flow chart 100 in
[0049] A user's physical location may also be monitored, for example to identify stores that the user visits in person. Various examples of mobile phone tracking software are available which can provide this data. The data may then be used alone or in conjunction with data gathered from other sources to identify an item of interest.
[0050] Once items of interest have been identified by any of the above described methods, data of the item(s) of interest is stored 104 in a storage device such as a server or cache folder. At 106, personal information relating to the user is retrieved from at least one online source. Examples of personal data include, but are not limited to: age, dress size, height, race, body type and gender. Online sources may include social media platforms such as Instagram®, Facebook®, Snapchat®, YouTube® and Twitter®. Personal and/or behavioural data may also be retrieved and cross referenced from other third party datasets.
[0051] At 110, contextual data relating to the user is retrieved from at least one source. Contextual data includes, but is not limited to, location, real time weather forecast, time of day. Situational data relating to the user may also be retrieved. Situational data relates to the user's personal circumstances. A user's relationship status, whether they own a dog, whether they have kids, whether they have an invite to attend a function in the near future are some non-limiting examples of a user's situational data.
[0052] At 114, a model simulation is created. The object to be modelled relates to the items of interest. In the example where the items of interest are wearable items e.g. clothing garments, the model to be simulated is a human person said human person resembling the user as closely as possible based on the available personal data.
[0053] Once the model has been simulated, the model is programmed to perform 116 an action. The action is selected from a database of pre-programmed actions based on which action is most relevant to the consumer based on personal information, contextual information and/or situational data.
[0054] The selected action may also take into consideration the nature of the items of interest. For example, if the items of interest are athletic clothing, then the human model may be programmed to be perform a running action, if the items of interest include a party dress the human model may be programmed to perform a dancing action.
[0055] In the example where the item of interest is a city car, the action may be parallel parking. For an off road vehicle, the action may be driving over an undulating surface.
[0056] At 118, a data code for each item of interest is compared to a data code relating to a database of computer rendered images of a selection of items of interest.
[0057] In
[0058] At 120, the illusion of the simulated video image is refined by adding a motion tracking model and an item-specific dynamics model. The motion tracking model ensures that the item moves along with the model when the model is performing the selected action. The item-specific dynamics model ensures the item response to the moment of the model as it would do in a real-life scenario. The motion tracking model and the item-specific dynamics model give the viewer the illusion of viewing a real-world video image stream. The illusion of the video image stream may be further improved by adding a materials rendering model. This model produces a realistic impression to the materials that make up the item of interest.
[0059] In the example that the item of interest is clothing, the items will be items of clothing: t-shirts, dresses, coats, trousers etc and will motion track a human model performing an action. Taking a dress as a specific non-limiting example, the dress will move with the human model. The dress will have fixture points which fix the dress to the model, for example, at the shoulders and around the waist. The rest of the material will follow these fixed points based on momentum from the performed action, gravity and the material of the dress and is dictated by the item specific dynamics model in the form of a fabric dynamic model. The material rendering model will accurately simulate the materials of the dress in the form of a fabric renderer. The fabric renderer includes sheen, weave, texture and refraction of light. In essence, the fabric renderer ensures a true representation of what fabric and materials look like in real life.
[0060] The next stage to creating a computer-generated realistic video image stream related to the user is to provide 122 a contextualised background environment to the video. The context of the background will be specific to the user and may be derived from personal, contextual and/or situational data. For example, if the user lives in London and it is winter the background may be a snowy Oxford Street. The background may also take into consideration the item of interest. For example, if the item of interest is a bikini, the background may be a sunny beach in Ibiza.
[0061] The entire video file is rendered in a cloud environment 124. The video image is then delivered 126 through a video service. The video service serves 128 the video in an inline frame to the user. The inline frame could be an Instagram® story advertisement, a Google® display advertisement. Again, the choice of hosting platform for the inline frame can be based on contextual information.
[0062]
[0063]
[0064] As described in relation to flow diagram 300, once the video image stream has been created it is sent to a video storage device 306 to be served on a webpage 308 in an inline frame 307.
[0065] A caption 508a, 508b relating to the video can be included. In the example video 504a the caption 508a reads “AFTER WORK DATE?”. Example video 504b displays the same “AFTER WORK DATE?” caption 508b. The caption 508a, 508b may be generated based on situational data of the user. In the example of videos 504a and 504b, the situational data that may have prompted the caption 508a, 508b is information that the user subscribes to a dating site, that they have recently changed their relationship status on a social media platform such as Facebook®.
[0066] The human model 510 is selected from a database of human models using user related personal data. For example, the human model 510 is a white, 5 foot 7 inch women of slim build aged around 25 years old. Thus, the human model 510 has used variables: ethnicity, height, gender, body shape and age to generate a human model that represents the user.
[0067]
[0068] The apparatus 600 further comprises a third database 618 of item image files and a third processing unit 620. The processing unit 620 uses the information about the item(s) of interest stored in the first data storage unit 604 to retrieve an item image file of a computer image of the item of interest from the third database 618. The apparatus 600 further includes a fourth processing unit 622 for applying an item-specific dynamics model and a motion tracking model to the item image file. The apparatus 600 includes a compilation module 624 for compiling the selected model simulation, action simulation and image file into a single computer-generated video image stream. The video image stream will display the item of interest overlaid onto the model simulation performing the action.
[0069] The method of using artificial intelligence to produce a personalised video image described herein has been described in relation to the online retail industry and the automotive industry. However, the method, apparatus and programme described herein can be used in other industries such as: electronics consumer goods, home furnishings, food and groceries to name a few non-limiting examples.
[0070] The term “user” herein may be taken, for example, to mean prospective or actual customer. Composite images may be created which include a representation of an item of interest overlaid/combined with an image of a model, which model may be of a thing, a person, an environment or a contextually relevant scenario. The model may be selected according to the nature of the item of interest.
[0071] The images created in accordance with the present invention may be used in an advertising network, such as (but not limited to) the Google (®) Display Network and/or social media platforms such as (but not limited to) Facebook (®) and Instagram (®).
[0072] The images may be “stills”—i.e. non-moving images or more preferably moving video images depicting action.
[0073] The images may include computer generated images of items, such as garments, in combination with model images, which model images may themselves be real or computer generated.
[0074] A user's online activity may be monitored using tracking technology, such as tracking pixels. Physical location of a user may be monitored using mobile (phone) tracking software, so that when a user visits a particular location, such as a store, that data may be acquired and used as part of the process to determine items that may be of interest to the user.
[0075] When rendering images of an item of interest, and/or of a model in conjunction with the item, machine learning techniques may be employed such as (but not limited to) Generative Adversarial Network techniques, for example to predict how an item will look and/or move in a particular situation or from a particular angle.
[0076] Whilst endeavouring in the foregoing specification to draw attention to those features of the invention believed to be of particular importance, it should be understood that the applicant claims protection in respect of any patentable feature or combination of features referred to herein, and/or shown in the drawings, whether or not particular emphasis has been placed thereon.