METHOD OF AUGMENTING A DATASET USED IN FACIAL EXPRESSION ANALYSIS
20230282028 · 2023-09-07
Inventors
Cpc classification
International classification
Abstract
In a computer-implemented method of augmenting a dataset used in facial expression analysis, a first facial image and a second facial image are added to a training/testing dataset and mapped to two respective points in a continuous dimensional emotion space. The position of a third point in the continuous dimensional emotion space between the first two points is determined. Augmentation is achieved when a labelled facial image is derived from the third point based on its position relative to the first and second facial expression.
Claims
1. A computer-implemented method of augmenting a dataset used in facial expression analysis, the method comprising the steps of: adding to a dataset a first facial image and a second facial image; mapping the first and second facial images to a first expression point and second expression point respectively in a continuous dimensional emotion space; determining a position of a third expression point in the continuous dimensional emotion space, the third expression point being a position between the first and second expression point; generating an expression variation image; assigning a first dimensional label to the expression variation image; and adding the expression variation image to the dataset, wherein the expression variation image and first dimensional label are derived from the position of the third expression point, based on its position relative to the first and second facial image.
2. The method according to claim 1, wherein the expression variation image is obtained by applying a first morph function to interpolate the first facial image and the second facial image and the intensity variation image is obtained by applying a second morph function to interpolate the neutral facial image and the first facial image.
3. The method according to claim 1, wherein the third expression point is determined by applying an expression distance to the first expression point or second expression point.
4. The method according to claim 3, wherein the expression distance is preselected.
5. The method according to claim 1, wherein the first and second facial images are each assigned a label comprising the following expressions: happy, surprised, afraid, angry, disgusted or sad and are mapped to positions that correspond to the labels in the continuous dimensional emotion space.
6. The method according to claim 1, wherein the first and second facial images are apex expressions.
7. The method according to claim 1, further comprising the step of: adding to the dataset, a neutral facial image; mapping the neutral facial image to a neutral point in the continuous dimensional emotion space; determining a first intensity point in the continuous dimensional emotion space, the first intensity point being a position between the neutral facial image and one of: the first facial image, the second facial image or the expression variation image; generating an intensity variation image; assigning a second dimensional label to the first variation image; and adding the intensity variation image to the dataset; wherein the intensity variation image and the second dimensional label are based on the first intensity point's relative position between the neutral facial image and the first facial image.
8. The method according to claim 7, wherein the first intensity point is determined by applying an intensity distance to the neutral expression point or the first facial image.
9. The method according to claim 7, wherein the intensity distance is preselected.
10. The method according to claim 7, wherein the expression variation image is obtained by applying a first morph function to interpolate the first facial image and the second facial image and the intensity variation image is obtained by applying a second morph function to interpolate the neutral facial image and the first facial image.
11. A storage medium comprising machine readable instructions stored thereon for causing a computer system to perform a method according to claim 1.
12. A computer-implemented method of augmenting a dataset used in facial expression analysis comprising the steps of: adding to a dataset, a neutral facial image and a first facial image; mapping the neutral facial image to a neutral point and the first facial image to a first expression point, the neutral point and first expression point being located in a continuous dimensional emotion space; determining a first intensity point on an intensity plane in the continuous dimensional emotion space, the first intensity point being a position between the neutral facial image and the first expression; generating an intensity variation image; and assigning a dimensional label to the intensity variation image; wherein the intensity variation image and the dimensional label are based on the first intensity point's relative position between the neutral facial image and the first facial image.
13. The method according to claim 12, wherein the intensity variation image is obtained by applying a second morph function to interpolate the neutral facial image and the first facial image.
14. The method according to claim 13, wherein the continuous dimensional emotion space is a valence-arousal circumplex.
15. A storage medium comprising machine readable instructions stored thereon for causing a computer system to perform a method according to claim 12.
Description
BRIEF DESCRIPTION OF DRAWINGS
[0014]
[0015]
[0016]
DETAILED DESCRIPTION OF THE INVENTION
[0017] In the following detailed description, reference is made to the accompanying drawings, which form a part hereof. The illustrative embodiments described in the detailed description, drawings and claims are not meant to be limiting. Other embodiments can be utilized, and other changes can be made, without departing from the spirit or scope of the subject matter presented herein.
[0018] The present invention provides a method of augmenting a dataset used in Facial Expression Analysis systems 100 specifically, by it being premised on the observation that many expression variations can be approximated by generating high-quality morphings between the existing available images of a categorical dataset. While expected to be typically employed to augment training datasets used to train machine learning models used in facial expression analysis, the method can likewise be used to augment a testing dataset.
[0019] Referring to
[0020] In one embodiment, the first facial image and second facial image both depict an apex expression, meaning the facial expression is a peak intensity.
[0021] Referring to
[0022] The first and second facial images are mapped to a first and second expression point respectively in a continuous dimensional emotion space (step 102). A non-limiting example of a continuous two-dimensional emotion space is the valence-arousal space of the circumplex model which will be discussed further as a non-limiting example.
[0023] Since the first and second facial images are already labelled, the first and second expression points can be very specific coordinates in the valence-arousal space of the circumplex model. Referring to
[0024] Next, a position of a third expression point in the continuous dimensional emotion space is determined, the third expression point being a position between the first and second expression point (step 103). In one embodiment, the position of the third expression point is determined by applying an expression distance to either the first or second expression point, so long as the third expression point lies therebetween. In one embodiment, further expression points can be determined by applying the expression distance to the third expression point and so on to obtain a series of expression points between the first and second expression point. Although not strictly required, in one embodiment, the expression distance is preselected based on the specific needs of the training or testing dataset, such as expression granularity, augmentation factor and symmetry between the points. In one embodiment, the expression distance is an angle increment of 15° starting from either the first or second expression point and strikes a balance between the aforementioned criteria. In one embodiment, the expression is preselected. The method as described above can also be applied to a neutral point and any expression image, including the first facial image and second facial image and newly generated images as a result of augmentation such as an expression variation image described below to generate an intensity variation image with the third expression point being analogous to the first intensity point for the purposes of mapping.
[0025] In one embodiment, the first and second facial images correspond to points that are immediately adjacent to each other when mapped in the continuous dimensional emotion space. By way of example, and once again referring again to
[0026] With the position of the third expression determined, an expression variation image can be generated, and a first dimensional label can be assigned to the expression variation image, the expression variation image and first dimensional label being derived from the position of the third expression point, based on its position relative to the first and second facial expression (step 104). In a non-limiting example, and assuming a two-dimensional valence arousal space of the circumplex model.
[0027] Let F.sub.i.sup.E denote the face image of subject i with facial expression E. For categorical datasets, usually E∈{NE,HA,SU,AF,AN,DI,SA}. Let θ.sup.E E denote the specific angle of each expression in the polar valence-arousal space, as estimated from emotion studies. See J. A. Russell, “A circumplex model of affect,” Journal of Personality and Social Psychology, vol. 39, no. 6, p. 1161-1178, 1980; and G. Paltoglou and M. Thelwall, “Seeing stars of valence and arousal in blog posts,” IEEE Trans. Affective Computing, vol. 4, no. 1, pp. 116-123, January 2013, each of which is incorporated herein by reference in its entirety. Let I.sub.i.sup.E∈[0, 1] denote the intensity of expression E of subject i. Zero expression intensity I.sup.e=0 coincides with NE (by definition I.sup.NE=0), while I.sup.E=1 represents the highest possible intensity of expression. Let M.sub.p (F.sub.i.sup.source, F.sub.i.sup.target, r) be a morphing function, based on p facial landmarks, that returns a new face image, which is the result of morphing F.sub.i.sup.source towards F.sub.i.sup.target with a ratio r∈[0, 1]; when r=0 the morphed image is identical to F.sub.i.sup.source and when r=1 it is identical to F.sub.i.sup.target. Any known contemporary morphing approach can be used for this, such as Delaunay triangulation followed by local warping on groups of facial landmarks.
[0028] The augmentation method according some aspects the present invention is based on 2 types of morphings. In order to synthesize new expression variations, Apex to Apex morphing (1) is used, between the given apex expressions of the categorical dataset:
[0029] Where A, A.sub.1 and A.sub.2 are apex expressions from the parent dataset, and r is a ratio in the interval [0, 1].
[0030] In order to synthesize new intensity variations, Neutral to Apex morphing (2) is used, between the NE image and a given (or interpolated) apex image:
[0031] Referring to
[0032] Optionally, the expression variation image or expression variation image can be added to the dataset (step 105).
[0033] As is traditional in the field of the disclosed technology, features and embodiments are described, and illustrated in the drawings, in terms of various steps. Those skilled in the art will appreciate that these steps are physically implemented by a computer system including one or more computers implemented with electronic (or optical) circuits such as logic circuits, discrete components, microprocessors, hard-wired circuits, memory elements, wiring connections, and the like. The steps may be implemented by a computer-readable storage medium used in conjunction with the one or more computers, and comprising machine readable instructions stored thereon for causing a computer system to perform the steps. In the case of the one or more computers being implemented by microprocessors or similar, they may be programmed using software (e.g., microcode) to perform various functions discussed herein and may optionally be driven by firmware and/or software. Alternatively, certain steps or procedures may be implemented by dedicated hardware, or as a combination of dedicated hardware to perform some functions and a processor (e.g., one or more programmed microprocessors and associated circuitry) to perform other functions. In this sense, the various steps described can be performed automatically to result in a rapid process for performing facial expression analysis or for augmenting a dataset used in performing facial expression analysis.
[0034] While example embodiments have been illustrated and described above, it will be apparent to those skilled in the art that modifications and variations could be made without departing from the scope of the present invention as defined by the appended claims.