Image capture and processing
11530986 · 2022-12-20
Assignee
Inventors
- Abhijeet GHOSH (London, GB)
- Yuliya GITLINA (London, GB)
- Giuseppe Claudio GUARNERA (London, GB)
- Daljit Singh DHILLON (London, GB)
Cpc classification
G01N21/4738
PHYSICS
International classification
Abstract
A method of image processing includes receiving a first image of human skin. The first image corresponds to a first, uniform broadband illumination condition. The method also includes receiving a second image which has the same field of view and contents as the first image. The second image corresponds to a second illumination condition which comprises a uniform narrowband illumination condition. The method also includes processing the first and second images to fit parameter maps for a spectral bidirectional scattering surface reflectance distribution function skin model. The parameter maps include a modelled melanin concentration, a modelled haemoglobin concentration, a modelled melanin blend-type fraction and a modelled epidermal haemoglobin fraction. At least three of the parameter maps are independent.
Claims
1. A method of image processing, comprising: receiving a first image of human skin, the first image corresponding to a first, uniform broadband illumination condition; receiving a second image which has the same field of view and contents as the first image, the second image corresponding to a second illumination condition which comprises a uniform narrowband illumination condition; receiving or determining first spectral data corresponding to the uniform broadband illumination condition and second spectral data corresponding to the second illumination condition; processing the first and second images to fit parameter maps for a spectral bidirectional scattering surface reflectance distribution function skin model, the parameter maps comprising a modelled melanin concentration, a modelled haemoglobin concentration, a modelled melanin blend-type fraction and a modelled epidermal haemoglobin fraction; wherein at least three of the parameter maps are independent; wherein processing the first and second images to fit parameter maps comprises applying a neural network model to inputs comprising: spectral information comprising the first and second spectral data; the first image; and the second image; wherein a final stage of the neural network outputs an output RGB albedo image, and wherein the neural network is configured to determine the parameter maps which minimise differences between the output RGB albedo image and the first image.
2. A method according to claim 1, wherein the uniform narrowband illumination condition corresponds to a blue narrowband illumination condition.
3. A method according to claim 1, wherein: the second illumination condition consists of the uniform narrowband illumination condition; or the second illumination condition comprises the uniform narrowband illumination condition superposed with the uniform broadband illumination condition.
4. A method according to claim 1, wherein the second illumination condition comprises the uniform narrowband illumination condition superposed with the uniform broadband illumination condition, and wherein applying a neural network model to inputs comprising the spectral information, the first image and the second image comprises: generating a narrowband illumination image based on the first and second images; and passing the first image, the narrowband illumination image and the spectral information as inputs to the neural network model.
5. A method according to claim 1, wherein determining first spectral data and second spectral data comprises: receiving a third image of a colour test card, the third image corresponding to the uniform broadband illumination condition; determining first spectral data corresponding to the broadband illumination condition based on the third image; receiving a fourth image of the colour test card, the fourth image corresponding to the second illumination condition; determining second spectral data corresponding to the second illumination condition based on the fourth image.
6. A method comprising: receiving an albedo image showing human skin; receiving or determining spectral data corresponding to an illumination condition used to obtain the albedo image; processing the albedo image to fit parameter maps for a spectral bidirectional scattering surface reflectance distribution function skin model, the parameter maps comprising a modelled melanin concentration, a modelled haemoglobin concentration, a modelled melanin blend-type fraction and a modelled epidermal haemoglobin fraction; wherein at least three of the parameter maps are independent; wherein processing the albedo image to fit parameter maps comprises applying a neural network model to inputs comprising: the albedo image; and the spectral data; wherein a final stage of the neural network outputs an output RGB albedo image, and wherein the neural network is configured to determine the parameter maps which minimise differences between the output RGB albedo image and the albedo image.
7. A method according to claim 6, wherein determining spectral data comprises: receiving a calibration image of a colour test card, the calibration image corresponding to the illumination condition; determining spectral data corresponding to the illumination condition based on the calibration image.
8. A method according to claim 6, further comprising: receiving a first input parameter map related to melanin concentration and corresponding to the albedo image; receiving a second input parameter map related to haemoglobin concentration and corresponding to the albedo image; wherein inputs to the neural network model further comprise: the first input parameter map; and the second input parameter map.
9. A method of image processing, comprising: receiving a first image of human skin, the first image corresponding to a first, uniform broadband illumination condition; receiving a second image which has the same field of view and contents as the first image, the second image corresponding to a second illumination condition which comprises a uniform narrowband illumination condition; receiving or determining first spectral data corresponding to the uniform broadband illumination condition and second spectral data corresponding to the second illumination condition; processing the first and second images to fit parameter maps for a spectral bidirectional scattering surface reflectance distribution function skin model, the parameter maps comprising a modelled melanin concentration, a modelled haemoglobin concentration, a modelled melanin blend-type fraction and a modelled epidermal haemoglobin fraction; wherein at least three of the parameter maps are independent; wherein processing first and second images to fit parameter maps comprises using a three-dimensional spectral look-up table or using a four-dimensional spectral look-up table; wherein the spectral look-up table is constructed based on spectral information comprising the first spectral data and the second spectral data, and further based on the spectral skin reflectance model.
10. A method according to claim 9, wherein the uniform narrowband illumination condition corresponds to a blue narrowband illumination condition.
11. A method according to claim 9, wherein: the second illumination condition consists of the uniform narrowband illumination condition; or the second illumination condition comprises the uniform narrowband illumination condition superposed with the uniform broadband illumination condition.
12. A method according to claim 9, wherein determining first spectral data and second spectral data comprises: receiving a third image of a colour test card, the third image corresponding to the uniform broadband illumination condition; determining first spectral data corresponding to the broadband illumination condition based on the third image; receiving a fourth image of the colour test card, the fourth image corresponding to the second illumination condition; determining second spectral data corresponding to the second illumination condition based on the fourth image.
13. A method comprising: receiving an albedo image showing human skin; receiving or determining spectral data corresponding to an illumination condition used to obtain the albedo image; processing the albedo image to fit parameter maps for a spectral bidirectional scattering surface reflectance distribution function skin model, the parameter maps comprising a modelled melanin concentration, a modelled haemoglobin concentration, a modelled melanin blend-type fraction and a modelled epidermal haemoglobin fraction; wherein at least three of the parameter maps are independent; wherein processing the albedo image to fit parameter maps comprises using a three-dimensional spectral look-up table or using a four-dimensional spectral look-up table; wherein the spectral look-up table is constructed based on the spectral data, and further based on the spectral skin reflectance model.
14. A method according to claim 13, wherein determining spectral data comprises: receiving a calibration image of a colour test card, the calibration image corresponding to the illumination condition; determining spectral data corresponding to the illumination condition based on the calibration image.
15. A method comprising: receiving an albedo image showing human skin; receiving or determining spectral data corresponding to an illumination condition used to obtain the albedo image; receiving a first input parameter map related to melanin concentration and corresponding to the albedo image; receiving a second input parameter map related to haemoglobin concentration and corresponding to the albedo image; processing the albedo image to fit parameter maps for a spectral bidirectional scattering surface reflectance distribution function skin model, the parameter maps comprising a modelled melanin concentration, a modelled haemoglobin concentration, a modelled melanin blend-type fraction and a modelled epidermal haemoglobin fraction; wherein at least three of the parameter maps are independent; wherein processing the albedo image to fit parameter maps comprises: processing the albedo image to fit first and second intermediate parameter maps for a two parameter spectral bidirectional scattering surface reflectance distribution function skin model, wherein the first intermediate parameter map comprises an intermediate melanin concentration and the second intermediate parameter map comprises an intermediate haemoglobin concentration, wherein the intermediate parameter maps are determined using a first two-dimensional spectral look up table constructed based on the spectral data; re-scaling the first and second input parameter maps based on the intermediate parameter maps; setting the parameter map corresponding to the modelled melanin concentration equal to the re-scaled first input parameter map; setting the parameter map corresponding to the modelled haemoglobin concentration equal to the re-scaled second input parameter map; processing the albedo image to fit the parameter maps corresponding to the modelled melanin blend-type fraction and the modelled epidermal haemoglobin fraction, wherein the parameter maps are determined using a second two-dimensional spectral look up table constructed based on the spectral data.
16. A method according to claim 15, wherein determining spectral data comprises: receiving a calibration image of a colour test card, the calibration image corresponding to the illumination condition; determining spectral data corresponding to the illumination condition based on the calibration image.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1) The patent or application file contains at least one drawing executed in color. Copies of this patent or patent application publication with color drawing(s) will be provided by the Office upon request and payment of the necessary fee.
(2) Certain embodiments of the present invention will now be described, by way of example, with reference to the accompanying drawings, in which:
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12)
(13)
(14)
(15)
(16)
(17)
(18)
(19)
(20)
(21)
(22)
(23)
(24)
(25)
(26)
(27)
(28)
(29)
(30)
(31)
(32)
(33)
(34)
DETAILED DESCRIPTION
(35) In this work, we propose a novel practical spectral skin measurement approach (using an LED sphere) that, in conjunction with driving the spectral skin BSSRDF of [JSB*10] with appropriate model complexity, is suitable for facial capture of live subjects with realistic spectral appearance reproduction (Section 4). The acronym BSSRDF stands for Bidirectional Scattering Surface Reflectance Distribution Function We demonstrate that the proposed model complexity involving four parameters (melanin C.sub.m and hemoglobin C.sub.h concentration, melanin blend-type fraction β.sub.m, and epidermal hemoglobin C.sub.he fraction) is required to match subject appearance in photographs, which may not be possible using a reduced model with just two free parameters (melanin and hemoglobin), and our proposed measurement protocol combining two complementary broad and narrow-band spectral illumination conditions provides higher quality estimates of spectral parameters than those obtained using just broadband illumination. Additionally, we demonstrate how to acquire ideal broadband and narrowband illumination measurements for parameter estimation within the practical constraints of LED illuminants and regular color cameras. Thus, as our primary contribution, we propose a minimal measurement and modeling complexity for data-driven reproduction of spatially varying spectral appearance of skin, including human faces.
(36) We also demonstrate how practical measurements with a hand-held off-the-shelf skin measurement device designed for dermatological applications (a Miravex Antera3D camera) can be adapted for realistic skin appearance reproduction and rendering (Section 5). Here, we demonstrate how to appropriately transform the output pigmentation and redness maps produced by the device into melanin and hemoglobin concentrations respectively, and augment them with additional model parameters (β.sub.m and C.sub.he) which are not provided by the device. Additionally, we demonstrate how neural networks can be employed for faster, improved parameter estimation given our measurements (Section 6). Finally, we demonstrate realistic rendering of subsurface scattering with our estimated parameters (in PBRT) using spatially-varying diffusion profiles (Section 7), achieving renderings of human faces using a biophysically based spectral BSSRDF that are, for the first time, comparable to photographs (see
(37) To summarize, our central high-level contributions in this work are as follows: Practical measurement of spectral skin reflectance suitable for facial capture in conjunction with appropriate model complexity of spectral BSSRDF for matching skin appearance in photographs. Adaption and augmentation of chromophore maps obtained from a hand-held dermatological skin measurement device for realistic rendering. Novel estimation of parameters from our measurements using neural networks, which is significantly faster than a look-up table search along with reduced quantization.
(38) We additionally propose the following practical contributions for spectral measurements with a color camera: An optimization procedure for combining a set of broad- and narrow-band LED illuminants to construct a metamer for desired D65 illumination. Novel indirect measurement of narrow-band LED response that enables higher quality measurement with sharper spectral isolation than direct measurement with a color camera. A genetic programming algorithm for estimation of unknown illumination spectra from a single color chart observation.
3. SPECTRAL BSSRDF MODEL COMPLEXITY
(39) We aim to drive a spectral skin BSSRDF model with practical measurements and hence prefer a model with an appropriately minimal complexity to simplify measurements while simultaneously having sufficient complexity to match the observed spatial variation in skin, particularly facial appearance. In this respect, we aim for a data-driven modeling of skin and facial appearance rather than striving for strict bio-physical accuracy of the estimated parameters. We choose the model of Jimenez et al. [JSB*10], originally proposed for facial measurements, as the starting point for our work. The model includes the following four parameters: melanin concentration (C.sub.m) in epidermis, melanin blend-type fraction β.sub.m (blend between eumelanin and pheumelanin), and hemoglobin concentration (C.sub.h) in dermis, and epidermis (C.sub.he), respectively (please see Supplemental material for details). However, unlike Jimenez et al., we have empirically found that all four parameters in the model need to be varied over the skin surface in order to closely match the appearance of real skin. Variation in β.sub.m is particularly useful for reconstructing facial appearance variation due to facial hair and around eyelids, while a higher fraction of epidermal hemoglobin C.sub.he is necessary to match the very reddish areas of a face such as the lips and cheeks. Hence, we allow β.sub.m to vary between 0.0 and 1.0, and C.sub.he to vary between 0.0 and 0.6. We also set epidermal thickness d to 0.33 mm instead of 0.25 mm suggested in previous work in order to better match the appearance of subjects with the above spectral model.
(40) This leads to a 4D spectral skin appearance model. In practice, in order to restrict the search space of the various parameters for model-fitting, we employ very coarse discretization for β.sub.m (10 bins) and C.sub.he (4 bins) parameters which have a more subtle effect on the overall appearance, while employing a large number of bins to model the dominant variation in C.sub.m and C.sub.h.
(41)
4. PRACTICAL SPECTRAL ACQUISITION FOR FACES
(42) We now present our practical measurement protocol for robustly estimating the four parameters of the BSSRDF model with a minimal set of measurements suitable for facial capture. We employ a multispectral LED sphere equipped with a combination of narrow band Red, Green, and Blue LEDs, and three types of broad band LEDs (warm 2700K, neutral 4000K, and cool 5700K which we refer to as W27, W40, and W57 respectively), and 9 color DSLR cameras (Canon 800D) for multiview acquisition of a subject. The LEDs on the sphere are all cross-polarized w.r.t. the cameras, allowing specular cancellation according to the method of [GFT*11]. We also measured the individual spectral distributions of our illuminants using a spectrometer (Sekonic SpectroMaster C700) placed at the center of the LED sphere (see
(43) 4.1. Measurement Protocol
(44) With this setup, when we restricted ourselves to a single observation (as a baseline), we found the best individual LED illumination on our LED sphere for estimating model parameters to be uniform W57 (cool white LED) illumination (see
(45) Ideal broadband measurement: Given the 6 types of LEDs in our LED sphere, we instead create a more ideal broadband illumination by computing a weighted combination of all 6 LEDs to create a D65 metamer spectra (D65′) which we instead employ for our broadband measurements. We notice an even higher contrast in skin color, particularly coloration due to skin pigmentation and redness, under the D65 metamer illumination and we make the observation that D65 spectrum (blue dominant) balances the somewhat skewed red-dominant reflectance spectrum of human skin (see Supplemental material). This is consistent with studies on the human visual system [CXW19] where D65 spectrum has been reported to be most desirable for discerning differences in skin color. Note that true D65 illumination is not possible with the LEDs on our LED sphere. Instead, the D65 metamer is created to be an approximation of the ideal D65 spectra achieved by the available LEDs such that it minimizes the color difference between the 24 color patches measured on an Xrite color chart vs the reference 24 colors on an ideal color chart in sRGB color space (which assumes ideal D65 spectrum). Our computed metamer spectrum can be seen in comparison to the ideal D65 spectrum in
(46) Narrow-band measurement: Directly recording the reflectance response of a subject under blue LED illumination unfortunately results in some colors being outside the gamut of most off-the-shelf color cameras, which typically work in the sRGB and Adobe RGB colour spaces, with the chromaticity of the narrow band being noticeably outside both colour spaces (
(47) 4.2. Isolation of Blue Response
(48) We propose an indirect scheme in order to measure skin response under blue LED illumination. We capture two photographs of a subject, the first one under broadband lighting (I.sub.W) and the second one under broadband plus blue LED (I.sub.W+nb). We then separately apply to both the images a Chromatic Adaptation Transform (CAT) (defined in Appendix B), to predict colours appearance under D65 lighting and within the sRGB gamut, and computationally recover the desired narrow band response I.sub.nb as follows:
I.sub.nb=δ×(CAT(I.sub.W+nb)−CAT(I.sub.W)/δ), (1)
(49) where the factor δ accounts for the difference in intensities of the LEDs in the two conditions.
(50) In our case, we preferably employ D65 metamer illumination for the broadband measurement (I.sub.W=I.sub.D65′) which greatly simplifies its CAT transform to identity matrix. However, the transform is general and can be employed with any broadband illumination (e.g., W=W57 cool white LED) for computing the spectral isolation as per Equation 1. The above scheme offers an additional advantage of actually measuring the skin response under a slightly narrower band than the one offered directly by blue LED (
(51) The capture process can be visually seen in
(52) 4.3. Results
(53) With the measurement protocol described in Section 4.1, we do a joint look-up table search for best matching color values (in CIELAB space) under simulated D65 metamer illumination (
(54)
5. PRACTICAL SKIN MEASUREMENTS WITH ANTERA3D
(55) The focus of the previous section was on practical spectral measurements of skin in a controlled setup suitable for facial capture. For more free-form measurement of skin, we employ a hand-held off-the-shelf device—Antera 3D© (Miravex Limited, Ireland), a camera for image acquisition and corresponding software for analysis of single skin patches (56×56 mm.sup.2). This is an instrument employed in dermatology: it has been compared with most commonly used devices in dermatological research and is reported to be robust, sensitive and precise for skin colour analysis [MFCN15, LWA*18]. For a single measurement, the camera is placed onto a skin patch without applying excessive pressure. The typical measurement procedure along with sample results for a cheek patch is shown in
(56) This is a good starting point for employing the data for spectral rendering of skin. However, the device does not capture all of the parameters we have identified as necessary for reproduction of skin appearance. Importantly, for the two parameters that are provided, the device provides chromophore concentrations in terms redness and pigmentation which do not directly map as C.sub.m and C.sub.h for the BSSRDF model. Hence, we have to undertake a number of steps in order to adapt the Antera measurements for driving the skin appearance model.
(57) 5.1. Parameter Remapping and Estimation
(58) Given that Antera estimates only the two primary parameters (pig-mentation and redness) related to melanin and hemoglobin concentration, in a first step we adapt these maps based on Jiminez et al.'s reduced 2D model of skin appearance. In order to do this, we employ the albedo map provided by Antera and estimate corresponding C.sub.m and C.sub.h for the Jimenez model using the same look-up table search procedure (in CIELAB space) described in the previous section. Given our best fit to the 2D model, we then scale antera's maps for pigmentation and redness in an appropriate manner to match the mean and variance of our estimated C.sub.m and C.sub.h parameters (using color space matching), and set these scaled pigmentation and redness maps as our final estimate of C.sub.m and C.sub.h for reconstruction.
(59) Note that this above remapping step requires us to simulate a 2D lookup table (fixed β.sub.m and C.sub.he) under the same illumination spectra employed for measuring Antera's albedo map. However, this information of the illumination spectrum for the Antera albedo is not provided by the device or the vendor and, being a proprietary device, we have no control over its LED illumination system in order to make a direct measurement of the appropriate LEDs: the device cycles through all the LEDs very rapidly, making direct measurement of spectrum (e.g., with a spectrometer) difficult. Hence, we estimate the unknown illumination spectrum for the albedo measurement using corresponding measurements of color squares on an Xrite color chart and then solve for the illumination spectrum (see
(60) After recovering the illumination spectrum for Antera's albedo, we can remap Antera's chromophore maps to the appropriate scale using a 2D look-up table based on the reduced Jimenez model. We then fix the C.sub.m and C.sub.h parameters and then re-fit the albedo data to the complete 4D model, this time searching for appropriate values of β.sub.m and C.sub.he parameters in a 4D look-up table for fixed values of C.sub.m and C.sub.h.
6. NEURAL PARAMETER ESTIMATION
(61) Thus far, we described how to estimate the spectral parameters of skin from our measurements using a look-up table search (for best matching color values in CIELAB space). This process is slow and the results can be prone to image noise and quantization due to discrete values in the look-up table. Hence, instead we also explored a neural prediction approach for obtaining the spectral parameters from our measurements using a cascaded feed-forward multilayer perceptron (MLP) architecture (see
(62) Each of the 4 parameters of our model is estimated by a different MLP (see
(63) To achieve higher accuracy, when the RGB input photograph is provided by our LED sphere setup, C.sub.mNNET and β.sub.mNNET can also take in input the synthesized response to pure blue LED illumination. Additionally, if the RGB input is provided by Antera, its remapped C.sub.m and C.sub.h measurements can be used to bypass respectively C.sub.mNNET and C.sub.hNNET.
(64) 6.1. Design and Training of the MLPs
(65) The first layer of each MLP is the input layer, followed by two hidden layers (L1 and L2) and the output layer. It is well know that two-hidden-layer feedforward networks can approximate complex nonlinear mappings with arbitrary accuracy given enough nodes in the hidden layers and training data. In our pipeline, we train the MLPs on synthetic RGB data, i.e the 4D LUTs provided by the spectral skin reflectance model, augmented by zero-mean white Gaussian noise to simulate photon shot noise. The noise variance is directly estimated from homogeneous areas in a photograph of a color chart. We then analyze the effect of the noise on the estimated parameters in order to derive a suitable noise model to inject in the skin model parameters used to train the MLPs.
(66) To determine the number of nodes N.sub.L.sub.
(67)
(68) Since such upper bounds might overfit the input data [Gua03], to avoid overfitting and to increase the generalization capabilities of our MLPs we set N.sub.L.sub.
(69) We train the parameter networks C.sub.mNNET, β.sub.mNNET, C.sub.heNNET and C.sub.hNNET once for each input type (e.g., broadband-only or broadband+synth. blue) with D65′ as the broadband spectrum for LED sphere data, and the recovered Antera spectrum as the broadband for Antera data. However, the RGBAlbedoNNET used for albedo reconstructions from the input parameters is specific for a given illumination spectrum, and needs to be trained separately for different target illumination spectras.
(70) 6.2. Results
(71)
(72)
7. ADDITIONAL RESULTS AND RENDERING
(73) We now present additional set of results with our proposed practical spectral measurements for faces, and skin-patch measurements with the Antera3D device.
(74)
(75) We note that parameters estimated using a single broadband measurement of the albedo encodes some amount of subsurface scattering in the parameter maps due to subsurface scattering being baked in the albedo (also observed by Jimenez et al. [JSB*10]). Since our look-up table based reconstruction of the albedo does not do explicit simulations of subsurface scattering, we actually found closer matches (in CIELAB space) to the input photographs when reconstructing the albedo using parameters estimated under a single broadband condition compared to our proposed approach of combining the broadband measurement with narrow-band blue response (also true for Antera data where full 4D search results in better match to input photograph). The reconstructions with our proposed measurements are slightly sharper, with less baked-in subsurface scattering compared to the input photograph which we attribute to the sharp descattered measurement of the narrow-band response. This is actually a desirable outcome for rendering of subsurface scattering with the measured parameter maps. Jimenez et al. employed their measured maps to only reconstruct the albedo which they employed as a modulation texture to a homogeneous subsurface rendering with a fixed diffusion profile. In contrast, our estimated maps allow us to render subsurface scattering with spatially varying albedo and diffusion profiles as explained next.
(76) 7.1. Rendering Subsurface Scattering
(77) Inspired by the work of Donner et al. [DWd*08], we employ our estimated parameters to render heterogeneous subsurface scattering in skin. We employed PBRT v2 [PH10] to generate renderings with pseudo-heterogeneous subsurface scattering driven by our estimated spatially varying parameters of the spectral BSSRDF model. Given a facial scan of a subject obtained with multiview acquisition in the LED sphere (we employ COLMAP [SF16, SZPF16] for the base geometry reconstruction), we first project the input data of broadband D65′ response (which is also the input to COLMAP for geometry reconstruction) and synthesized narrow-band blue response into the UV texture-space of the face scan (see
(78) In order to render heterogeneous subsurface scattering, we modified the provided subsurface scattering implementation in PBRT from the default dipole diffusion kernel to our specified spatially varying profiles implementing two-layered diffusion. For each sampled color generated by the skin model under chosen illumination spectrum, we first pre-compute corresponding spectral reflectance and transmittance profiles for epidermis and dermis separately. We use a dipole model for dermis and a multipole model for epidermis as suggested by [DJ06]. We then convolve these spectral profiles according to Kubelka-Munk formula [DJ05] and store the overall radial reflectance profile function in linear RGB color space. At each point on the surface, PBRT framework extracts the corresponding precomputed spectral reflectance profile from the tabulated set and integrates it over a given radial distance and performs the same operation for all other sampled points on the geometry to add contribution from different spectral profiles, thereby rendering heterogeneous subsurface scattering (more details in Supplemental material). While we convert the spectral profile contributions after integration to RGB within PBRT, it is possible to pre-compute the profiles in sRGB space for usage in a standard RGB rendering pipeline.
(79)
(80)
(81) Note that the facial geometry and spectral measurements for the face renderings were acquired using a multiview capture setup (9 DSLR cameras) placed around the LED sphere. For rendering subsurface scattering in PBRT v2, we had to significantly downsample the mesh vertices, thereby rendering a smooth base mesh seen in
(82) 7.2. Limitations and Discussion
(83) The employed 4D skin BSSRDF model is well suited to reconstruct the appearance of skin and facial hair but has limitations and cannot well reconstruct the appearance of dominant veins or tattoos in skin (see Supplemental material). This is because veins and tattoos cannot be modeled with melanin and hemoglobin concentrations [DWd*08]. We currently do not model any fluorescence in skin, although our broadband measurements likely include some effects of dermal fluorescence [GZAK00]. Through our analysis, we show that parameter estimation using just a single broadband illumination is possible, although suboptimal. The quality and structural physiological correlation of the estimated parameters increases with multiple measurements. In this respect, our proposed two shot acquisition with the complementary spectral illumination conditions is a practical middle ground between the highly accurate measurements possible with detailed spectral imaging as demonstrated by [DWd*08] (also employed by the Antera camera), and just single broadband illumination previously employed for practical measurements of faces. Our choice of illuminants is also a function of the LEDs available in our facial capture setup, and the optimal choices may vary slightly for other spectral illumination setups. However, we demonstrate through our analysis the general trend of the desired illumination conditions for skin measurements, and a method for approximating desirable D65 spectrum using combination of available LEDs. Note that an LED sphere is not a strict requirement for the measurements since the method only requires uniform illumination, and hence can be adapted for many facial capture setups. When employing the Antera camera, we are rather restricted to only using the processed outputs of its proprietary software which does not allow access to the raw data of the spectral measurements for further analysis. However, we demonstrate how to adapt black-box measurements from such a custom dermatological scanning device for realistic rendering.
(84) For the LED sphere measurements, we currently estimate parameters assuming uniform illumination with no occlusions. However, for faces there is partial ambient occlusion around eye sockets and the nose that is baked into the measurements and by extension in our estimated parameter maps. We also do not currently explicitly account for the effect of exitant Frensel in these measurements which slightly affects the parameters estimated for surfaces seen at a grazing angle (most visible in the β.sub.m map).
(85) This is not really a problem for our 3D renderings which employ data seen from 9 different viewpoints and hence the facial parameter maps in the UV parameterization of the geometry is composed of mostly near normal incidence estimates. Finally, our look-up table based reconstructions reproduce the coloration of albedo texture under uniform illumination, but do not model lateral scattering of light within skin which requires explicit rendering of subsurface scattering.
8. CONCLUSION
(86) In summary, we have presented novel practical spectral measurements of skin reflectance using both a dedicated spectral illumination setup (LED sphere) and an off-the-shelf skin measurement device (Antera3D), and employed them to drive a spectral skin BSSRDF model with appropriate complexity to match the appearance of real skin. In this respect, our main contribution is proposing a sweet spot both for measurement and data-driven modeling complexity for reproducing the appearance of skin, including human faces. Our additional contributions include investigating desirable illumination spectra realizable with common LEDs, practical analysis of the gamut limitations of regular RGB color cameras for measuring response to narrow band LED illumination, and proposing a novel indirect measurement protocol that overcomes the gamut limitation and achieves improved spectral isolation compared to direct measurement with a color camera. We also demonstrate how to adapt practical hand-held physiological measurements from a dermatological skin measurement device to our application of realistic rendering which can have a significant impact for dermatological visualizations. Additionally, we also demonstrate how neural networks can be employed for much more efficient parameter estimation and spectral reconstructions given various types of measurements. We see this as a promising initial step towards efficient machine-learning based spectral skin rendering and diagnostics. We demonstrate highly realistic reconstructions of skin with our approach, including renderings of human faces using a biophysically based skin BSSRDF that are, for the first time, comparable to photographs. Future work in this direction could investigate practical measurements and modeling of changes in skin parameters due to skin dynamics or physiological factors of interest for medical diagnostics, as well as applications of any skin products.
Appendix A: D65 Metamer
(87) We address the problem of reproducing a desired spectral illumination (D65) using the LEDs in our LED sphere relying on the faithful reproduction of the appearance of a color chart. We first acquire a set of images of the color chart with known reflectance, individually under each of the n=6 LEDs. The measurements are taken at 3 different LED intensities, under uniform spherical illumination. Given the knowledge of the camera spectral sensitivity CSS, measured with a monochromator, this data allows us to recover the per-channel (ch) non-linear response of the camera sensor γ.sub.ch, and the LEDs relative intensities α.sub.k.
(88) This is similar to the approach of LeGendre et al. [LYL*16], except that instead of focusing only on the color chart appearance to a given camera, we also aim at maximizing the faithfulness of perceived color appearance to a human observer. The reason for simultaneously accounting for a percetual metric besides camera sensitivity in our optimization is that the Chromatic Adaptation Transform (Appendix B)
(89)
employed for spectral isolation of narrow-band response, and look-up table search for the spectral parameters are all based on perceptual metrics. In order to find a spectrum which satisfies the above, given the weights w.sub.c and w.sub.h for the camera and perceptual terms respectively, we need to find a set of coefficients α*.sub.k, for the n=6 LEDs which minimizes the following equation:
where DLj, ch are the per-channel digital levels of the patch j of the color chart, Lab.sub.D65,j are the CIELab values of the color chart patches under D65 illumination, XYZ.fwdarw.Lab is a standard conversion using the D65 reference white [WS82], and [X.sub.j,α′.sub.
(90)
Appendix B: Chromatic Adaptation Transform
(91) The Chromatic Adaptation Transform (CAT) of an image CAT (I.sub.III) is computed in the CIE XYZ tristimulus values space, where its predicted [X.sup.D65Y.sup.D65Z.sup.D65].sup.T values under D65 illumination are derived as follows:
(92)
(93) In the above, [X.sub.w.sup.IIIY.sub.w.sup.IIIZ.sub.w.sup.III].sup.T and [X.sub.w.sup.D65Y.sub.w.sup.D65Z.sub.w.sup.D65].sup.T respectively represent the tristimulus values of the source III and D65 illuminants. The 3×3 matrix M.sub.CAT models human color perception at the LMS cone response level [BS10b], and is derived by numerical optimization [BS10a]. Please note that, given the typical overlapping design of camera sensors spectral sensitivity, the red and green channels of I.sub.nb might contain non-zero signal, similarly to a direct acquisition of a photograph under narrow band illumination.
Appendix C: Estimating Antera's LED Spectrum
(94) The lack of control over Antera's LEDs, both in terms of switching sequence and speed, makes it difficult to directly measure the lighting spectra using a spectrometer given the typical integration time of over 1 second per measurement. Moreover, the individual LEDs have different orientations, thus requiring an integrating sphere for accurate spectral measurements. Hence, we opted for spectral recovery by means of a Genetic Algorithm (GA) based optimization by providing as input to our algorithm a sequence of photographs of a standard reference colour chart, with Np=24 colour patches of known spectral reflectance R.
(95) Our GA begins with a random guess for the Antera's LEDs spectra S: the optimization is driven by a fitness function ƒ, which measures the differences between the acquired digital levels and the simulated ones according to a standard camera model:
(96)
where DG.sub.n,ch is the acquired digital level of the ch channel of the colour patch n and γ.sub.ch models the per-channel non-linear response of the sensor.
(97) The above equation implies the joint estimation of the incident spectra S and the Antera's spectral sensitivity CSS. Hence, a naïve implementation would be extremely underconstrained. However, the specifications of the Antera camera report that it can be used as a colorimeter, thus implying that the spectral sensitivity must fulfill Luther's condition (i.e. it is a linear transformation of the CIE 1931 2-degree Colour Matching Functions), greatly reducing the search space. Furthermore, we enforce the recovery of the LEDs spectra by augmenting ƒ(S,CSS) with a smoothness constraint:
(98)
where K is a scale constant related to the desired resolution in nm of S and CSS; a and S respectively indicates standard deviation and derivative.
Supplementary Material
S1. Details of the Spectral Skin BSSRDF Model
(99) Various terms, parameters and coefficients involved in the spectral skin BSSRDF model of [JSB*10] employed in this work are tabulated along with their descriptions in Table 1. The BSSRDF employs a multipole model for scattering in the thin epidermis, and a dipole model for scattering in the thicker dermis. To model subsurface scattering using the multipole and dipole formulations for epidermis and dermis layers, we need to first compute their absorption and scattering coefficients.
(100) The wavelength λ dependent spectral absorption coefficient for the epidermal layer is given as:
σ.sub.a.sup.epi=C.sub.m[β.sub.mσ.sub.a.sup.em(λ)+(1−β.sub.m)σ.sub.a.sup.pm(λ)]+C.sub.he[γσ.sub.a.sup.oxy(λ)+(1−γ)σ.sub.a.sup.deoxy(λ)]+(1−C.sub.m−C.sub.he)σ.sub.a.sup.base, (S1)
where, the absorption coefficients for eumelanin and pheomelanin (two types of melanin in skin) is computed as:
σ.sub.a.sup.em(λ)=6.6×10.sup.10×λ.sup.−3.33 mm.sup.−1, (S2)
σ.sub.a.sup.pm(λ)=2.9×10.sup.14×λ.sup.−4.75 mm.sup.−1, and (S3)
the baseline absorption coefficient σ.sup.base for the cellular matrix is defined as:
σ.sub.a.sup.base(λ)=0.0244+8.53e.sup.−(λ−154)/66.2 mm.sup.−1, (S4)
Note that X in above equations is defined in nanometers. The absorption coefficients σ.sub.a.sup.oxy and σ.sub.a.sup.deoxy for the oxygenated and deoxygenated hemoglobin are borrowed from measurements provided in medical literature [DJ06].
(101) Similar to epidermis, the absorption coefficient for the dermal layer is defined as:
σ.sub.a.sup.derm(λ)=C.sub.h(γσ.sub.a.sup.oxy(λ)+(1−γ)σ.sub.a.sup.deoxy(λ))+(1−C.sub.h)σ.sub.a.sup.base(λ). (S5)
(102) Next, the reduced scattering coefficient for the dermis is computed as:
σ′.sub.s.sup.derm(λ)=7.37λ.sup.−0.22+1.1×10.sup.11×λ.sup.−4, and (S6)
the reduced scattering coefficient for the epidermis is given by:
σ′.sub.s.sup.derm(λ)=14.74λ.sup.−0.22+2.2×10.sup.11×λ.sup.−4. (S7)
Using σ.sub.a and σ.sub.s′ as absorption and reduced scattering coefficients for the dipole formulation for dermis, its reflectance profile can be computed as explained by Donner and Jensen [DJ05].
(103) Similarly, they also explain how transmittance and reflectance profiles for the epidermis can be computed using its absorption and reduced scattering coefficients with a multipole model. These individual profiles are then convolved to compute the net reflectance profile which is then subject to surface integration to compute diffuse albedo observed due to subsurface scattering in skin. We refer the reader to Donner&Jensen [DJ06] for further details.
(104)
(105) S2. Measurements with LED Sphere
(106)
(107) Given the 6 types of LEDs in our LED sphere, we prefer to create an ideal broadband illumination by computing a weighted combination of all 6 LEDs to create a D65 metamer spectra (D65) which we employ for our broadband measurements. We notice a higher contrast in skin color, particularly coloration due to skin pigmentation and redness, under the D65 metamer illumination compared to any of the individual white LEDs including W57. This is consistent with our observation that D65 spectrum (blue dominant) balances the somewhat skewed red-dominant reflectance spectrum of human skin (see
(108)
(109)
(110) For baseline measurements in the LED sphere, we also did an analysis of which type of broadband illumination is most suitable for estimating the spectral parameters of skin. Across four different skin types ranging from Caucasian, Mediterranean, Asian, and South Asian, we consistently found the reconstruction accuracy of estimation using the cold spectrum broadband illumination (W57) to be higher for reconstructing the appearance of skin under both colder and warmer broadband spectrums. And we also found a clear ordering in decreasing order of accuracy for generalization to a different illumination spectrum from W57, followed by W40, and then W27. This is why we selected the W57 as the choice for the baseline measurement. Measurements under each of these broadband condition were most accurate for reproducing the appearance under their own spectral conditions, pointing to some overfitting to the measurement spectrum. This issue is mitigated to quite an extent when we employ the D65 metamer illumination for broadband measurements, improving the generalization to a different illumination spectrum.
(111) S3. Measurements with Antera3D
(112)
(113) S4. Additional Results and Rendering
(114)
(115) S4.1. Rendering Subsurface Scattering
(116) In order to render heterogeneous subsurface scattering, we modified the provided subsurface scattering implementation in PBRT from the default dipole diffusion kernel to our specified spatially varying profiles implementing two-layered diffusion. We precompute and store the overall radial reflectance profile due to two-layered diffusion per surface point in linear RGB color space. PBRT framework identifies chromophore parameters mapped to that location and extracts the corresponding precomputed reflectance profile from the tabulated set of all sampled profiles generated with the coloration model for the illumination spectrum. Thereafter, PBRT integrates the selected reflectance profile over radial distances and performs the same operation for all other sampled points on the geometry to add contribution from different spectral profiles, thereby rendering heterogeneous subsurface scattering. The pipeline was implemented by modifying Diffusion-Reflectance structure from the dipole subsurface integrator, which reads in chromophore fractions and finds the index of spectral reflectance profile within the precomputed tabulated set. Then at runtime, for each generated radial distance from the current point on the surface the distance function will extract the radial profile for the closest sampled point and add it to the overall color contribution, thus integrating the spectral reflectance profile according to geometry and spatial variation in chromophores.
(117)
(118)
(119) S5. Limitation
(120) The employed 4D skin BSSRDF model is well suited to reconstruct the appearance of skin and facial hair but has limitations and cannot well reconstruct the appearance of dominant veins or tattoos in skin. An example of this can be seen in
(121) Modifications
(122) It will be appreciated that various modifications may be made to the embodiments hereinbefore described. Such modifications may involve equivalent and other features which are already known in the processing of images and/or in the design, manufacture and use of image illuminating and/or capturing apparatuses and component parts thereof and which may be used instead of, or in addition to, features already described herein. Features of one embodiment may be replaced or supplemented by features of another embodiment.
(123) Although claims have been formulated in this application to particular combinations of features, it should be understood that the scope of the disclosure of the present invention also includes any novel features or any novel combination of features disclosed herein either explicitly or implicitly or any generalization thereof, whether or not it relates to the same invention as presently claimed in any claim and whether or not it mitigates any or all of the same technical problems as does the present invention. The applicants hereby give notice that new claims may be formulated to such features and/or combinations of such features during the prosecution of the present application or of any further application derived therefrom.