1
Performance Geometry Capture for Spatially Varying Relighting Andrew Jones Andrew Gardner Mark BolasIan McDowallPaul Debevec USC Institute for Creative Technologies University of Southern CaliforniaFakespace LabsIn image-based relighting, novel illumination on a subject is syn- thesized based on images acquired in different basis lighting con- ditions. Most commonly, the basis images of the subject are taken under a variety of different directional lighting conditions. A lin- ear combination of the color channels of the different lighting di- rections is formed to produce the rendering under novel illumina- tion. When the lighting directions are densely distributed through- out the sphere, any distant lighting environment can be simulated accurately. (a) (b) (c) (d) Figure 1: (a) One of 29 basis lighting directions (b) One of 24 struc- tured light patterns (c) Indirect illumination computed from the re- covered geometry and albedo (d) Rendering with spatially varying light including both direct and indirect illumination effects. This standard image-based relighting technique is not capable of simulating spatially-varying incident illumination, such as dappled illumination or partial shadows on the subject, since no basis il- lumination condition shows such effects. [Masselus et al. 2003] expanded the lighting basis to include spatially-varying conditions, but at the expense of greatly increased capture times. Recent work [Wenger et al. 2005] has shown that a traditional lighting basis can be captured for live human performances using a high speed cam- era and time-multiplexed illumination using LEDs. In this sketch, we present a process for simulating spatially-varying illumination on such datasets by including additional structured light patterns from a video projector within the illumination basis. Our processed data is an animated 3D model with both geometric and reflectance information. We show how this augmented dataset can be used to simulate spatially-varying illumination effects in an image-based relighting process. Our geometry and reflectance datasets are captured 24 times per second using a Luxeon V LED-based light stage, a Vision Research high-speed camera, and a specially modified 1024 × 768 pixel high- speed DLP projector. The lights, camera, and projector are syn- chronized at 1500 frames per second. For each 24th of a second, the subject is lit by a series of 29 basis lighting directions (e.g. Fig. 1(a)) followed by 24 structured light patterns (e.g. Fig. 1(b)). We recover surface normals and spatially-varying diffuse color (albedo) using the photometric stereo based techniques described in [Wenger et al. 2005]. A surface mesh of the geometry is recovered from the structured lighting patterns. We use a gradient-descent optimization to refine the geometry to match the estimated surface normals. The normals are also used to partially fill in missing ge- ometry resulting from projector occlusions. The geometric model allows us to determine where each point in the image lies within the three-dimensional volume of a spatially- varying light source, such as a projected stained glass window (Fig. 2). With this information, each point on the face can be scaled as if illuminated by a different lighting environment, giving the ap- pearance of spatially-varying illumination. However, this process does not correctly model second-order illumination effects. For ex- ample, if a spatially-varying light source illuminates just a person’s shoulder, in reality this would produce indirect illumination on the underside of the chin. We have developed a process for simulat- ing such effects by performing a global illumination computation on the 3D model texture-mapped by the recovered albedo. For a given basis lighting direction, we estimate and remove the existing bounced light in the image. We multiply the resulting image by the projected spatially varying illumination. Finally, we add new sim- ulated indirect illumination (Fig. 1(c)). The final rendering com- bines spatially-varying illumination with fill light from a conven- tional light probe (Fig. 1(d)). We have demonstrated our technique on a four-second sequence of an actor’s face, shown in the video. To our knowledge this is the first sequence capturing time-varying geometry and directionally- varying surface reflectance, and the first simulation of indirect illu- mination effects from a traditional image-based relighting dataset. In our current work, we are extending this process to multiple pro- jectors to capture more complete subject geometry and to more accurately simulate subsurface scattering effects resulting from spatially-varying illumination. Figure 2: An HDR image of a stained glass window is used to create a spatially-varying point light source. References MASSELUS, V., PEERS, P., DUTR ´ E, P., AND WILLEMS, Y. D. 2003. Relighting with 4d incident light fields. ACM Transactions on Graphics 22, 3 (July), 613–620. WENGER, A., GARDNER, A., UNGER, J., TCHOU, C., HAWKINS, T., AND DE- BEVEC, P. 2005. Performance relighting and reflectance transformation with time-multiplexed illumination. Conditionally Accepted to ACM Transactions on Graphics (SIGGRAPH 2005).

Performance Geometry Capture for Spatially Varying Relightinggl.ict.usc.edu/Research/SpatialRelighting/spatial_sketch.pdf · Performance Geometry Capture for Spatially Varying Relighting

  • Upload
    others

  • View
    30

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Performance Geometry Capture for Spatially Varying Relightinggl.ict.usc.edu/Research/SpatialRelighting/spatial_sketch.pdf · Performance Geometry Capture for Spatially Varying Relighting

Performance Geometry Capture for Spatially Varying Relighting

Andrew Jones Andrew Gardner Mark Bolas† Ian McDowall‡ Paul DebevecUSC Institute for Creative Technologies University of Southern California† Fakespace Labs‡

In image-based relighting, novel illumination on a subject is syn-thesized based on images acquired in different basis lighting con-ditions. Most commonly, the basis images of the subject are takenunder a variety of different directional lighting conditions. A lin-ear combination of the color channels of the different lighting di-rections is formed to produce the rendering under novel illumina-tion. When the lighting directions are densely distributed through-out the sphere, any distant lighting environment can be simulatedaccurately.

(a) (b)

(c) (d)

Figure 1: (a) One of 29 basis lighting directions (b) One of 24 struc-tured light patterns (c) Indirect illumination computed from the re-covered geometry and albedo (d) Rendering with spatially varyinglight including both direct and indirect illumination effects.

This standard image-based relighting technique isnot capable ofsimulating spatially-varying incident illumination, such as dappledillumination or partial shadows on the subject, since no basis il-lumination condition shows such effects. [Masselus et al. 2003]expanded the lighting basis to include spatially-varying conditions,but at the expense of greatly increased capture times. Recent work[Wenger et al. 2005] has shown that a traditional lighting basis canbe captured for live human performances using a high speed cam-era and time-multiplexed illumination using LEDs. In this sketch,we present a process for simulating spatially-varying illuminationon such datasets by including additional structured light patternsfrom a video projector within the illumination basis. Our processeddata is an animated 3D model with both geometric and reflectanceinformation. We show how this augmented dataset can be used tosimulate spatially-varying illumination effects in an image-basedrelighting process.

Our geometry and reflectance datasets are captured 24 times persecond using a Luxeon V LED-based light stage, a Vision Researchhigh-speed camera, and a specially modified 1024×768 pixel high-speed DLP projector. The lights, camera, and projector are syn-chronized at 1500 frames per second. For each 24th of a second,the subject is lit by a series of 29 basis lighting directions (e.g. Fig.1(a)) followed by 24 structured light patterns (e.g. Fig. 1(b)).

We recover surface normals and spatially-varying diffuse color(albedo) using the photometric stereo based techniques described in[Wenger et al. 2005]. A surface mesh of the geometry is recoveredfrom the structured lighting patterns. We use a gradient-descentoptimization to refine the geometry to match the estimated surfacenormals. The normals are also used to partially fill in missing ge-ometry resulting from projector occlusions.

The geometric model allows us to determine where each point inthe image lies within the three-dimensional volume of a spatially-varying light source, such as a projected stained glass window (Fig.2). With this information, each point on the face can be scaled asif illuminated by a different lighting environment, giving the ap-pearance of spatially-varying illumination. However, this processdoes not correctly model second-order illumination effects. For ex-ample, if a spatially-varying light source illuminates just a person’sshoulder, in reality this would produce indirect illumination on theunderside of the chin. We have developed a process for simulat-ing such effects by performing a global illumination computationon the 3D model texture-mapped by the recovered albedo. For agiven basis lighting direction, we estimate and remove the existingbounced light in the image. We multiply the resulting image by theprojected spatially varying illumination. Finally, we add new sim-ulated indirect illumination (Fig. 1(c)). The final rendering com-bines spatially-varying illumination with fill light from a conven-tional light probe (Fig. 1(d)).

We have demonstrated our technique on a four-second sequence ofan actor’s face, shown in the video. To our knowledge this is thefirst sequence capturing time-varying geometry and directionally-varying surface reflectance, and the first simulation of indirect illu-mination effects from a traditional image-based relighting dataset.In our current work, we are extending this process to multiple pro-jectors to capture more complete subject geometry and to moreaccurately simulate subsurface scattering effects resulting fromspatially-varying illumination.

Figure 2: An HDR image of a stained glass window is used to createa spatially-varying point light source.

References

MASSELUS, V., PEERS, P., DUTRE, P.,AND WILLEMS , Y. D. 2003. Relighting with4d incident light fields.ACM Transactions on Graphics 22, 3 (July), 613–620.

WENGER, A., GARDNER, A., UNGER, J., TCHOU, C., HAWKINS , T., AND DE-BEVEC, P. 2005. Performance relighting and reflectance transformation withtime-multiplexed illumination. Conditionally Accepted to ACM Transactions onGraphics (SIGGRAPH 2005).