Deep Tone Mapping Operator for High Dynamic Range Images
Item Type:Conference Paper
Citation:Rana, A., Singh, P., Valenzise, G., Dufaux, F., Komodakis, N. & Smolic, A., Deep Tone Mapping Operator for High Dynamic Range Images, IEEE Transaction of Image Processing, 29, pp. 1285-1298, 2019, ISBN: 1057-7149., 2019
AR - Deep Tone mAPPING.pdf (PDF) 7.751Mb
A computationally fast tone mapping operator (TMO) that can quickly adapt to a wide spectrum of high dynamic range (HDR) content is quintessential for visualization on varied low dynamic range (LDR) output devices such as movie screens or standard displays. Existing TMOs can successfully tone-map only a limited number of HDR content and require an extensive parameter tuning to yield the best subjective-quality tone-mapped output. In this paper, we address this problem by proposing a fast, parameter-free and scene-adaptable deep tone mapping operator (DeepTMO) that yields a high-resolution and high-subjective quality tone mapped output. Based on conditional generative adversarial network (cGAN), DeepTMO not only learns to adapt to vast scenic-content (e.g., outdoor, indoor, human, structures, etc.) but also tackles the HDR related scene-specific challenges such as contrast and brightness, while preserving the fine-grained details. We explore 4 possible combinations of Generator-Discriminator architectural designs to specifically address some prominent issues in HDR related deep-learning frameworks like blurring, tiling patterns and saturation artifacts. By exploring different influences of scales, loss-functions and normalization layers under a cGAN setting, we conclude with adopting a multi-scale model for our task. To further leverage on the large-scale availability of unlabeled HDR data, we train our network by generating targets using an objective HDR quality metric, namely Tone Mapping Image Quality Index (TMQI). We demonstrate results both quantitatively and qualitatively, and showcase that our DeepTMO generates high-resolution, high-quality output images over a large spectrum of real-world scenes. Finally, we evaluate the perceived quality of our results by conducting a pair-wise subjective study which confirms the versatility of our method.
Author: Smolic, Aljosa; Rana, Aakanksha; Singh, Praveer; Valenzise, Giuseppe; Dufaux, Frederic; Komodakis, Nikos
Other Titles:IEEE Transaction of Image Processing, 29 , pp. 1285-1298, 2019, ISBN: 1057-7149.
Type of material:Conference Paper
Availability:Full text available