• No results found

Deep HDR Estimation with Generative Detail Reconstruction

N/A
N/A
Protected

Academic year: 2022

Share "Deep HDR Estimation with Generative Detail Reconstruction"

Copied!
7
0
0

Laster.... (Se fulltekst nå)

Fulltekst

(1)

EUROGRAPHICS 2021 / N. Mitra and I. Viola (Guest Editors)

(2021),

Deep HDR estimation with generative detail reconstruction:

Supplementary Materials

Y. Zhang and T. O. Aydın DisneyResearch|Studios

In this supplementary material, we provide further implementa- tion and network architecture details. Additional visual results are shown for a more thorough ablation study as well as comparison between our results and the current state-of-the-art.

1. Additional Objective evaluations

In Table1, we report additional objective evaluations that utilize the perceptually uniformity (PU) encoding. Our proposed method also outperforms existing HDR reconstruction methods in terms of PU-PSNR and PU-SSIM.

Method PU-PSNR PU-MS-SSIM

mean σ mean σ

HDRCNN [EKD17] 44.4923 5.5717 0.9343 0.0973

DrTMO [EKM17] 33.7938 1.9358 0.8519 0.1133

ExpandNet [MBRHD18] 35.6328 5.1561 0.8660 0.1161 Santos et al. [STKK20] 45.8392 6.8725 0.9483 0.1867

Baseline

B-net only 43.3657 2.1593 0.9226 0.1145 D-net only 42.9325 2.0148 0.9002 0.1127 R-net only 43.7787 1.9572 0.9126 0.1132

(B+Dpart.)-net 44.8475 2.4824 0.9367 0.1213

(B+D)-net 45.2573 2.1467 0.9432 0.1144 Proposed (B+D+R)-net 47.6317 5.9907 0.9655 0.1082

Table 1: PU encoded HDR reconstruction performance of our method compared to various baselines and state-of-the-art meth- ods.

2. Network architectures

We report our implementation details of the network architectures.

For simplicity, we denote the layers with K (kernel size), D (dila- tion), S (stride size) and C (number of channels).

Base layer reconstruction network & Refinement network

Encoder:K3S1C64 - K3S1C128 - K3S1C256 - K3S1C512 - K3S1C512

Decoder: K3S1C256 - K3S1C128 - K3S1C64 - K3S1C64 - K3S1C3 - clip

Detail reconstruction network is constructed by two stages stacked together. The first stage is convolutional network with par- tial convolutional layers. Nearest upsampling (NU) has been em- ployed in the decoder of this network and skip connections have been added from each encoder layer to its corresponding decoder layer. The second stage, a contextual inpainting network contains two parallel encoders: dilated convolutional branch and contextual inpainting branch. The output of the two parallel encoders are con- catenated together to feed into the decoder. All convolutional layers except the last one are followed by ReLU activation function.

Stage 1 Encoder: K7S2C64 - K5S2C128 - K5S2C256 - [K3S2C512]×3.

Stage 1 Decoder:[K3S1C512]×2 - K3S1C256 - K3S1C128 - K3S1C64 - K3S1C3 - K3S1C3 - sigmoid.

Stage 2 Dilated Convolutional Branch: K5S1D1C32 - K3S2D1C32 - K3S1D1C64 - K3S2D1C64 - K3S1D1C128 - K3S1D1C128 - K3S1D2C128 -

K3S1D4C128 - K3S1D8C128 - K3S1D16C128 - concat.

Stage 2 Contextual Attention Branch: K5S1D1C32 - K3S2D1C32 - K3S1D1C64 - K3S2D1C128 - K3S1D1C128 - K3S1D1C128 - ReLU - contextual attention layer [YLY18] - K3S1D1C128 - K3S1D1C128 - concat.

Stage 2 Decoder K3S1D1C128 - K3S1D1C128 - NU(2×) - K3S1D1C64 - K3S1D1C64 - UN(2×) - K3S1D1C32 - K3S1D1C16 - K3S1D1C3 - clip.

(2)

3. Additional ablation examples

3.1. Detail reconstruction network with & without contextual detail inpainting component

Figure1and2shows example results of the detail reconstruction network with and without the contextual detail inpainting compo- nent.

4. Additional comparisons with other methods

Additional examples of the comparisons between our model and other existing methods can be found in Figure3and4.

(3)

Detail layer

Input Detail layer Output w/

Part. Conv. only Detail layer Output w/

(Part. + cont.) inpainting Detail layer reference

Figure 1:Example results of the detail reconstruction network where the network with and without the contextual detail inpainting subnet- work component.

(4)

Detail layer

Input Detail layer Output w/

Part. Conv. only Detail layer Output w/

(Part. + cont.) inpainting Detail layer reference

Figure 2:Example results of the detail reconstruction network where the network with and without the contextual detail inpainting subnet- work component.

(5)

Input SDR Image (b) Input SDR (c) DrTMO (d) ExpandNet (e) HDRCNN (f) Ours (g) HDR Ref. (h) HDR-VDP-2 Ours vs HDR Ref.

(6)

Input SDR Image (b) Input SDR (c) DrTMO (d) ExpandNet (e) HDRCNN (f) Ours (g) HDR Ref. (h) HDR-VDP-2 Ours vs HDR Ref.

Figure 4:Additional comparison of our method’s results with state-of-the-art techniques. Input SDR image (a) and corresponding crops (b) are shown on the right. We present results from DrTMO [EKM17] (c), ExpandNet [MBRHD18] (d), HDRCNN [EKD17] (e) and our model (f). The reference HDR is shown in (g), and HDR-VDP-2 visible difference map is given in (h).

(7)

References

[EKD17] EILERTSEN G., KRONANDER J., DENES G., MANTIUK R. K., UNGERJ.: HDR image reconstruction from a single exposure using deep CNNs.ACM Transactions on Graphics 36, 6 (2017).1,5,6 [EKM17] ENDOY., KANAMORIY., MITANIJ.: Deep Reverse Tone

Mapping.ACM Trans. Graph 36, 6 (2017).1,5,6

[MBRHD18] MARNERIDES D., BASHFORD-ROGERS T., HATCHETT J., DEBATTISTAK.: ExpandNet: A Deep Convolutional Neural Net- work for High Dynamic Range Expansion from Low Dynamic Range Content.arXiv:1803.02266 [cs](Mar. 2018). arXiv: 1803.02266.1,5, 6

[STKK20] SANTOSM. S., TSANGR., KHADEMIKALANTARIN.: Sin- gle image hdr reconstruction using a cnn with masked features and per- ceptual loss. ACM Transactions on Graphics 39, 4 (7 2020). doi:

10.1145/3386569.3392403.1

[YLY18] YUJ., LINZ., YANGJ., SHENX., LUX., HUANGT. S.:

Generative image inpainting with contextual attention. InProceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2018), pp. 5505–5514.1

Referanser

RELATERTE DOKUMENTER

In this paper, the offline identification is considered for two specific purposes. Firstly, based on the offline iden- tification results, the desired set of frequency points to

One focus group member expressed a radical attitude concerning the people working as social workers who do not have training in social work: […] … Everybody working in this

Flexibility exercise training for adults with fibromyalgia (Review) 24.. two parallel arms, and one had three parallel arms. Owing to the differences between interventions

The industrial heat demand will vary, and hence, running the heat pump model with the best mixture for the two different source temperature conditions at different load

points that are close to regions in the point set with insufficient sampling, and then by copying appropriate local neighbourhood patches (so-called fragments) from a candidate set

The successful coupling of these two different visualisation systems as well as the benefit of a fast optical network for parallel data access and for distributed collaboration could

Figure 1: We propose a novel generative approach for image inpainting by incorporating the semantic information through conditional feature modulation and using dual discriminators

Semantic-Aware Generative Approach for Image Inpainting. Deepankar Chanda and Nima