2019
DOI: 10.1016/j.image.2019.05.013
|View full text |Cite
|
Sign up to set email alerts
|

DeepFlash: Turning a flash selfie into a studio portrait

Abstract: We present a method for turning a flash selfie taken with a smartphone into a photograph as if it was taken in a studio setting with uniform lighting. Our method uses a convolutional neural network trained on a set of pairs of photographs acquired in an ad-hoc acquisition campaign. Each pair consists of one photograph of a subject's face taken with the camera flash enabled and another one of the same subject in the same pose illuminated using a photographic studio-lighting setup. We show how our method can ame… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
23
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
2
1

Relationship

1
5

Authors

Journals

citations
Cited by 18 publications
(23 citation statements)
references
References 48 publications
(46 reference statements)
0
23
0
Order By: Relevance
“…Enhancement of low-illuminated areas (red), and estimation of natural skin and air tone of people (green). We compare with SRIE (Fu et al, 2016), LIME (Guo et al, 2017), and DeepFlash (Capece et al, 2019). we used a reduced set of the entire FAID dataset for our experiments. Finally, our custom dataset has 969 pairs of images for training and 116 for testing and all images were resized to 320 × 240 or 240 × 320 depending on their orientation.…”
Section: Methodsmentioning
confidence: 99%
See 4 more Smart Citations
“…Enhancement of low-illuminated areas (red), and estimation of natural skin and air tone of people (green). We compare with SRIE (Fu et al, 2016), LIME (Guo et al, 2017), and DeepFlash (Capece et al, 2019). we used a reduced set of the entire FAID dataset for our experiments. Finally, our custom dataset has 969 pairs of images for training and 116 for testing and all images were resized to 320 × 240 or 240 × 320 depending on their orientation.…”
Section: Methodsmentioning
confidence: 99%
“…The encoder part of the network represents the first 13 convolutional blocks of the VGG-16 (Simonyan and Zisserman, 2015), and the weights of the encoder are initialized with a pre-trained model for face-recognition (Parkhi et al, 2015). (Capece et al, 2019) implemented a pre-processing step before feed the inputs and targets to their architecture, each pair of images are filtered. After this step the encoder-decoder architecture is trained to learn a lowfrequency relationship between the flash and the ambient image.…”
Section: Image-to-image Translationmentioning
confidence: 99%
See 3 more Smart Citations