PowerPoint - People

Download Report

Transcript PowerPoint - People

Joint Noise Level Estimation from Personal Photo Collections
YiChang
1*,2
Shih
1
Kwatra
Vivek
Troy
1Google Research
1
Chinen
Hui
2MIT CSAIL
1
Fang
Sergey
*Internship work at Google
Goal
Contributions
 Given a set of face images from the same person, taken
under different lighting and cameras, estimate the noise
levels in each image
 Key observation: given two noisy images, the noise
levels are correlated if they share the same underlying
2
2
image content, since 𝜎1 βˆ’ 𝜎2 = π‘£π‘Žπ‘Ÿ[𝑰𝒏,𝟏 ] βˆ’ π‘£π‘Žπ‘Ÿ[𝑰𝒏,𝟐 ]
Overview
Starting from a face image collection:
 Preprocess: geometrically and photometrically align the
images with affine transform and color match
 We formulate the estimation as maximizing the joint
probability distribution between all images’ noise levels
 The joint distribution is conditioned on the pair-wise
2
2
relative noise levels {πœŒπ‘–π‘— |πœŒπ‘–π‘— β‰œ πœŽπ‘– βˆ’ πœŽπ‘— }. We use a twostage optimization that first estimates {πœŒπ‘–π‘— }, then {πœŽπ‘– }
 𝑰𝑛 = π‘°π‘œπ‘Ÿπ‘–π‘” + 𝒏, i.i.d, zero mean. 𝜎 = noise level β‰œ 𝑠𝑑𝑑[𝒏]
 This is difficult because we cannot decouple 𝒏 from 𝑰𝑛
1
Ioffe
 Two-stage optimization:
οƒ˜ Estimating {πœŒπ‘–π‘— }: We take a patch-based method. We
first find the patch correspondence between π‘°π’Š and 𝑰𝒋 ,
βˆ—
then find the best estimated relative noise {πœŒπ‘–π‘— } from
the patch pairs.
βˆ—
2
2
βˆ—
οƒ˜ With {πœŒπ‘–π‘— }, estimate {πœŽπ‘– } by constraining πœŽπ‘– βˆ’ πœŽπ‘— = πœŒπ‘–π‘—
Pair-wise Relative Noise πœŒπ‘–π‘— Estimation
Results
Ground Truth Experiment and Comparison
 The two faces are not perfectly aligned
 We show one example below with estimated
noise levels and denoised result using BM3D +
our method for noise parameter
 Add synthetic Gaussian noise with different parameters
 Compare estimated noise levels and denoised result by BM3D
q
β€’ π‘π‘π‘ž = exp(βˆ’πœ…π‘π‘ž 𝒑1𝑝 βˆ’ 𝒑2π‘ž
2
Input
BM3D
Input
y=x Line (Ground Truth)
Liu et. al. (Mean)
Metric-Q (Mean)
Our Method (Mean)
BM3D
Mean PSNR (Metric-Q)
Mean PSNR (Our Method)
Mean PSNR(Best BM3D)
30
I2
βˆ—
𝜌12
=
Οƒ=23.2
Οƒ=14.5
𝑝,π‘ž π‘π‘π‘ž πœπ‘π‘ž
𝑝,π‘ž π‘π‘π‘ž
), confidence that (𝑝, π‘ž) is a true correspondence
26
40.00
22
18
14
10
6
 More subjects
38.00
37.00
36.00
35.00
33.00
2
Οƒ=38.2
39.00
34.00
2
Οƒ=23.9
β€’ For computational efficiency, we selected the best 5 π‘ž s for each 𝑝
41.00
Mean PSNR
I1
 Compute pair-wise relative noise by aggregating πœπ‘π‘ž :
q
Estimated Noise Sigma
 We break down the image into patches, and
estimate the patch-wise relative noise
levels πœπ‘π‘ž by πœπ‘π‘ž β‰œ π‘£π‘Žπ‘Ÿ[𝒑1𝑝 ] βˆ’ π‘£π‘Žπ‘Ÿ[𝒑2π‘ž ]
p
7
12
0.00
17
Synthetic noise
10.00
15.00
20.00
Mean Noise Sigma
True Noise Sigma
Clean image
5.00
Our method
Metric-Q
Best BM3D
Absolute Noise Level Estimation with Global Optimization
 We estimate {πœŽπ‘– } conditioning on

2
πœŽπ‘–
2
𝑖≠𝑗 𝑀𝑖𝑗 πœŽπ‘–
βˆ—
{πœŒπ‘–π‘— }
2
πœŽπ‘—
=argmin
βˆ’
βˆ’
𝑀𝑖𝑗 : similarity between two faces
βˆ— 2
πœŒπ‘–π‘—
User Study
2
πœŽπ‘–
βˆ—
πœŒπ‘–π‘—
 Solving a linear system
2
πœŽπ‘—
 The system is under-determined, up to adding
a constant number.
- option 1: assign some images to be zero noise
- option 2: assuming the collection contains clean
images, assign the least noisy one to be zero. We use this one for evaluations
 Based on BM3D denoised result, decide which
one is preferable
 Ran on 71 images, each is evaluated by 3 users
Input image
Our method
Metric-Q
35%
24%
41%
Οƒ=46.3
Οƒ=35
Metric-Q
Οƒ
PSNR
19.25
22.33
20.31
34.65
27
34.53
23
34.65
Selected References
C. Liu, R. Szeliski, S. Kang, C. Zitnick, and W. Freeman. Automatic estimation and removal of
noise from a single image. IEEE Transactions on Pattern Analysis and Machine Intelligence,
30(2), 2008
X. Zhu and P. Milanfar. Automatic parameter selection for denoising algorithms using a noreference measure of image content. IEEE Transactions on Image Processing, 19(12), 2010.
Acknowledgements
Ours
Equally good
Οƒ=9.1
Οƒ=27
We thank MIT Graphics and Vision group for helpful discussion. We would like to thank the
volunteers who participated in the user study.