Anda belum login :: 27 Nov 2024 19:38 WIB
Home
|
Logon
Hidden
»
Administration
»
Collection Detail
Detail
Unifying Probabilistic and Variational Estimation
Oleh:
Hamza, A. B.
;
Krim, H.
;
Unal, G. B.
Jenis:
Article from Bulletin/Magazine
Dalam koleksi:
IEEE Signal Processing Magazine vol. 19 no. 5 (2002)
,
page 37-47.
Topik:
estimation
;
unifying
;
probabilistic
;
variational estimation
Ketersediaan
Perpustakaan Pusat (Semanggi)
Nomor Panggil:
SS26.6
Non-tandon:
1 (dapat dipinjam: 0)
Tandon:
tidak ada
Lihat Detail Induk
Isi artikel
A maximum a posteriori (MAP) estimator using a Markov or a maximum entropy random field model for a prior distribution may be viewed as a minimizer of a variational problem.Using notions from robust statistics, a variational filter referred to as a Huber gradient descent flow is proposed. It is a result of optimizing a Huber functional subject to some noise constraints and takes a hybrid form of a total variation diffusion for large gradient magnitudes and of a linear diffusion for small gradient magnitudes. Using the gained insight, and as a further extension, we propose an information - theoretic gradient descent flow which is a result of minimizing a functional that is a hybrid between a negentropy variational integral and a total variation. Illustrating examples demonstrate a much improved performance of the approach in the presence of Gaussian and heavy tailed noise. In this article, we present a variational approach to MAP estimation with a more qualitative and tutorial emphasis. The key idea behind this approach is to use geometric insight in helping construct regularizing functionals and avoiding a subjective choice of a prior in MAP estimation. Using tools from robust statistics and information theory, we show that we can extend this strategy and develop two gradient descent flows for image denoising with a demonstrated performance.
Opini Anda
Klik untuk menuliskan opini Anda tentang koleksi ini!
Kembali
Process time: 0.015625 second(s)