

Author: Ruggiero Valeria Serafini Thomas Zanella Riccardo Zanni Luca
Publisher: Springer Publishing Company
ISSN: 0925-5001
Source: Journal of Global Optimization, Vol.48, Iss.1, 2010-09, pp. : 145-157
Disclaimer: Any content in publications that violate the sovereignty, the constitution or regulations of the PRC is not accepted or approved by CNPIEC.
Abstract
The ability of the modern graphics processors to operate on large matrices in parallel can be exploited for solving constrained image deblurring problems in a short time. In particular, in this paper we propose the parallel implementation of two iterative regularization methods: the well known expectation maximization algorithm and a recent scaled gradient projection method. The main differences between the considered approaches and their impact on the parallel implementations are discussed. The effectiveness of the parallel schemes and the speedups over standard CPU implementations are evaluated on test problems arising from astronomical images.
Related content







