Spelling suggestions: "subject:"sparsity constraints"" "subject:"sparsity eonstraints""
1 |
Multiframe Superresolution Techniques For Distributed Imaging SystemsShankar, Premchandra M. January 2008 (has links)
Multiframe image superresolution has been an active research area for many years. In this approach image processing techniques are used to combine multiple low-resolution (LR) images capturing different views of an object. These multiple images are generally under-sampled, degraded by optical and pixel blurs, and corrupted by measurement noise. We exploit diversities in the imaging channels, namely, the number of cameras, magnification, position, and rotation, to undo degradations. Using an iterative back-projection (IBP) algorithm we quantify the improvements in image fidelity gained by using multiple frames compared to single frame, and discuss effects of system parameters on the reconstruction fidelity. As an example, for a system in which the pixel size is matched to optical blur size at a moderate detector noise, we can reduce the reconstruction root-mean-square-error by 570% by using 16 cameras and a large amount of diversity in deployment.We develop a new technique for superresolving binary imagery by incorporating finite-alphabet prior knowledge. We employ a message-passing based algorithm called two-dimensional distributed data detection (2D4) to estimate the object pixel likelihoods. We present a novel complexity-reduction technique that makes the algorithm suitable even for channels with support size as large as 5x5 object pixels. We compare the performance and complexity of 2D4 with that of IBP. In an imaging system with an optical blur spot matched to pixel size, and four 2x2 undersampled LR images, the reconstruction error for 2D4 is 300 times smaller than that for IBP at a signal-to-noise ratio of 38dB.We also present a transform-domain superresolution algorithm to efficiently incorporate sparsity as a form of prior knowledge. The prior knowledge that the object is sparse in some domain is incorporated in two ways: first we use the popular L1 norm as the regularization operator. Secondly we model wavelet coefficients of natural objects using generalized Gaussian densities. The model parameters are learned from a set of training objects and the regularization operator is derived from these parameters. We compare the results from our algorithms with an expectation-maximization (EM) algorithm for L1 norm minimization and also with the linear minimum mean squared error (LMMSE) estimator.
|
2 |
Phase Retrieval with Sparsity ConstraintsLoock, Stefan 07 June 2016 (has links)
No description available.
|
3 |
Real-time MRI and Model-based Reconstruction Techniques for Parameter Mapping of Spin-lattice RelaxationWang, Xiaoqing 18 October 2016 (has links)
No description available.
|
4 |
The impact of a curious type of smoothness conditions on convergence rates in l1-regularizationBot, Radu Ioan, Hofmann, Bernd 31 January 2013 (has links) (PDF)
Tikhonov-type regularization of linear and nonlinear ill-posed problems in abstract spaces under sparsity constraints gained relevant attention in the past years. Since under some weak assumptions all regularized solutions are sparse if the l1-norm is used as penalty term, the l1-regularization was studied by numerous authors although the non-reflexivity of the Banach space l1 and the fact that such penalty functional is not strictly convex lead to serious difficulties. We consider the case that the sparsity assumption is narrowly missed. This means that the solutions may have an infinite number of nonzero but fast decaying components. For that case we formulate and prove convergence rates results for the l1-regularization of nonlinear operator equations. In this context, we outline the situations of Hölder rates and of an exponential decay of the solution components.
|
5 |
The impact of a curious type of smoothness conditions on convergence rates in l1-regularizationBot, Radu Ioan, Hofmann, Bernd January 2013 (has links)
Tikhonov-type regularization of linear and nonlinear ill-posed problems in abstract spaces under sparsity constraints gained relevant attention in the past years. Since under some weak assumptions all regularized solutions are sparse if the l1-norm is used as penalty term, the l1-regularization was studied by numerous authors although the non-reflexivity of the Banach space l1 and the fact that such penalty functional is not strictly convex lead to serious difficulties. We consider the case that the sparsity assumption is narrowly missed. This means that the solutions may have an infinite number of nonzero but fast decaying components. For that case we formulate and prove convergence rates results for the l1-regularization of nonlinear operator equations. In this context, we outline the situations of Hölder rates and of an exponential decay of the solution components.
|
Page generated in 0.0735 seconds