It's the long entry of the week that features everything from sensors, template matching to theoretical phase transition like results. But first here is a challenge on Classifying Heart Sounds Challenge by Peter Bentley, Glenn Nordehn, Miguel Coimbra, Shie Mannor, Rita Getz.

Also here is Jason Laska's thesis presentation slides entitled: Regime Change Sampling Rate vs. Bit-Depth in Compressive Sensing

I was wondering when this would happen, it came out at the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML PKDD) and the presentation is Fast approximate text document clustering using Compressive Sampling by Laurence A. F. Park, The abstract reads:

Document clustering involves repetitive scanning of a document set, therefore as the size of the set increases, the time required for the clustering task increases and may even become impossible due to computational constraints. Compressive sampling is a feature sampling technique that allows us to perfectly reconstruct a vector from a small number of samples, provided that the vector is sparse in some known domain. In this article, we apply the theory behind compressive sampling to the document clustering problem using k-means clustering. We provide a method of computing high accuracy clusters in a fraction of the time it would have taken by directly clustering the documents. This is performed by using the Discrete Fourier Transform and the Discrete Cosine Transform. We provide empirical results showing that compressive sampling provides a 14 times increase in speed with little reduction in accuracy on 7,095 documents, and we also provide a very accurate clustering of a 231,219 document set, providing 20 times increase in speed when compared to performing k-means clustering on the document set. This shows that compressive clustering is a very useful tool that can be used to quickly compute approximate clusters.

From 50 hours down to 2, it's not bad, not bad at all.

Efficient Adaptive Compressive Sensing Using Sparse Hierarchical Learned Dictionaries by Akshay Soni, Jarvis Haupt. The abstract reads:

Recent breakthrough results in compressed sensing (CS) have established that many high dimensional objects can be accurately recovered from a relatively small number of non- adaptive linear projection observations, provided that the objects possess a sparse representation in some basis. Subsequent efforts have shown that the performance of CS can be improved by exploiting the structure in the location of the non-zero signal coefficients (structured sparsity) or using some form of online measurement focusing (adaptivity) in the sensing process. In this paper we examine a powerful hybrid of these two techniques. First, we describe a simple adaptive sensing procedure and show that it is a provably effective method for acquiring sparse signals that exhibit structured sparsity characterized by tree-based coefficient dependencies. Next, employing techniques from sparse hierarchical dictionary learning, we show that representations exhibiting the appropriate form of structured sparsity can be learned from collections of training data. The combination of these techniques results in an effective and efficient adaptive compressive acquisition procedure.

Optimal Phase Transitions in Compressed Sensing by Yihong Wu, Sergio Verdu. The abstract reads:

Compressed sensing deals with efficient recovery of analog signals from linear measurements. This paper presents a statistical study of compressed sensing by modeling the input signal as random processes. Three classes of encoders are considered, namely optimal nonlinear, optimal linear and random linear encoders. Focusing on optimal decoders, we investigate the fundamental tradeoff between measurement rate and reconstruction fidelity gauged by error probability and noise sensitivity in the absence and presence of measurement noise, respectively. Optimal phase transition thresholds are determined as a functional of the input distribution and compared to suboptimal thresholds achieved by various popular reconstruction algorithms. In particular, we show that Gaussian sensing matrices incur no penalty on the phase transition threshold with respect to optimal nonlinear encoding. Our results also provide a rigorous justification of previous results based on replica heuristics in the weak-noise regime.

Generalized sampling, inﬁnite-dimensional compressed sensing, and semi-random sampling for asymptotically incoherent dictionaries by Ben Adcock, Anders C. Hansen, Evelyn Herrholz, Gerd Teschke.. The abstract reads:

Recent developments in sampling in abstract Hilbert spaces have led to a new theory of compressed sensing for inﬁnitedimensional signals. In this paper, we continue with this theme by introducing a new type of subsampling for inﬁnite-dimensional sparse recovery problems, known as semi-random sampling. As we demonstrate, this allows for subsampling in problems which previously had not been amenable to more conventional compressed sensing tools. More speciﬁcally, semi-random sampling allows one to overcome the so-called incoherence barrier, which limits the potential for subsampling via standard random sampling techniques. The key to this improvement is a property known as asymptotic incoherence. In the ﬁnal part of this paper we provide speciﬁc estimates for this property in several important cases, and illustrate via numerical example the beneﬁt of semi-random sampling.

Generalized Sampling: Extension to Frames and Inverse and Ill-Posed Problems by by Ben Adcock, Anders C. Hansen, Evelyn Herrholz, Gerd Teschke. The abstract reads:

Generalized sampling is new framework for sampling and reconstruction in inﬁnite-dimensionalHilbert spaces. Given measurements (inner products) of an element with respect to one basis, it allows one to reconstruct in another, arbitrary basis, in a way that is both convergent and numerically stable. However, generalized sampling is thus far only valid for sampling and reconstruction in systems that comprise bases. Thus, in the ﬁrst part of this paper we extend this framework from bases to frames, and provide fundamental sampling theorems for this more general case. The second part of the paper is concerned with extending the idea of generalized sampling to the solution of inverse and ill-posed problems. In particular, we introduce two generalized sampling frameworks for such problems, based on regularized and non-regularized approaches. We furnish evidence of the usefulness of the proposed theories by providing a number of numerical experiments.

At ASU in APM/EEE 598 Reverse Engineering of Complex Dynamical Networks taught by Rodrigo Platte

- 11/07 - Robert Thompson - Reverse Engineering Dynamical Systems From Time-Series Data Using l1-Minimization (PDF) (Platte)
- 11/21 - Lei Ying - "Predicting communication networks from time series by compressive sensing" (PDF) (Lai)
- 11/21 - Riqi Su - "Predict network collective behavior via compressive sensing" (PDF) (Lai)

.

Upscaling Ambisonic sound scenes using compressed sensing techniques by Wabnitz, Andrew; Epain, Nicolas; McEwan, Alistair; Jin, Craig. The abstract reads:

This paper considers the application of compressed sensing to spherical acoustics in order to improve spatial sound field reconstruction. More specifically, we apply compressed sensing techniques to a set of Ambisonic sound signals to obtain a super-resolution plane-wave decomposition of the original sound field. That is to say, we investigate using the plane-wave decomposition to increase the spherical harmonic order of the Ambisonic sound scene. We refer to this as upscaling the Ambisonic sound scene. A focus of the paper is using sub-band analysis to make the plane-wave decomposition more robust. Results show that the sub-band analysis does indeed improve the robustness of the plane-wave decomposition when dominant overlapping sources are present or in noisy or diffuse sound conditions. Upscaling Ambisonic sound scenes allows more loudspeakers to be used for spatial sound field reconstruction, resulting in a larger sweet spot and improved sound quality.

David let me know of the Advance program for the 2012 IEEE INTERNATIONAL SOLID-STATE CIRCUITS CONFERENCE

page 4222.4 A 256×256 CMOS Image Sensor With ΔΣ-Based Single-Shot Compressed Sensing Y. Oike, A. El Gamal, Stanford University, Stanford, CA; Sony, Atsugi, JapanPage 532:50 Is Compressed Sensing Relevant to Image Sensors? Abbas El Gamal, Stanford University, Stanford, CA1:45 Image and Depth from a Conventional Camera with a Coded Aperture, Bill Freeman, Massachusetts Institute of Technology, Cambridge, MA

Greedy Sparsity-Constrained Optimization by Sohail Bahmani, Petros Boufounos, and Bhiksha Raj. The abstract reads:

Finding optimal sparse solutions to estimation problems, particularly in underdetermined regimes has recently gained much attention. Most existing literature study linear models in which the squared error is used as the measure of discrepancy to be minimized. However, in many applications discrepancy is measured in more general forms such as log-likelihood. Regularization by `1-norm has been shown to induce sparse solutions, but their sparsity level can be merely suboptimal. In this paper we present a greedy algorithm, dubbed Gradient Support Pursuit (GraSP), for sparsity-constrained optimization. Quantiﬁable guarantees are provided for GraSP when cost functions have the “Stable Hessian Property”.and

Secure Binary Embeddings for Privacy Preserving Nearest Neighbors by Petros Boufounos and Shantanu Rane. The abstract reads:

Abstract—We present a novel method to securely determine whether two signals are similar to each other, and apply it to approximate nearest neighbor clustering. The proposed method relies on a locality sensitive hashing scheme based on a secure binary embedding, computed using quantized random projections. Hashes extracted from the signals preserve information about the distance between the signals, provided this distance is small enough. If the distance between the signals is larger than a threshold, then no information about the distance is revealed. Theoretical and experimental justiﬁcation is provided for this property. Further, when the randomized embedding parameters are unknown, then the mutual information between the hashes of any two signals decays to zero exponentially fast as a function of the `2 distance between the signals. Taking advantage of this property, we suggest that these binary hashes can be used to perform privacy-preserving nearest neighbor search with signiﬁcantly lower complexity compared to protocols which use the actual signals.

FUSION FRAMES AND THE RESTRICTED ISOMETRY PROPERTY by BERNHARD G. BODMANN, JAMESON CAHILL, AND PETER G. CASAZZA, The abstract reads:

We will show that tight frames satisfying the restricted isometry property give rise to nearly tight fusion frames which are nearly orthogonal and hence are nearly equi-isoclinic. We will also show how to replace parts of the RIP frame with orthonormal sets while maintaining the RIP property

When Compressive Sampling Meets Multicast: Outage Analysis and Subblock Network Coding by Chandrashekhar Thejaswi PS, Tuan Tran, and Junshan Zhang. The abstract reads:

Abstract—This paper studies multicasting compressively sampled signals from a source to many receivers, over lossy wireless channels. Our focus is on the network outage from the perspective of signal distortion across all receivers, for both cases where the transmitter may or may not be capable of reconstructing the compressively sampled signals. Capitalizing on extreme value theory, we characterize the network outage in terms of key system parameters, including the erasure probability, the number of receivers and the sparse structure of the signal. We show that when the transmitter can reconstruct the compressively sensed signal, the strategy of using network coding to multicast the reconstructed signal coefﬁcients can reduce the network outage signiﬁcantly. We observe, however, that the traditional network coding could result in suboptimal performance with powerlaw decay signals. Thus motivated, we devise a new method, namely subblock network coding, which involves fragmenting the data into subblocks, and allocating time slots to different subblocks, based on its priority. We formulate the corresponding optimal allocation as an integer programming problem. Since integer programming is often intractable, we develop a heuristic algorithm that prioritizes the time slot allocation by exploiting the inherent priority structure of power-law decay signals.Numerical results show that the proposed schemes outperform the traditional methods with signiﬁcant margins.

Analog Sparse Approximation with Applications to Compressed Sensing by Adam S. Charles, Pierre Garrigues, Christopher J. Rozell. The abstract reads:

Recent research has shown that performance in signal processing tasks can often be significantly improved by using signal models based on sparse representations, where a signal is approximated using a small number of elements from a fixed dictionary. Unfortunately, inference in this model involves solving non-smooth optimization problems that are computationally expensive. While significant efforts have focused on developing digital algorithms specifically for this problem, these algorithms are inappropriate for many applications because of the time and power requirements necessary to solve large optimization problems. Based on recent work in computational neuroscience, we explore the potential advantages of continuous time dynamical systems for solving sparse approximation problems if they were implemented in analog VLSI. Specifically, in the simulated task of recovering synthetic and MRI data acquired via compressive sensing techniques, we show that these systems can potentially perform recovery at time scales of 10-20{\mu}s, supporting datarates of 50-100 kHz (orders of magnitude faster that digital algorithms). Furthermore, we show analytically that a wide range of sparse approximation problems can be solved in the same basic architecture, including approximate $\ell^p$ norms, modified $\ell^1$ norms, re-weighted $\ell^1$ and $\ell^2$, the block $\ell^1$ norm and classic Tikhonov regularization.

Compressed Sensing with General Frames via Optimal-dual-based $\ell_1$-analysis by Yulong Liu, Tiebin Mi, Shidong Li. The abstract reads:

Compressed sensing allowing sparse representations with respect to frames is seen to have much greater range of practical applications. In such settings, one approach to recover the signal is known as $\ell_1$-analysis. We expand in this article the performance analysis of this approach by providing a weaker recovery condition. We also base our analysis on general frames and alternative dual frames (as analysis operators). As one application to such a general-dual-based approach and performance analysis, we also outline an optimal-dual-based technique to demonstrate the effectiveness of using alternative dual frames as analysis operators.

Sparse Group Selection Through Co-Adaptive Penalties by Zhou Fang. The abstract reads:

Recent work has focused on the problem of conducting linear regression when the number of covariates is very large, potentially greater than the sample size. To facilitate this, one useful tool is to assume that the model can be well approximated by a fit involving only a small number of covariates -- a so called sparsity assumption, which leads to the Lasso and other methods. In many situations, however, the covariates can be considered to be structured, in that the selection of some variables favours the selection of others -- with variables organised into groups entering or leaving the model simultaneously as a special case. This structure creates a different form of sparsity. In this paper, we suggest the Co-adaptive Lasso to fit models accommodating this form of `group sparsity'. The Co-adaptive Lasso is fast and simple to calculate, and we show that it holds theoretical advantages over the Lasso, performs well under a broad set of conclusions, and is very competitive in empirical simulations in comparison with previously suggested algorithms like the Group Lasso and the Adaptive Lasso.

Explicit Group Testing Scheme that Identifies Almost All Defective Configurations by Arya Mazumdar. The abstract reads:

In a group testing scheme a set of tests is designed to identify a small number t of defective items among a large set (of size N) of items. In the non-adaptive scenario the set of tests has to be designed in one-shot. In this setting designing a testing scheme is equivalent to the construction of a disjunct matrix, an M x N matrix where the union of supports of any t columns does not contain the support of any other column. In principle, one wants to have such a matrix with minimum possible number M of rows (tests).In this paper we relax the definition of disjunct matrix. The new definition allows one to come up with group testing schemes where almost all (as opposed to all) possible sets of defective items are identifiable. Our main contribution is to show that, it is possible to explicitly construct disjunct matrices with the relaxed definition with much smaller number of rows than possible with the original definition. As a consequence of our result, for any absolute constant \epsilon >0 and t proportional to any positive power of N, it is possible to explicitly construct a group testing scheme that identifies (1-\epsilon) proportion of all possible defective sets of size t using only O(t^{3/2}\sqrt{log(N/\epsilon)}) tests. Without our relaxation, the best known scheme requires O(t^2 log N) tests.

Efficient High-Dimensional Inference in the Multiple Measurement Vector Problem by Justin Ziniel, Philip Schniter. The abstract reads:

In this work, a Bayesian approximate message passing algorithm is proposed for solving the multiple measurement vector (MMV) problem in compressive sensing, in which a collection of sparse signal vectors that share a common support are recovered from undersampled noisy measurements. The algorithm, AMP-MMV, is capable of exploiting temporal correlations in the amplitudes of non-zero coefficients, and provides soft estimates of the signal vectors as well as the underlying support. Central to the proposed approach is an extension of recently developed approximate message passing techniques to the MMV setting. Aided by these techniques, AMP-MMV offers a computational complexity that is linear in all problem dimensions. In order to allow for automatic parameter tuning, an expectation-maximization algorithm that complements AMP-MMV is described. Finally, a detailed numerical study demonstrates the power of the proposed approach and its particular suitability for application to high-dimensional problems.

The Graphical Lasso: New Insights and Alternatives by Rahul Mazumder, Trevor Hastie. The abstract reads:

The graphical lasso [Banerjee et al., 2008, Friedman et al., 2007b] is a popular approach for learning the structure in an undirected Gaussian graphical model, using $\ell_1$ regularization to control the number of zeros in the precision matrix ?${\B\Theta}={\B\Sigma}^{-1}$. The R package glasso is popular, fast, and allows one to efficiently build a path of models for different values of the tuning parameter. Convergence of GLASSO can be tricky; the converged precision matrix might not be the inverse of the estimated covariance, and occasionally it fails to converge with warm starts. In this paper we explain this behavior, and propose new algorithms that appear to outperform GLASSO. We show that in fact glasso is solving the dual of the graphical lasso penalized likelihood, by block coordinate descent. In this dual, the target of estimation is $\B\Sigma$, the covariance matrix, rather than the precision matrix $\B\Theta$. We propose similar primal algorithms P-GLASSO and DP-GLASSO, that also operate by block-coordinate descent, where $\B\Theta$ is the optimization target. We study all of these algorithms, and in particular different approaches to solving their coordinate subproblems. We conclude that DP-GLASSO is superior from several points of view.

Compressive Echelle Spectroscopy by Lina Xu, Mark A. Davenport, Matthew A. Turner, Ting Sun, Kevin F. Kelly. The abstract reads:

Building on the mathematical breakthroughs of compressive sensing (CS), we developed a 2D spectrometer system that incorporates a spatial light modulator and a single detector. For some wavelengths outside the visible spectrum, when it is too expensive to produce the large detector arrays, this scheme gives us a better solution by using only one pixel. Combining this system with the “smashed filter” technique, we hope to create an efficient IR gas sensor. We performed Matlab simulations to evaluate the effectiveness of the smashed filter for gas tracing.

Template Matching via l1 Minimization and Its Application to Hyperspectral Data by Zhaohui Guo and Stanley Osher. The abstract reads:

Detecting and identifying targets or objects that are present in hyperspectral ground images are of great interest. Applications include land and environmental monitoring, mining, military, civil search-and-rescue operations, and so on. We propose and analyze an extremely simple and efﬁcient idea for template matching based on l1 minimization. The designed algorithm can be applied in hyperspectral classiﬁcation and target detection. Synthetic image data and real hyperspectral image (HSI) data are used to assess the performance, with comparisons to other approaches, e.g. spectral angle map (SAM), adaptive coherence estimator (ACE), generalized-likelihood ratio test (GLRT) and matched ﬁlter. We demonstrate that this algorithm achieves excellent results with both high speed and accuracy by using Bregman iteration.

RESOLUTION OF CROSSING FIBERS WITH CONSTRAINED COMPRESSED SENSING USING DIFFUSION TENSOR MRI by Bennett A. Landman , John A. Bogovic, Hanlin Wan, Fatma El Zahraa ElShahaby, Pierre-Louis Bazin, Jerry L. Prince. The abstract reads:

Diffusion tensor imaging (DTI) is widely used to characterize tissue micro-architecture and brain connectivity. In regions of crossing fibers, however, the tensor model fails because it cannot represent multiple, independent intra-voxel orientations. Most of the methods that have been proposed to resolve this problem require diffusion magnetic resonance imaging (MRI) data that comprise large numbers of angles and high b-values, making them problematic for routine clinical imaging and many scientific studies. We present a technique based on compressed sensing that can resolve crossing fibers using diffusion MRI data that can be rapidly and routinely acquired in the clinic (30 directions, b-value equal to 700 s/mm2). The method assumes that the observed data can be well fit using a sparse linear combination of tensors taken from a fixed collection of possible tensors each having a different orientation. A fast algorithm for computing the best orientations based on a hierarchical compressed sensing algorithm and a novel metric for comparing estimated orientations are also proposed. The performance of this approach is demonstrated using both simulations and in vivo images. The method is observed to resolve crossing fibers using conventional data as well as a standard q-ball approach using much richer data that requires considerably more image acquisition time.

A Compressive Sensing and Unmixing Scheme for Hyperspectral Data Processing by Chengbo Li, Ting Sun, Kevin F. Kelly and Yin Zhang. The abstract reads:

Hyperspectral data processing typically demands enormous computational resources in terms of storage, computation and I/O throughputs, especially when real-time processing is desired. In this paper, we investigate a lowcomplexity scheme for hyperspectral data compression and reconstruction. In this scheme, compressed hyperspectral data are acquired directly by a device similar to the single-pixel camera [5] based on the principle of compressive sensing. To decode the compressed data, we propose a numerical procedure to directly compute the unmixed abundance fractions of given endmembers, completely bypassing high-complexity tasks involving the hyperspectral data cube itself. The reconstruction model is to minimize the total variation of the abundance fractions subject to a preprocessed ﬁdelity equation with a signiﬁcantly reduced size, and other side constraints. An augmented Lagrangian type algorithm is developed to solve this model. We conduct extensive numerical experiments to demonstrate the feasibility and efﬁciency of the proposed approach, using both synthetic data and hardware-measured data. Experimental and computational evidences obtained from this study indicate that the proposed scheme has a high potential in real-world applications.

A Compressed Sensing Approach to 3D Weak Lensing by Adrienne Leonard, François-Xavier Dupé, Jean-Luc Starck. The abstract reads:

(Abridged) Weak gravitational lensing is an ideal probe of the dark universe. In recent years, several linear methods have been developed to reconstruct the density distribution in the Universe in three dimensions, making use of photometric redshift information to determine the radial distribution of lensed sources. In this paper, we aim to address three key issues seen in these methods; namely, the bias in the redshifts of detected objects, the line of sight smearing seen in reconstructions, and the damping of the amplitude of the reconstruction relative to the underlying density. We consider the problem under the framework of compressed sensing (CS). Under the assumption that the data are sparse in an appropriate dictionary, we construct a robust estimator and employ state-of-the-art convex optimisation methods to reconstruct the density contrast. For simplicity in implementation, and as a proof of concept of our method, we reduce the problem to one-dimension, considering the reconstruction along each line of sight independently. Despite the loss of information this implies, we demonstrate that our method is able to accurately reproduce cluster haloes up to a redshift of z=1, deeper than state-of-the-art linear methods. We directly compare our method with these linear methods, and demonstrate minimal radial smearing and redshift bias in our reconstructions, as well as a reduced damping of the reconstruction amplitude as compared to the linear methods. In addition, the CS framework allows us to consider an underdetermined inverse problem, thereby allowing us to reconstruct the density contrast at finer resolution than the input data.

Compressive Phase Retrieval From Squared Output Measurements Via Semidefinite Programming by Henrik Ohlsson, Allen Y. Yang, Roy Dong, S. Shankar Sastry. The abstract reads:

Given a linear system in a real or complex domain, linear regression aims to recover the model parameters from a set of observations. Recent studies in compressive sensing have successfully shown that under certain conditions, a linear program, namely, l1-minimization, guarantees recovery of sparse parameter signals even when the system is underdetermined. In this paper, we consider a more challenging problem: when the phase of the output measurements from a linear system is omitted. Using a lifting technique, we show that even though the phase information is missing, the sparse signal can be recovered exactly by solving a simple semidefinite program when the sampling rate is sufficiently high. This is an interesting finding since the exact solutions to both sparse signal recovery and phase retrieval are combinatorial. Besides, this also extends the type of applications that compressive sensing can be applied to those where only output magnitudes can be observed. We demonstrate the accuracy of the algorithms through extensive simulation and a practical experiment.

The goal of the following paper is absolutely not clear to me. The author first decomposes an image as a series of wavelets and then use a sparse measurement matrix to compressed the wavelet coordinates (and produce compressed samples). He then reconstruct the original image using this smaller set of compressed samples to check whether it looks similar to the original image. This is all fine, but I did not see the purpose for this: is it for decreasing the comparison between original image that one tries to see if compressed samples can be comparable ? I don't know.

Compressed sensing of astronomical images: orthogonal wavelets domains by Vasil Kolev. The abstract reads:

A simple approach for orthogonal wavelets in compressed sensing (CS) applications is presented. We compare efficient algorithm for different orthogonal wavelet measurement matrices in CS for image processing from scanned photographic plates (SPP). Some important characteristics were obtained for astronomical image processing of SPP. The best orthogonal wavelet choice for measurement matrix construction in CS for image compression of images of SPP is given. The image quality measure for linear and nonlinear image compression method is defined.

Fast Algorithms for Sparse Recovery with Perturbed Dictionary by Xuebing Han, Hao Zhang. The abstract reads:

In this paper, for sparse recovery of large underdetermined linear systems, we propose a new kind of fast algorithms, based on totally least square (TLS) method and FOCUSS (FOCal Underdetermined System Solver). The problem about sparse recovery was considered, when perturbations appear in both the measurements and the dictionary (sensing matrix) (here we can call the system model as TLS model). The objective function to be optimized is deduced through a maximum a posteriori (MAP) estimation. Then a new FOCUSS algorithm, named TLS-FOCUSS, is extended with main idea of TLS, to reduce the impact of the perturbation of dictionary and measurements to the performance of sparse recovery. Compared with other recovery algorithms on TLS model, TLS-FOCUSS algorithm is not only near-optimum but also fast, thus fit for large scale computation. In order to reduce the complexity of algorithm further, another suboptimal algorithm named SD-FOCUSS is proposed. Another breakthrough of the paper is that SD-FOCUSS can be applied in MMV (multiple measurement vectors) TLS model which field is not researched at present. The convergence of the TLS-FOCUSS algorithm and SD-FOCUSS algorithm is established with mathematical proof. The new algorithms based on TLS model are proved to be efficient and high-performance. The simulation results illustrate the advantage of TLS-FOCUSS and SD-FOCUSS on accuracyand stability compared with the other algorithms.

Continuous Measurement Quantum State Tomography of Atomic Ensembles by Carlos A. Riofrío. The abstract reads:

Quantum state tomography is a fundamental tool in quantum information processing. It allows us to estimate the state of a quantum system by measuring different observables on many identically prepared copies of the system. This is, in general, a very time-consuming task that requires a large number of measurements. There are, however, systems in which the data acquisition can be done more efficiently. In fact, an ensemble of quantum systems can be prepared and manipulated by external fields while being continuously and collectively probed, producing enough information to estimate its state. This provides a basis for continuous measurement quantum tomography. In this protocol, an ensemble of identically prepared systems is collectively probed and controlled in a time-dependent manner to create an informationally complete continuous measurement record. The measurement history is then inverted to determine the state at the initial time. We use two different estimation methods: maximum likelihood and compressed sensing. The general formalism is applied to the case of reconstruction of the quantum state encoded in the magnetic sub-levels of a large-spin alkali atom, ${}^{133}$Cs. We apply this protocol to the case of reconstruction of states in the full 16-dimensional electronic-ground subspace ($F=3 \oplus F=4$), controlled by microwaves and radio-frequency magnetic fields. We present an experimental demonstration of continuous measurement quantum tomography in an ensemble of cold cesium atoms with full control of its 16-dimensional Hilbert space. We show the exquisite level of control achieved in the lab and the excellent agreement between the theory discussed in this dissertation and the experimental results. This allows us to achieve fidelities >95% for low complexity quantum states, and >92% for arbitrary random states, which is a formidable accomplishment for a space of this size.

Distributed Representation of Geometrically Correlated Images with Compressed Linear Measurements by Vijayaraghavan Thirumalai, Pascal Frossard. The abstract reads:

This paper addresses the problem of distributed coding of images whose correlation is driven by the motion of objects or positioning of the vision sensors. It concentrates on the problem where images are encoded with compressed linear measurements. We propose a geometry-based correlation model in order to describe the common information in pairs of images. We assume that the constitutive components of natural images can be captured by visual features that undergo local transformations (e.g., translation) in different images. We first identify prominent visual features by computing a sparse approximation of a reference image with a dictionary of geometric basis functions. We then pose a regularized optimization problem to estimate the corresponding features in correlated images given by quantized linear measurements. The estimated features have to comply with the compressed information and to represent consistent transformation between images. The correlation model is given by the relative geometric transformations between corresponding features. We then propose an efficient joint decoding algorithm that estimates the compressed images such that they stay consistent with both the quantized measurements and the correlation model. Experimental results show that the proposed algorithm effectively estimates the correlation between images in multi-view datasets. In addition, the proposed algorithm provides effective decoding performance that compares advantageously to independent coding solutions as well as state-of-the-art distributed coding schemes based on disparity learning.

Compressive Sensing Framework for Speech Signal Synthesis Using a Hybrid Dictionary by Yue Wang, Zhixing Xu, Gang Li, Liping Chang and Chuanrong Hong. The abstract reads:

Compressive sensing (CS) is a promising focus in signal processing ﬁeld, which offers a novel view of simultaneous compression and sampling. In this framework a sparse approximated signal is obtained with samples much less than that required by the Nyquist sampling theorem if the signal is sparse on one basis. Encouraged by its exciting potential application in signal compression, we use CS framework for speech synthesis problems. The linear prediction coding (LPC) is an efﬁcient tool for speech compression, as the speech is considered to be an AR process. It is also known that a speech signal is quasi-periodic in its voiced parts, hence a discrete fourier transform (DFT) basis will provide a better approximation. Thus we propose a hybrid dictionary combined with the LPC model and the DFT model as the basis of speech signal. The orthogonal matching pursuit (OMP) is employed in our simulations to compute the sparse representation in the hybrid dictionary domain. The results indicate good performance with our proposed scheme, offering a satisfactory perceptual quality.

Accuracy guaranties for `1 recovery of block-sparse signals by Anatoli Juditsky, Fatma Kilinc Karzan, Arkadi Nemirovski, Boris Polyak. The abstract reads:

We discuss new methods for the recovery of signals with block-sparse structure, based on `1-minimization.Our emphasis is on verifiable conditions on the problem parameters (sensing matrix and the block structure) for accurate recovery and e ciently computable bounds for the recovery error. These bounds are then optimized with respect to the method parameters to construct the estimators with improved statistical properties. To justify the proposed approach we provide an oracle inequality which links the properties of the recovery algorithms and the best estimation performance. We also propose a new matching pursuit algorithm for block-sparse recovery

Fighting the curse of dimensionality: compressive sensing in exploration seismology by Felix J. Herrmann, Michael P. Friedlander, Ozgur Yılmaz. The abstract reads:

Many seismic exploration techniques rely on the collection of massive data volumes that are mined for information during processing. This approach has been extremely successful, but current efforts toward higher-resolution images in increasingly complicated regions of the Earth continue to reveal fundamental shortcomings in our typical workﬂows. The “curse of dimensionality” is the main roadblock, and is exempliﬁed by Nyquist’s sampling criterion, which disproportionately strains current acquisition and processing systems as the size and desired resolution of our survey areas continues to increase. We offer an alternative sampling strategy that leverages recent insights from compressive sensing towards seismic acquisition and processing for data that are traditionally considered to be undersampled. The main outcome of this approach is a new technology where acquisition and processing related costs are no longer determined by overly stringent sampling criteria. Compressive sensing is a novel nonlinear sampling paradigm, effective for acquiring signals that have a sparse representation in some transform domain. We review basic facts about this new sampling paradigm that revolutionized various areas of signal processing, and illustrate how it can be successfully exploited in various problems in seismic exploration to effectively ﬁght the curse of dimensionality.

Generalized sampling and the stable and accurate reconstruction of piecewise analytic functions from their Fourier coefﬁcients by Ben Adcock, Anders C. Hansen. The abstract reads:

Here is Postdoc at University of New Mexico:Suppose that the ﬁrst m Fourier coefﬁcients of a piecewise analytic function are given. Direct expansion in a Fourier series suffers from the Gibbs phenomenon and lacks uniform convergence. Nonetheless, in this paper we show that, under very broad conditions, it is always possible to recover an n-term expansion in a different system of functions using only these coefﬁcients. Such an expansion can be made arbitrarily close to the best possible n-term expansion in the given system. Thus, if a piecewise polynomial basis is employed, for example, exponential convergence can be restored. The resulting method is linear, numerically stable and can be implemented efﬁciently in only O (nm) operations. A key issue is how the parameter m must scale in comparison to n to ensure recovery. We derive analytical estimates for this scaling for large classes of polynomial and piecewise polynomial bases. In particular, we show that in many important cases, including the case of piecewise Chebyshev polynomials, this scaling is quadratic: m = O`n2´. Therefore, using a system of polynomials that the user is essentially free to choose, one can restore exponential accuracy in n and root exponential accuracy in m. Thisgeneralizes a result proved recently for piecewise Legendre polynomials. The method developed in this paper is part of new numerical framework for sampling and reconstruction in abstract Hilbert spaces, known as generalized sampling. This paper extends previous work by the authors by introducing a substantially more ﬂexible methodology which allows for sampling and reconstruction with respect to different inner products. In the ﬁnal part of this paper we illustrate the application of generalized sampling to a related family of problems.

Postdoctoral Fellowship in Diffusion Tensor MR Spectroscopic Imaging in Human BrainDepartment of NeurologyUniversity of New Mexico School of MedicineApplicants are invited to apply for an NIH funded postdoctoral fellow position in the Department of Neurology at the University of New Mexico School of Medicine (http://hsc.unm.edu/som/neuro/lab/). The goal of the fellowship project is to develop novel ultra-fast MR spectroscopic imaging methods to measure the diffusion tensor of metabolites in human brain. The project involves development of Proton-Echo-Planar-Spectroscopic-Imaging (PEPSI) with parallel imaging and compressed sensing. In vivo studies will be conducted in healthy adults and in children with autism in collaboration with the University of Washington and Seattle Children’s Hospital. Facilities include whole body Siemens TIM Trio scanners equipped with state-of-the-art 32 channel array coils.Successful candidates will hold a PhD in biomedical MR spectroscopy, MR physics, MR engineering, or related fields. This project requires experience in MR spectroscopic imaging, parallel MRI reconstruction and compressed sensing. Strong candidates will have experience in the areas of MR pulse sequence programming (Siemens Syngo platform preferred) and diffusion tensor imaging. A statistical modeling background is desirable, but not required. A demonstrable record of peer reviewed journal publications and strong expertise in C/C++, and/or Matlab are essential. This 2 year position offers an excellent opportunity to be involved in inter-disciplinary research in a thriving functional neuroimaging environment. Strong collaborations exist between clinical and basic science departments at UNM, with the MIND Research Network (http://www.themindinstitute.org/), and with national and international research centers. A career path to research faculty is a possibility for outstanding candidates.Please send your curriculum vitae, a letter describing your interest, background, and qualifications, and 3 letters ofrecommendation to: sposse@unm.eduStefan Posse, PhDProfessor of NeurologyDepartment of NeurologyUniversity of New Mexico School of MedicineMSC 10 56201 University of New MexicoAlbuquerque, NM 87131

And finally a talk that tool place last week:

Artificial Intelligence Lab Seminar2011 Nov 25 at 15:30PAS 2464Oleg V. Michailovich, Assistant Professor, Department of ECE University of WaterlooThe unique ability of diffusion-weighted MRI (DW-MRI) to generate contrast based on the morphological properties of white matter opens the door to developing qualitatively new methods of early detection and diagnosis of many brain-related disorders. Unfortunately, practical implementation of DW-MRI still poses a number of challenges which hamper its wide-spread integration into standard clinical practice. Chief among these is the problem of prohibitively long scanning times, which necessitates the development of time-efficient methods for acquisition of diffusion data. In many such methods, however, the acceleration entails a trade-off between the time efficiency and the accuracy of signal reconstruction. In such a case, it is imperative for one to be able to understand the effect the above trade-off might have on the accuracy of diagnostic inference. Accordingly, the objective of this talk is twofold. First, using high-angular resolution diffusion imaging (HARDI) as a specific instance of DW-MRI, we will introduce the notion of a directional diffusion structure which, in combination with multidimensional scaling, allows representing HARDI data in a lower dimensional Euclidean space. Subsequently, based on this representation, we will develop an algorithm for detection and classification of first episode schizophrenia. Finally, the above algorithm will be applied to HARDI data acquired by means of compressed sensing and we will demonstrate that the resulting classification error increases insignificantly when the sampling density is reduced to as low as a fourth of its conventional value.

Liked this entry ? subscribe to Nuit Blanche's feed, there's more where that came from. You can also subscribe to Nuit Blanche by Email, explore the Big Picture in Compressive Sensing or the Matrix Factorization Jungle and join the conversations on compressive sensing, advanced matrix factorization and calibration issues on Linkedin.

## No comments:

Post a Comment