2013
Luengo, David; Via, Javier; Monzon, Sandra; Trigano, Tom; Artés-Rodríguez, Antonio
Cross-Products LASSO Proceedings Article
En: 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 6118–6122, IEEE, Vancouver, 2013, ISSN: 1520-6149.
Resumen | Enlaces | BibTeX | Etiquetas: Approximation methods, approximation theory, concave programming, convex programming, Cost function, cross-product LASSO cost function, Dictionaries, dictionary, Encoding, LASSO, learning (artificial intelligence), negative co-occurrence, negative cooccurrence phenomenon, nonconvex optimization problem, Signal processing, signal processing application, signal reconstruction, sparse coding, sparse learning approach, Sparse matrices, sparsity-aware learning, successive convex approximation, Vectors
@inproceedings{Luengo2013,
title = {Cross-Products LASSO},
author = {David Luengo and Javier Via and Sandra Monzon and Tom Trigano and Antonio Art\'{e}s-Rodr\'{i}guez},
url = {http://ieeexplore.ieee.org/articleDetails.jsp?arnumber=6638840},
issn = {1520-6149},
year = {2013},
date = {2013-01-01},
booktitle = {2013 IEEE International Conference on Acoustics, Speech and Signal Processing},
pages = {6118--6122},
publisher = {IEEE},
address = {Vancouver},
abstract = {Negative co-occurrence is a common phenomenon in many signal processing applications. In some cases the signals involved are sparse, and this information can be exploited to recover them. In this paper, we present a sparse learning approach that explicitly takes into account negative co-occurrence. This is achieved by adding a novel penalty term to the LASSO cost function based on the cross-products between the reconstruction coefficients. Although the resulting optimization problem is non-convex, we develop a new and efficient method for solving it based on successive convex approximations. Results on synthetic data, for both complete and overcomplete dictionaries, are provided to validate the proposed approach.},
keywords = {Approximation methods, approximation theory, concave programming, convex programming, Cost function, cross-product LASSO cost function, Dictionaries, dictionary, Encoding, LASSO, learning (artificial intelligence), negative co-occurrence, negative cooccurrence phenomenon, nonconvex optimization problem, Signal processing, signal processing application, signal reconstruction, sparse coding, sparse learning approach, Sparse matrices, sparsity-aware learning, successive convex approximation, Vectors},
pubstate = {published},
tppubtype = {inproceedings}
}
2012
Campo, Adria Tauste; Vazquez-Vilar, Gonzalo; i Fabregas, Albert Guillen; Koch, Tobias; Martinez, Alfonso
Achieving Csiszár's Source-Channel Coding Exponent with Product Distributions Proceedings Article
En: 2012 IEEE International Symposium on Information Theory Proceedings, pp. 1548–1552, IEEE, Cambridge, MA, 2012, ISSN: 2157-8095.
Resumen | Enlaces | BibTeX | Etiquetas: average probability of error, Channel Coding, code construction, codewords, Csiszár's source-channel coding, Decoding, Encoding, error probability, error statistics, Joints, Manganese, product distributions, random codes, random-coding upper bound, source coding, source messages, Upper bound
@inproceedings{Campo2012a,
title = {Achieving Csisz\'{a}r's Source-Channel Coding Exponent with Product Distributions},
author = {Adria Tauste Campo and Gonzalo Vazquez-Vilar and Albert Guillen i Fabregas and Tobias Koch and Alfonso Martinez},
url = {http://ieeexplore.ieee.org/lpdocs/epic03/wrapper.htm?arnumber=6283524},
issn = {2157-8095},
year = {2012},
date = {2012-01-01},
booktitle = {2012 IEEE International Symposium on Information Theory Proceedings},
pages = {1548--1552},
publisher = {IEEE},
address = {Cambridge, MA},
abstract = {We derive a random-coding upper bound on the average probability of error of joint source-channel coding that recovers Csiszár's error exponent when used with product distributions over the channel inputs. Our proof technique for the error probability analysis employs a code construction for which source messages are assigned to subsets and codewords are generated with a distribution that depends on the subset.},
keywords = {average probability of error, Channel Coding, code construction, codewords, Csiszár's source-channel coding, Decoding, Encoding, error probability, error statistics, Joints, Manganese, product distributions, random codes, random-coding upper bound, source coding, source messages, Upper bound},
pubstate = {published},
tppubtype = {inproceedings}
}
2009
Vinuelas-Peris, Pablo; Artés-Rodríguez, Antonio
Sensing Matrix Optimization in Distributed Compressed Sensing Proceedings Article
En: 2009 IEEE/SP 15th Workshop on Statistical Signal Processing, pp. 638–641, IEEE, Cardiff, 2009, ISBN: 978-1-4244-2709-3.
Resumen | Enlaces | BibTeX | Etiquetas: Compressed sensing, Computer Simulation, computer simulations, correlated signal, Correlated signals, correlation theory, Dictionaries, distributed coding strategy, distributed compressed sensing, Distributed control, efficient projection method, Encoding, joint recovery method, Matching pursuit algorithms, Optimization methods, orthogonal matching pursuit, Projection Matrix Optimization, sensing matrix optimization, Sensor Network, Sensor phenomena and characterization, Sensor systems, Signal processing, Sparse matrices, Technological innovation
@inproceedings{Vinuelas-Peris2009,
title = {Sensing Matrix Optimization in Distributed Compressed Sensing},
author = {Pablo Vinuelas-Peris and Antonio Art\'{e}s-Rodr\'{i}guez},
url = {http://ieeexplore.ieee.org/lpdocs/epic03/wrapper.htm?arnumber=5278496},
isbn = {978-1-4244-2709-3},
year = {2009},
date = {2009-01-01},
booktitle = {2009 IEEE/SP 15th Workshop on Statistical Signal Processing},
pages = {638--641},
publisher = {IEEE},
address = {Cardiff},
abstract = {Distributed compressed sensing (DCS) seeks to simultaneously measure signals that are each individually sparse in some domain(s) and also mutually correlated. In this paper we consider the scenario in which the (overcomplete) bases for common component and innovations are different. We propose and analyze a distributed coding strategy for the common component, and also the use of efficient projection (EP) method for optimizing the sensing matrices in this setting. We show the effectiveness of our approach by computer simulations using the orthogonal matching pursuit (OMP) as joint recovery method, and we discuss the configuration of the distribution strategy.},
keywords = {Compressed sensing, Computer Simulation, computer simulations, correlated signal, Correlated signals, correlation theory, Dictionaries, distributed coding strategy, distributed compressed sensing, Distributed control, efficient projection method, Encoding, joint recovery method, Matching pursuit algorithms, Optimization methods, orthogonal matching pursuit, Projection Matrix Optimization, sensing matrix optimization, Sensor Network, Sensor phenomena and characterization, Sensor systems, Signal processing, Sparse matrices, Technological innovation},
pubstate = {published},
tppubtype = {inproceedings}
}