Fast Dictionary Learning for Sparse Representations of Speech Signals

1
Bharathi
Bharathi
2
Maria G. Jafari
Maria G. Jafari
3
Mark D. Plumbley
Mark D. Plumbley
1 Kingas College London, London, U.K

Send Message

To: Author

GJCST Volume 14 Issue E8

Article Fingerprint

ReserarchID

CSTNWS92JS3

Fast Dictionary Learning for Sparse Representations of Speech Signals Banner
  • English
  • Afrikaans
  • Albanian
  • Amharic
  • Arabic
  • Armenian
  • Azerbaijani
  • Basque
  • Belarusian
  • Bengali
  • Bosnian
  • Bulgarian
  • Catalan
  • Cebuano
  • Chichewa
  • Chinese (Simplified)
  • Chinese (Traditional)
  • Corsican
  • Croatian
  • Czech
  • Danish
  • Dutch
  • Esperanto
  • Estonian
  • Filipino
  • Finnish
  • French
  • Frisian
  • Galician
  • Georgian
  • German
  • Greek
  • Gujarati
  • Haitian Creole
  • Hausa
  • Hawaiian
  • Hebrew
  • Hindi
  • Hmong
  • Hungarian
  • Icelandic
  • Igbo
  • Indonesian
  • Irish
  • Italian
  • Japanese
  • Javanese
  • Kannada
  • Kazakh
  • Khmer
  • Korean
  • Kurdish (Kurmanji)
  • Kyrgyz
  • Lao
  • Latin
  • Latvian
  • Lithuanian
  • Luxembourgish
  • Macedonian
  • Malagasy
  • Malay
  • Malayalam
  • Maltese
  • Maori
  • Marathi
  • Mongolian
  • Myanmar (Burmese)
  • Nepali
  • Norwegian
  • Pashto
  • Persian
  • Polish
  • Portuguese
  • Punjabi
  • Romanian
  • Russian
  • Samoan
  • Scots Gaelic
  • Serbian
  • Sesotho
  • Shona
  • Sindhi
  • Sinhala
  • Slovak
  • Slovenian
  • Somali
  • Spanish
  • Sundanese
  • Swahili
  • Swedish
  • Tajik
  • Tamil
  • Telugu
  • Thai
  • Turkish
  • Ukrainian
  • Urdu
  • Uzbek
  • Vietnamese
  • Welsh
  • Xhosa
  • Yiddish
  • Yoruba
  • Zulu

For dictionary-based decompositions of certain types, it has been observed that there might be a link between sparsity in the dictionary and sparsity in the decomposition. Sparsity in the dictionary has also been associated with the derivation of fast and efficient dictionary learning algorithms. Therefore, in this paper we present a greedy adaptive dictionary learning algorithm that sets out to find sparse atoms for speech signals. The algorithm learns the dictionary atoms on data frames taken from a speech signal. It iteratively extracts the data frame with minimum sparsity index, and adds this to the dictionary matrix. The contribution of this atom to the data frames is then removed, and the process is repeated. The algorithm is found to yield a sparse signal decomposition, supporting the hypothesis of a link between sparsity in the decomposition and dictionary. The algorithm is applied to the problem of speech representation and speech denoising, and its performance is compared to other existing methods.

29 Cites in Articles

References

  1. P Földiák (1990). Forming sparse representations by local anti-Hebbian learning.
  2. B Olshausen,D Field (1996). Emergence of simplecell receptive field properties by learning a sparse code for natural images.
  3. T Shan,L Jiao (2005). New evidences for sparse coding strategy employed in visual neurons: From the image processing and nonlinear approximation viewpoint.
  4. Michael Zibulevsky,Barak Pearlmutter (2001). Blind Source Separation by Sparse Decomposition in a Signal Dictionary.
  5. Ö Yilmaz,S Rickard (2004). Blind separation of speech mixtures via time-frequency masking.
  6. R (2002). Sparse decomposition of stereo signals with Matching Pursuit and application to blind separation of more than two sources from a stereo mixture.
  7. M Elad,M Aharon (2006). Image denoising via sparse redundant representations over learned dictionaries.
  8. R Gribonval,K Schnass (2008). Some recovery conditions for basis learning by L1-minimization.
  9. L Rebollo-Neira (2004). Dictionary redundancy elimination.
  10. Michael Lewicki,Terrence Sejnowski (2000). Learning Overcomplete Representations.
  11. Irina Gorodnitsky,John George,Bhaskar Rao (1995). Neuromagnetic source imaging with FOCUSS: a recursive weighted minimum norm algorithm.
  12. M Aharon,M Elad,A Bruckstein (2006). K-SVD: An algorithm for designing overcomplete dictionaries for sparse representations.
  13. Mark Plumbley (2007). Dictionary Learning for L1-Exact Sparse Coding.
  14. J Mairal,F Bach,J Ponce,G Sapiro (2009). Online dictionary learning for sparse coding.
  15. R Rubinstein,M Zibulevsky,M Elad (2010). Double sparsity: Learning sparse dictionaries for sparse signal approximation.
  16. Maria Jafari,Emmanuel Vincent,Samer Abdallah,Mark Plumbley,Mike Davies (2008). An adaptive stereo basis method for convolutive blind audio source separation.
  17. S Abdallah,M Plumbley (2001). If edges are the independent components of natural images, what are the independent components of natural sounds?.
  18. M Goodwin,M Vetterli (1999). Matching pursuit and atomic signal models based on recursive filter banks.
  19. V Tan,C Févotte (2005). A study of the effect of source sparsity for various transforms on blind audio source separation performance.
  20. Maria Jafari,Mark Plumbley (2008). An adaptive orthogonal sparsifying transform for speech signals.
  21. S Mallat,Z Zhang (1993). Matching pursuit with timefrequency dictionaries.
  22. S Mallat,G Davis,Z Zhang (1994). Adaptive timefrequency decompositions.
  23. S Haykin (1999). Neural Networks: A Comprehensive Foundation.
  24. Frederic Font (2021). SOURCE: a Freesound Community Music Sampler.
  25. R Rubinstein Unknown Title.
  26. S Valiollahzadeh,H Firouzi,M Babaie-Zadeh,C Jutten (2009). Image denoising using sparse representations.
  27. A Hyvaerinen,J Karhunen,E Oja (2000). Independent Component Analysis.
  28. N Hurley,S Rickard (2009). Comparing measures of sparsity.
  29. W Kleijn (1993). Encoding speech using prototype waveforms.

Funding

No external funding was declared for this work.

Conflict of Interest

The authors declare no conflict of interest.

Ethical Approval

No ethics committee approval was required for this article type.

Data Availability

Not applicable for this article.

Bharathi. 2014. \u201cFast Dictionary Learning for Sparse Representations of Speech Signals\u201d. Global Journal of Computer Science and Technology - E: Network, Web & Security GJCST-E Volume 14 (GJCST Volume 14 Issue E8): .

Download Citation

Issue Cover
GJCST Volume 14 Issue E8
Pg. 37- 44
Journal Specifications

Crossref Journal DOI 10.17406/gjcst

Print ISSN 0975-4350

e-ISSN 0975-4172

Classification
Not Found
Version of record

v1.2

Issue date

December 30, 2014

Language

English

Experiance in AR

The methods for personal identification and authentication are no exception.

Read in 3D

The methods for personal identification and authentication are no exception.

Article Matrices
Total Views: 8803
Total Downloads: 2223
2026 Trends
Research Identity (RIN)
Related Research

Published Article

For dictionary-based decompositions of certain types, it has been observed that there might be a link between sparsity in the dictionary and sparsity in the decomposition. Sparsity in the dictionary has also been associated with the derivation of fast and efficient dictionary learning algorithms. Therefore, in this paper we present a greedy adaptive dictionary learning algorithm that sets out to find sparse atoms for speech signals. The algorithm learns the dictionary atoms on data frames taken from a speech signal. It iteratively extracts the data frame with minimum sparsity index, and adds this to the dictionary matrix. The contribution of this atom to the data frames is then removed, and the process is repeated. The algorithm is found to yield a sparse signal decomposition, supporting the hypothesis of a link between sparsity in the decomposition and dictionary. The algorithm is applied to the problem of speech representation and speech denoising, and its performance is compared to other existing methods.

Our website is actively being updated, and changes may occur frequently. Please clear your browser cache if needed. For feedback or error reporting, please email [email protected]
×

This Page is Under Development

We are currently updating this article page for a better experience.

Request Access

Please fill out the form below to request access to this research paper. Your request will be reviewed by the editorial or author team.
X

Quote and Order Details

Contact Person

Invoice Address

Notes or Comments

This is the heading

Lorem ipsum dolor sit amet, consectetur adipiscing elit. Ut elit tellus, luctus nec ullamcorper mattis, pulvinar dapibus leo.

High-quality academic research articles on global topics and journals.

Fast Dictionary Learning for Sparse Representations of Speech Signals

Maria G. Jafari
Maria G. Jafari
Mark D. Plumbley
Mark D. Plumbley

Research Journals