Dictionary Learning With Residual Codes
Loading...
Files
Date
2017
Authors
Türkan, Mehmet
Journal Title
Journal ISSN
Volume Title
Publisher
Institute of Electrical and Electronics Engineers Inc.
Open Access Color
Green Open Access
No
OpenAIRE Downloads
OpenAIRE Views
Publicly Funded
No
Abstract
In conventional sparse representations based dictionary learning algorithms, initial dictionaries are generally assumed to be proper representatives of the system at hand. However, this may not be the case, especially in some systems restricted to random initialization. Therefore, a supposedly optimal state-update based on such an improper model might lead to undesired effects that will be conveyed to successive learning iterations. In this paper, we propose a dictionary learning method which includes a general error-correction process that codes the residual left over from a less intensive initial learning attempt and then adjusts the sparse codes accordingly. Experimental observations show that such additional step vastly improves rates of convergence in high-dimensional cases, also results in better converged states in the case of random initialization. Improvements also scale up with more lenient sparsity constraints. © 2017 IEEE.
Description
25th Signal Processing and Communications Applications Conference, SIU 2017 -- 15 May 2017 through 18 May 2017 -- 128703
Keywords
dictionary learning, residual codes, sparse approximation, Sparse coding, Codes (symbols), Learning algorithms, Signal processing, Dictionary learning, Dictionary learning algorithms, Rates of convergence, residual codes, Sparse approximations, Sparse coding, Sparse representation, Sparsity constraints, Education
Fields of Science
0202 electrical engineering, electronic engineering, information engineering, 02 engineering and technology
Citation
WoS Q
N/A
Scopus Q
N/A

OpenCitations Citation Count
N/A
Source
2017 25th Signal Processing and Communications Applications Conference, SIU 2017
Volume
Issue
Start Page
1
End Page
4
PlumX Metrics
Citations
Scopus : 1
Captures
Mendeley Readers : 4
Google Scholar™


