Publication:

DESIGNING TRANSFORMER NETWORKS FOR SPARSE RECOVERY OF SEQUENTIAL DATA USING DEEP UNFOLDING

 
cris.virtual.department#PLACEHOLDER_PARENT_METADATA_VALUE#
cris.virtual.department#PLACEHOLDER_PARENT_METADATA_VALUE#
cris.virtual.orcid0000-0001-9300-5860
cris.virtual.orcid0000-0001-9240-2370
cris.virtualsource.department90f2bec3-f84d-4738-9103-ba2cd2f04cbc
cris.virtualsource.departmenta67c0e76-79f3-4b00-81b6-908af3916467
cris.virtualsource.orcid90f2bec3-f84d-4738-9103-ba2cd2f04cbc
cris.virtualsource.orcida67c0e76-79f3-4b00-81b6-908af3916467
dc.contributor.authorDe Weerdt, Brent
dc.contributor.authorEldar, Yonina C.
dc.contributor.authorDeligiannis, Nikolaos
dc.date.accessioned2026-03-31T07:22:23Z
dc.date.available2026-03-31T07:22:23Z
dc.date.createdwos2026-02-21
dc.date.issued2023
dc.description.abstractDeep unfolding models are designed by unrolling an optimization algorithm into a deep learning network. These models have shown faster convergence and higher performance compared to the original optimization algorithms. Additionally, by incorporating domain knowledge from the optimization algorithm, they need much less training data to learn efficient representations. Current deep unfolding networks for sequential sparse recovery consist of recurrent neural networks (RNNs), which leverage the similarity between consecutive signals. We redesign the optimization problem to use correlations across the whole sequence, which unfolds into a Transformer architecture. Our model is used for the task of video frame reconstruction from low-dimensional measurements and is shown to outperform state-of-the-art deep unfolding RNN and Transformer models, as well as a traditional Vision Transformer on several video datasets.
dc.description.wosFundingTextThis work was supported by the FWO Flanders Ph.D. Fellowship Strategic Basic Research under Grant 1S44523N and imec.icon Surv-AI-llance.
dc.identifier.doi10.1109/icassp49357.2023.10094712
dc.identifier.issn1520-6149
dc.identifier.urihttps://imec-publications.be/handle/20.500.12860/58974
dc.language.isoeng
dc.provenance.editstepusergreet.vanhoof@imec.be
dc.publisherIEEE
dc.source.conferenceIEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)
dc.source.conferencedate2023-06-04
dc.source.conferencelocationRhodos
dc.source.journalICASSP 2023 - 2023 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP
dc.source.numberofpages5
dc.subject.keywordsINVERSE PROBLEMS
dc.subject.keywordsNEURAL-NETWORKS
dc.subject.keywordsRECONSTRUCTION
dc.subject.keywordsALGORITHM
dc.subject.keywordsSIGNAL
dc.title

DESIGNING TRANSFORMER NETWORKS FOR SPARSE RECOVERY OF SEQUENTIAL DATA USING DEEP UNFOLDING

dc.typeProceedings paper
dspace.entity.typePublication
imec.internal.crawledAt2026-02-23
imec.internal.sourcecrawler
Files
Publication available in collections: