hh.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Establishing strong imputation performance of a denoising autoencoder in a wide range of missing data problems
Department of Astronomy and Theoretical Physics, Lund University, Lund, Sweden.
Department of Astronomy and Theoretical Physics, Lund University, Lund, Sweden.
Department of Astronomy and Theoretical Physics, Lund University, Lund, Sweden.
Halmstad University, School of Information Technology, Halmstad Embedded and Intelligent Systems Research (EIS), CAISR - Center for Applied Intelligent Systems Research. Department of Astronomy and Theoretical Physics, Lund University, Lund, Sweden.ORCID iD: 0000-0003-1145-4297
2019 (English)In: Neurocomputing, ISSN 0925-2312, E-ISSN 1872-8286, Vol. 65, p. 137-146Article in journal (Refereed) Published
Abstract [en]

Dealing with missing data in data analysis is inevitable. Although powerful imputation methods that address this problem exist, there is still much room for improvement. In this study, we examined single imputation based on deep autoencoders, motivated by the apparent success of deep learning to efficiently extract useful dataset features. We have developed a consistent framework for both training and imputation. Moreover, we benchmarked the results against state-of-the-art imputation methods on different data sizes and characteristics. The work was not limited to the one-type variable dataset; we also imputed missing data with multi-type variables, e.g., a combination of binary, categorical, and continuous attributes. To evaluate the imputation methods, we randomly corrupted the complete data, with varying degrees of corruption, and then compared the imputed and original values. In all experiments, the developed autoencoder obtained the smallest error for all ranges of initial data corruption. © 2019 Elsevier B.V.

Place, publisher, year, edition, pages
Amsterdam: Elsevier, 2019. Vol. 65, p. 137-146
Keywords [en]
Deep learning, Autoencoder, Imputation, Missing data
National Category
Other Computer and Information Science
Identifiers
URN: urn:nbn:se:hh:diva-41245DOI: 10.1016/j.neucom.2019.07.065ISI: 000484072600014Scopus ID: 2-s2.0-85069939556OAI: oai:DiVA.org:hh-41245DiVA, id: diva2:1378263
Funder
Swedish Foundation for Strategic Research Available from: 2019-12-13 Created: 2019-12-13 Last updated: 2019-12-13Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records BETA

Ohlsson, Mattias

Search in DiVA

By author/editor
Ohlsson, Mattias
By organisation
CAISR - Center for Applied Intelligent Systems Research
In the same journal
Neurocomputing
Other Computer and Information Science

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 18 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf