dc.contributor.author | Grósz, Tamás | en_US |
dc.contributor.author | Kurimo, Mikko | en_US |
dc.contributor.editor | Archambault, Daniel and Nabney, Ian and Peltonen, Jaakko | en_US |
dc.date.accessioned | 2020-05-24T13:27:44Z | |
dc.date.available | 2020-05-24T13:27:44Z | |
dc.date.issued | 2020 | |
dc.identifier.isbn | 978-3-03868-113-7 | |
dc.identifier.uri | https://doi.org/10.2312/mlvis.20201103 | |
dc.identifier.uri | https://diglib.eg.org:443/handle/10.2312/mlvis20201103 | |
dc.description.abstract | In the past few years, Deep Neural Networks (DNN) have become the state-of-the-art solution in several areas, including automatic speech recognition (ASR), unfortunately, they are generally viewed as black boxes. Recently, this started to change as researchers have dedicated much effort into interpreting their behavior. In this work, we concentrate on visual interpretation by depicting the hidden activation vectors of the DNN, and propose the usage of deep Autoencoders (DAE) to transform these hidden representations for inspection. We use multiple metrics to compare our approach with other, widely-used algorithms and the results show that our approach is quite competitive. The main advantage of using Autoencoders over the existing ones is that after the training phase, it applies a fixed transformation that can be used to visualize any hidden activation vector without any further optimization, which is not true for the other methods. | en_US |
dc.publisher | The Eurographics Association | en_US |
dc.subject | Computing methodologies | |
dc.subject | Dimensionality reduction and manifold learning | |
dc.subject | Speech recognition | |
dc.subject | Neural networks | |
dc.title | Visual Interpretation of DNN-based Acoustic Models using Deep Autoencoders | en_US |
dc.description.seriesinformation | Machine Learning Methods in Visualisation for Big Data | |
dc.description.sectionheaders | Papers | |
dc.identifier.doi | 10.2312/mlvis.20201103 | |
dc.identifier.pages | 25-29 | |