Title: | Algorithms for Data-Efficient Training of Deep Neural Networks |
Author(s): | Verma, Vikas |
Date: | 2020 |
Language: | en |
Pages: | 97 + app. 131 |
Department: | Tietotekniikan laitos Department of Computer Science |
ISBN: | 978-952-64-0160-7 (electronic) 978-952-64-0159-1 (printed) |
Series: | Aalto University publication series DOCTORAL DISSERTATIONS, 198/2020 |
ISSN: | 1799-4942 (electronic) 1799-4934 (printed) 1799-4934 (ISSN-L) |
Supervising professor(s): | Kannala, Juho, Prof., Aalto University, Department of Computer Science, Finland |
Thesis advisor(s): | Bengio, Yoshua, Prof., Mila - Quebec AI Institute (Mila - Institut québécois d'intelligence artificielle), Canada; Raiko, Tapani, Prof., Aalto University, Department of Computer Science, Finland; Karhunen, Juha, Prof., Aalto University, Department of Computer Science, Finland |
Subject: | Computer science |
Keywords: | deep neural networks, machine learning |
Archive | yes |
|
|
Abstract:Deep Neural Networks ("deep learning") have become a ubiquitous choice of algorithms for Machine Learning applications. These systems often achieve human-level or even super-human level performances across a variety of tasks such as computer vision, natural language processing, speech recognition, reinforcement learning, generative modeling and healthcare. This success can be attributed to their ability to learn complex representations directly from the raw input data, completely eliminating the hand-crafted feature extraction from the pipeline. However, there still exists a caveat: due to the extremely large number of trainable parameters in Deep Neural Networks, their generalization ability depends heavily on the availability of a large amount of labeled data.
|
|
Parts:[Publication 1]: Vikas Verma, Alex Lamb, Christopher Beckham, Amir Najafi, Ioannis Mitliagkas, David Lopez-Paz, Yoshua Bengio. Manifold Mixup: Better Representations by Interpolating Hidden States. In Proceedings of the 36th International Conference on Machine Learning (ICML 2019), Long Beach, California, USA, volume 97, pages: 6438–6447, 2019. Full text in Acris/Aaltodoc:http://urn.fi/URN:NBN:fi:aalto-201911156276. [Publication 2]: Christopher Beckham, Sina Honari, Vikas Verma, Alex Lamb, Farnoosh Ghadiri, R Devon Hjelm, Yoshua Bengio. On Adversarial Mixup Resynthesis. In 2019 Conference on Neural Information Processing Systems (NeurIPS 2019), Vancouver, Canada, pages:4346–4357, 2019.[Publication 3]: Vikas Verma, Alex Lamb, Juho Kannala, Yoshua Bengio. Interpolated Adversarial Training: Achieving Robust Neural Networks Without Sacrificing Too Much Accuracy. In Proceedings of the 12th ACM Workshop on Artificial Intelligence and Security (AISec’19), London, United Kingdom, pages:95-103, 2019. Full text in Acris/Aaltodoc: http://urn.fi/URN:NBN:fi:aalto-202001171609. DOI: 10.1145/3338501.3357369 View at Publisher [Publication 4]: Vikas Verma, Alex Lamb, Juho Kannala, Yoshua Bengio, David Lopez-Paz. Interpolation Consistency Training for Semi-Supervised Learning. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence (IJCAI-19), Macao, China, pages:3635–3641, 2019. DOI: 10.24963/ijcai.2019/504 View at Publisher [Publication 5]: Vikas Verma, Meng Qu, Alex Lamb, Yoshua Bengio, Juho Kannala, Jian Tang. GraphMix: Improved Training of Graph Neural Networks for Semi-Supervised Learning. Submitted for review, https://arxiv.org/pdf/1909.11715.pdf, January 2020, 8 pages[Publication 6]: Fan-Yun Sun, Jordan Hoffmann, Vikas Verma, Jian Tang. InfoGraph: Unsupervised and Semi-supervised Graph-Level Representation Learning via Mutual Information Maximization. In Eighth International Conference on Learning Representations (ICLR 2020, spotlight), Addis Ababa, Ethiopia, 2020[Publication 7]: Stanislaw Jastrzebski, Devansh Arpit, Nicolas Ballas, Vikas Verma, Tong Che, Yoshua Bengio. Residual Connections Encourage Iterative Inference. In 6th International Conference on Learning Representations (ICLR 2018), Vancouver, Canada, 2018 |
|
|
Unless otherwise stated, all rights belong to the author. You may download, display and print this publication for Your own personal use. Commercial use is prohibited.
Page content by: Aalto University Learning Centre | Privacy policy of the service | About this site