Benchmark of Self-supervised Graph Neural Networks
dc.contributor | Aalto-yliopisto | fi |
dc.contributor | Aalto University | en |
dc.contributor.advisor | Verma, Vikas | |
dc.contributor.author | Wang, Haishan | |
dc.contributor.school | Perustieteiden korkeakoulu | fi |
dc.contributor.supervisor | Solin, Arno | |
dc.date.accessioned | 2022-08-30T14:03:22Z | |
dc.date.available | 2022-08-30T14:03:22Z | |
dc.date.issued | 2022-07-29 | |
dc.description.abstract | A graph is an abstract data structure with abundant applications, such as social networks, biochemical molecules, and traffic maps. Graph neural networks (GNNs), deep learning tools which adapt to irregular non-Euclidean space, are designed for such graph data with heavy reliance on manual labels. Learning generalizable and reliable representation for unlabeled graph-structured data has become an attractive and trending task in academia because of the promising application scenarios. Recently, numerous SSL-GNN algorithms have been proposed with success on this task. However, the proposed methods are often evaluated with different architecture and evaluation processes on different small-scale datasets, resulting in unreliable model comparisons. To address this problem, this thesis aims to build a benchmark with a unified framework, a standard evaluation process, and replaceable blocks. In this thesis, a benchmark of SSL-GNNs algorithms is created with the implementation of 9 state-of-art algorithms. These algorithms are compared on this benchmark with consistent settings: shared structure of the GNN encoder, pre-training and fine-tuning scheme, and a unified evaluation protocol. Each model is pre-trained on large-scale datasets: ZINC-15 with two million molecular data and then fine-tuned on eight biophysical downstream datasets for the graph classification task. The experiment results support that two of the nine algorithms outperform others under the benchmark set. Furthermore, the comparison between algorithms also shows the correlation between the pre-training dataset and certain fine-tuning datasets, and the correlation is analyzed by the model mechanisms. The implemented benchmark and discoveries in this thesis are expected to promote transfer learning on graph representation learning. | en |
dc.format.extent | 52+2 | |
dc.format.mimetype | application/pdf | en |
dc.identifier.uri | https://aaltodoc.aalto.fi/handle/123456789/116441 | |
dc.identifier.urn | URN:NBN:fi:aalto-202208305254 | |
dc.language.iso | en | en |
dc.programme | Master’s Programme in Computer, Communication and Information Sciences | fi |
dc.programme.major | Machine Learning, Data Science and Artificial Intelligence (Macadamia) | fi |
dc.programme.mcode | SCI3044 | fi |
dc.subject.keyword | machine learning | en |
dc.subject.keyword | benchmark | en |
dc.subject.keyword | graph neural networks | en |
dc.subject.keyword | self-supervised learning | en |
dc.subject.keyword | pre-training and fine-tuning | en |
dc.title | Benchmark of Self-supervised Graph Neural Networks | en |
dc.type | G2 Pro gradu, diplomityö | fi |
dc.type.ontasot | Master's thesis | en |
dc.type.ontasot | Diplomityö | fi |
local.aalto.electroniconly | yes | |
local.aalto.openaccess | yes |
Files
Original bundle
1 - 1 of 1
No Thumbnail Available
- Name:
- master_Wang_Haishan_2022.pdf
- Size:
- 703.21 KB
- Format:
- Adobe Portable Document Format