Benchmark of Self-supervised Graph Neural Networks

dc.contributorAalto-yliopistofi
dc.contributorAalto Universityen
dc.contributor.advisorVerma, Vikas
dc.contributor.authorWang, Haishan
dc.contributor.schoolPerustieteiden korkeakoulufi
dc.contributor.supervisorSolin, Arno
dc.date.accessioned2022-08-30T14:03:22Z
dc.date.available2022-08-30T14:03:22Z
dc.date.issued2022-07-29
dc.description.abstractA graph is an abstract data structure with abundant applications, such as social networks, biochemical molecules, and traffic maps. Graph neural networks (GNNs), deep learning tools which adapt to irregular non-Euclidean space, are designed for such graph data with heavy reliance on manual labels. Learning generalizable and reliable representation for unlabeled graph-structured data has become an attractive and trending task in academia because of the promising application scenarios. Recently, numerous SSL-GNN algorithms have been proposed with success on this task. However, the proposed methods are often evaluated with different architecture and evaluation processes on different small-scale datasets, resulting in unreliable model comparisons. To address this problem, this thesis aims to build a benchmark with a unified framework, a standard evaluation process, and replaceable blocks. In this thesis, a benchmark of SSL-GNNs algorithms is created with the implementation of 9 state-of-art algorithms. These algorithms are compared on this benchmark with consistent settings: shared structure of the GNN encoder, pre-training and fine-tuning scheme, and a unified evaluation protocol. Each model is pre-trained on large-scale datasets: ZINC-15 with two million molecular data and then fine-tuned on eight biophysical downstream datasets for the graph classification task. The experiment results support that two of the nine algorithms outperform others under the benchmark set. Furthermore, the comparison between algorithms also shows the correlation between the pre-training dataset and certain fine-tuning datasets, and the correlation is analyzed by the model mechanisms. The implemented benchmark and discoveries in this thesis are expected to promote transfer learning on graph representation learning.en
dc.format.extent52+2
dc.format.mimetypeapplication/pdfen
dc.identifier.urihttps://aaltodoc.aalto.fi/handle/123456789/116441
dc.identifier.urnURN:NBN:fi:aalto-202208305254
dc.language.isoenen
dc.programmeMaster’s Programme in Computer, Communication and Information Sciencesfi
dc.programme.majorMachine Learning, Data Science and Artificial Intelligence (Macadamia)fi
dc.programme.mcodeSCI3044fi
dc.subject.keywordmachine learningen
dc.subject.keywordbenchmarken
dc.subject.keywordgraph neural networksen
dc.subject.keywordself-supervised learningen
dc.subject.keywordpre-training and fine-tuningen
dc.titleBenchmark of Self-supervised Graph Neural Networksen
dc.typeG2 Pro gradu, diplomityöfi
dc.type.ontasotMaster's thesisen
dc.type.ontasotDiplomityöfi
local.aalto.electroniconlyyes
local.aalto.openaccessyes

Files

Original bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
master_Wang_Haishan_2022.pdf
Size:
703.21 KB
Format:
Adobe Portable Document Format