A Comparative Study of Classical and Quantum Transformer Models and Their Applications
| dc.contributor | Aalto-yliopisto | fi |
| dc.contributor | Aalto University | en |
| dc.contributor.advisor | Raasakka, Matti | |
| dc.contributor.author | Shah, Jayaditya | |
| dc.contributor.school | Perustieteiden korkeakoulu | fi |
| dc.contributor.supervisor | Raasakka, Matti | |
| dc.date.accessioned | 2024-11-19T09:15:50Z | |
| dc.date.available | 2024-11-19T09:15:50Z | |
| dc.date.issued | 2024-09-29 | |
| dc.description.abstract | Transformers are a neural network architecture that enable the utilisation of a larger context frame than traditional neural networks when training deep learning models. With the advent of the ongoing decade, transformers have enabled the application of large language models, improved computer vision, generative models and other large-scale artificial intelligence systems. This thesis investigates the potential of quantum transformers to implement deep learning tasks focusing on the Quixer model, developed by Quantinuum, by comparing a classically simulated version of Quixer to classical transformers. The thesis is motivated by the need to optimize transformer components for large-scale applications, to address the quadratic complexity arising from the self attention mechanism. The results of the thesis indicate that Quixer performs in line with the classical baseline as published in the paper by Quantinuum when reproduced with the same dataset, and model performance follows the trend for another dataset of twice the size. Hence, providing a proof of concept quantum transformers can be considered an effective method for developing large-scale models in the future with the eventual improvement of quantum hardware. | en |
| dc.format.extent | 36+20 | |
| dc.format.mimetype | application/pdf | en |
| dc.identifier.uri | https://aaltodoc.aalto.fi/handle/123456789/131672 | |
| dc.identifier.urn | URN:NBN:fi:aalto-202411197190 | |
| dc.language.iso | en | en |
| dc.programme | Aalto Bachelor’s Programme in Science and Technology | fi |
| dc.programme.major | Quantum Technology | en |
| dc.programme.mcode | SCI3103 | fi |
| dc.subject.keyword | machine learning | en |
| dc.subject.keyword | physics | en |
| dc.subject.keyword | quantum circuits | en |
| dc.subject.keyword | quantum transformers | en |
| dc.subject.keyword | quantum information | en |
| dc.subject.keyword | transformer architecture | en |
| dc.title | A Comparative Study of Classical and Quantum Transformer Models and Their Applications | en |
| dc.type | G1 Kandidaatintyö | fi |
| dc.type.dcmitype | text | en |
| dc.type.ontasot | Bachelor's thesis | en |
| dc.type.ontasot | Kandidaatintyö | fi |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- Shah_Jayaditya_2024.pdf
- Size:
- 2.15 MB
- Format:
- Adobe Portable Document Format
Download (opens in new window)
Aalto login required (access for Aalto Staff only).