Dual-model augmentation for compositional semantic parsing via recursive log-likelihood guidance
| dc.contributor | Aalto-yliopisto | fi |
| dc.contributor | Aalto University | en |
| dc.contributor.advisor | Spilsbury, Sam | |
| dc.contributor.author | Sherafati, Mohammad | |
| dc.contributor.school | Perustieteiden korkeakoulu | fi |
| dc.contributor.school | School of Science | en |
| dc.contributor.supervisor | Marttinen, Pekka | |
| dc.date.accessioned | 2026-01-19T18:02:50Z | |
| dc.date.available | 2026-01-19T18:02:50Z | |
| dc.date.issued | 2025-12-26 | |
| dc.description.abstract | Compositional generalization remains a significant challenge for neural semantic parsers, particularly in low-resource settings. Data augmentation is a powerful method to address this by improving downstream task performance through expanded training diversity. While Large Language Models (LLMs) offer a promising avenue for such augmentation, their tendency to "hallucinate" invalid structures limits their utility in domains requiring strict syntactic adherence, such as the SMCalFlow dataset. This thesis proposes a Dual-Model Augmentation Cycle to resolve the trade-off between generative diversity and structural correctness. We decouple generation, assigned to a Large Language Model (Gemini), from verification, assigned to a fine-tuned small model (T5). Leveraging the small model as a likelihood estimate of the training distribution, we implement a "Judger" that evaluates synthetic examples using normalized log-likelihood scores. Low-probability sequences trigger an automated self-correction loop, where the Generator refines the output based on targeted, token-level feedback. We evaluate this framework on SMCalFlow, Overnight, and GeoQuery. Results demonstrate that our method significantly outperforms strong heuristic and grammar-induction baselines. On SMCalFlow (32-shot), our approach improves exact match accuracy from 37.0% to 52.0%,surpassing GECA (44.0\%) and QCFG (45.0%). Furthermore, in the ultra-low-resource Overnight domain, our method prevents negative transfer, improving accuracy to 19.0% where heuristic methods degraded performance. Distributional analysis confirms that the Judger effectively aligns the synthetic data with the natural data distribution. These findings suggest that small, specialized models act as indispensable critics for large foundation models, unlocking robust compositional generalization. | en |
| dc.format.extent | 78 | |
| dc.format.mimetype | application/pdf | en |
| dc.identifier.uri | https://aaltodoc.aalto.fi/handle/123456789/142061 | |
| dc.identifier.urn | URN:NBN:fi:aalto-202601191437 | |
| dc.language.iso | en | en |
| dc.programme | Master's Programme in Life Science Technologies | en |
| dc.programme | Master's Programme in Life Science Technologies | fi |
| dc.programme | Master's Programme in Life Science Technologies | sv |
| dc.programme.major | Machine Learning, Data Science and Artificial Intelligence | en |
| dc.subject.keyword | Large Language Models (LLMs) | en |
| dc.subject.keyword | semantic parsing | en |
| dc.subject.keyword | data augmentation | en |
| dc.subject.keyword | compositional generalization | en |
| dc.subject.keyword | self-correction | en |
| dc.subject.keyword | low-resource NLP | en |
| dc.subject.keyword | small model as judger | en |
| dc.title | Dual-model augmentation for compositional semantic parsing via recursive log-likelihood guidance | en |
| dc.type | G2 Pro gradu, diplomityö | fi |
| dc.type.ontasot | Master's thesis | en |
| dc.type.ontasot | Diplomityö | fi |
| local.aalto.electroniconly | yes | |
| local.aalto.openaccess | yes |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- master_Sherafati_Mohammad_2026.pdf
- Size:
- 10.37 MB
- Format:
- Adobe Portable Document Format