Hybird prompt optimization without fine-tuning: Enhancing information extraction and translation in vehicle maintenance records
| dc.contributor | Aalto-yliopisto | fi |
| dc.contributor | Aalto University | en |
| dc.contributor.advisor | Hedman, Anders | |
| dc.contributor.advisor | Li, Haibo | |
| dc.contributor.author | Wu, Zhiyuan | |
| dc.contributor.school | Sähkötekniikan korkeakoulu | fi |
| dc.contributor.school | School of Electrical Engineering | en |
| dc.contributor.supervisor | Zhou, Quan | |
| dc.date.accessioned | 2025-08-19T17:11:21Z | |
| dc.date.available | 2025-08-19T17:11:21Z | |
| dc.date.issued | 2025-06-10 | |
| dc.description.abstract | Unstructured, multilingual maintenance logs hamper automated fault analysis, yet fully fine-tuning large language models (LLMs) is prohibitively costly. We therefore adapt a 70 B-parameter LLaMA-33 model without gradient updates using a two-step Hybrid Prompt Optimization (HPO) scheme that merges expert-written instructions with DSPy’s automatic prompt search. The approach is trained and tested on 1000 real, Volvo Trucks records while running only on CPU. For evaluation we draw 10 disjoint test sets of 50 records each; scores reported here are the mean of those ten runs. Relative to a zero-shot baseline, manual prompts raise structured-field extraction accuracy by 23 %; DSPy adds 27 %; and HPO supplies a further 5 %, yielding a cumulative gain of ≈ 60 %. For full-text translation the same steps deliver 55 %, +2 %, and +2 %, respectively. Paired two-tailed t-tests across the 10 × 50 predictions confirm that all improvements remain significant at α = 0.05. Under the best prompt, the lightweight LLaMA-33-70B reaches 92 % of GPT-4o’s extraction quality and slightly surpasses it in translation—yet consumes only a fraction of the compute budget. These findings show that prompt engineering alone can unlock near–state-of-the-art multilingual extraction and translation, delivering a low-cost alternative to full model fine-tuning for industrial after-sales data. | en |
| dc.format.extent | 76 | |
| dc.format.mimetype | application/pdf | en |
| dc.identifier.uri | https://aaltodoc.aalto.fi/handle/123456789/138120 | |
| dc.identifier.urn | URN:NBN:fi:aalto-202508196349 | |
| dc.language.iso | en | en |
| dc.location | P1 | fi |
| dc.programme | Master's Programme in ICT Innovation | en |
| dc.programme.major | Autonomous Systems | en |
| dc.subject.keyword | prompt engineering | en |
| dc.subject.keyword | large language models | en |
| dc.subject.keyword | hybrid optimization | en |
| dc.subject.keyword | information extraction | en |
| dc.subject.keyword | multilingual translation | en |
| dc.subject.keyword | automotive industry | en |
| dc.title | Hybird prompt optimization without fine-tuning: Enhancing information extraction and translation in vehicle maintenance records | en |
| dc.type | G2 Pro gradu, diplomityö | fi |
| dc.type.ontasot | Master's thesis | en |
| dc.type.ontasot | Diplomityö | fi |
| local.aalto.electroniconly | yes | |
| local.aalto.openaccess | yes |
Files
Original bundle
1 - 1 of 1