| dc.contributor.author | Abishethvarman, V. | |
| dc.contributor.author | Luxshi, K. | |
| dc.contributor.author | Sujair, I. | |
| dc.contributor.author | Prasanth, S. | |
| dc.contributor.author | Kumara, B.T.G.S. | |
| dc.date.accessioned | 2026-03-07T03:55:27Z | |
| dc.date.available | 2026-03-07T03:55:27Z | |
| dc.date.issued | 2025 | |
| dc.identifier.uri | http://drr.vau.ac.lk/handle/123456789/1931 | |
| dc.description.abstract | Phrasal verbs pose a significant chal lenge in natural language understanding due to their context-dependent meanings. This study investi gates the impact of fine-tuning BERT on the task of context-aware phrasal verb disambiguation. Using a curated dataset of phrasal verbs with annotated meanings, we first evaluate a baseline model’s per formance on semantic similarity and text generation metrics. Subsequently, we fine-tune BERT on the dataset and re-assess its effectiveness. The results demonstrate consistent improvements across all met rics after fine-tuning: Cosine Similarity increased from 0.5889 to 0.6189, BLEU Score improved from 0.2570 to 0.3150, ROUGE-L rose from 0.4623 to 0.4901, Jaccard Similarity from 0.3484 to 0.3648, and METEORfrom0.3555 to 0.3607. These findings highlight that fine-tuning BERT enhances its ability to capture the nuanced meanings of phrasal verbs in context, which is critical for advancing semantic classification tasks. Future work will extend this approach to idiomatic expressions and broader lin guistic ambiguity. | en_US |
| dc.language.iso | en | en_US |
| dc.publisher | Faculty of Applied Science University of Vavuniya Sri Lanka | en_US |
| dc.subject | BERT | en_US |
| dc.subject | Transformer Models | en_US |
| dc.subject | Phrasal Verbs | en_US |
| dc.subject | Semantic Classification | en_US |
| dc.subject | Con textual Embeddings | en_US |
| dc.title | Transformer-Based Approach to Contextual Phrasal Verb Classification | en_US |
| dc.type | Conference full paper | en_US |
| dc.identifier.proceedings | 1st International Conference on Applied Sciences- 2025 | en_US |