Semantic analysis of phrasal verbs using BERT for contextual meaning disambiguation
| dc.contributor.author | Ibrahim, S. | |
| dc.contributor.author | Angel, M. | |
| dc.contributor.author | Abishethvarman, V. | |
| dc.contributor.author | Prasanth, S. | |
| dc.contributor.author | Kumara, B.T.G.S. | |
| dc.date.accessioned | 2025-11-05T17:56:00Z | |
| dc.date.available | 2025-11-05T17:56:00Z | |
| dc.date.issued | 2025-11-07 | |
| dc.description.abstract | Phrasal verbs remain a major challenge in natural language processing (NLP) due to their idiomatic, polysemous, and context-sensitive nature. Unlike literal expressions, their meanings are often non-compositional, making them difficult for transformer-based architectures such as Bidirectional Encoder Representations from Transformers (BERT) to interpret reliably. Previous research has applied BERT-based models to verb disambiguation but has rarely addressed phrasal verb semantics. Moreover, limited work has explored parameter-efficient fine-tuning methods specifically for semantics-related NLP tasks. This study bridges the above gap by adapting transformer models through parameter-efficient fine-tuning for improved phrasal verb understanding. The use of Quantised Low-Rank Adaptation (QLoRA) for supervised fine-tuning of BERT wass investigated in this study to enhance phrasal verb disambiguation. A curated dataset of approximately 10,000 phrasal verb instances was constructed and split into 70% training, 15% validation, and 15% test sets. The base BERT model was first evaluated to establish baseline performance, followed by QLoRA fine-tuning. Model performance was assessed using multiple lexical and semantic evaluation metrics, including cosine similarity, Bilingual Evaluation Understudy (BLEU), Recall-Oriented Understudy for Gisting Evaluation–Longest Common Subsequence (ROUGE-L), Jaccard similarity, and Metric for Evaluation of Translation with Explicit ORdering (METEOR). Across all metrics, QLoRA fine-tuning yielded measurable improvements: cosine similarity increased from 0.5889 to 0.6189, BLEU from 0.2570 to 0.3150, and ROUGE-L from 0.4623 to 0.4901. These results demonstrate that lightweight supervised adaptation enhances BERT’s ability to capture phrasal verb semantics while reducing computational overhead. The findings show that a fine-tuned foundational transformer model can excel in this baseline setting, indicating potential scalability to larger language models. This work advances idiomatic language processing in NLP and highlights the broader promise of parameter-efficient fine-tuning for resource-constrained semantic understanding tasks. | |
| dc.identifier.citation | Proceedings of the Postgraduate Institute of Science Research Congress (RESCON)-2025, University of Peradeniya,p92 | |
| dc.identifier.issn | 3051-4622 | |
| dc.identifier.uri | https://ir.lib.pdn.ac.lk/handle/20.500.14444/5997 | |
| dc.language.iso | en | |
| dc.publisher | Postgraduate Institute of Science (PGIS), University of Peradeniya, Sri Lanka | |
| dc.relation.ispartofseries | Volume 12 | |
| dc.subject | BERT | |
| dc.subject | Context-aware | |
| dc.subject | Fine-tuning | |
| dc.subject | Phrasal verbs | |
| dc.subject | QLoRA | |
| dc.subject | Semantic classification | |
| dc.title | Semantic analysis of phrasal verbs using BERT for contextual meaning disambiguation | |
| dc.type | Article |