The application of pre-trained transformer models to UK Court of Appeal legal judgments
Conference paper
Abbas, W., Zia, T., Tirunagari, S., Chennareddy, V., Dhami, M. and Windridge, D. 2025. The application of pre-trained transformer models to UK Court of Appeal legal judgments. Quan, T.T., Sombattheera, C., Pham, H.-A. and Tran, N.T. (ed.) 18th International Conference on Multi-disciplinary Trends in Artificial Intelligence. Ho Chi Minh City, Vietnam 03 - 05 Dec 2025 Singapore Springer. pp. 252-264 https://doi.org/10.1007/978-981-95-4963-4_21
| Type | Conference paper |
|---|---|
| Title | The application of pre-trained transformer models to UK Court of Appeal legal judgments |
| Authors | Abbas, W., Zia, T., Tirunagari, S., Chennareddy, V., Dhami, M. and Windridge, D. |
| Abstract | The emergence of Transformer-based Pre-trained Language Models (PLMs) has had a significant impact across a variety of Natural Language Processing (NLP) domains. Pre-training language models on curated legal corpora can assist researchers in developing models to improve performance on downstream legal NLP tasks. This paper reports experiments with pre-training and fine-tuning language models on British and Irish Legal Information Institute (BAILII) data, addressing specifically Appeal Court judgments (these being, in effect, meta-judgments on previous legal judgments). We pre-train BERT-based language models on this corpus and evaluate their effectiveness on BAILII-based domain-specific tasks such as named entity recognition (NER), multi-label classification (MLC), and question answering (QA). The performance of this is then compared to baseline RoBERTa (Robustly Optimized BERT Pretraining Approach) and DistilRoBERTa models with two publicly available PLMs specifically designed for legal text (i.e., LegalBERT and CaseLawBERT), all pre-trained on the BAILII dataset. Pre-training on BAILII improves the performance of the PLMs on downstream tasks, and domain-specific pre-training enables a relatively smaller model such as BERT to achieve performance at par with a larger model such as RoBERTa. The pre-trained PLMs are now publicly available for downstream tasks on BAILII. |
| Keywords | LegalAI; Large Language Models; Pretraining; Generative Answering |
| Sustainable Development Goals | 16 Peace, justice and strong institutions |
| Middlesex University Theme | Creativity, Culture & Enterprise |
| Research Group | Artificial Intelligence group |
| Conference | 18th International Conference on Multi-disciplinary Trends in Artificial Intelligence |
| Page range | 252-264 |
| Proceedings Title | Multi-disciplinary Trends in Artificial Intelligence: 18th International Conference, MIWAI 2025, Ho Chi Minh City, Vietnam, December 3–5, 2025, Proceedings, Part III |
| Series | Multi-disciplinary Trends in Artificial Intelligence |
| Editors | Quan, T.T., Sombattheera, C., Pham, H.-A. and Tran, N.T. |
| ISSN | 0302-9743 |
| Electronic | 1611-3349 |
| ISBN | |
| Paperback | 9789819549627 |
| Electronic | 9789819549634 |
| Publisher | Springer |
| Place of publication | Singapore |
| Copyright Year | 2026 |
| Publication dates | |
| Online | 22 Nov 2025 |
| Publication process dates | |
| Submitted | 2025 |
| Accepted | Sep 2025 |
| Deposited | 07 Jan 2026 |
| Output status | Published |
| Accepted author manuscript | License File Access Level Open |
| Copyright Statement | For the purpose of open access, the author has applied a Creative Commons Attribution (CC BY) licence to any Author Accepted Manuscript version arising from this submission. |
| Digital Object Identifier (DOI) | https://doi.org/10.1007/978-981-95-4963-4_21 |
| Web address (URL) of conference proceedings | https://doi.org/10.1007/978-981-95-4963-4 |
| Related Output | |
| Has version | https://zenodo.org/records/18162842 |
https://repository.mdx.ac.uk/item/324701
Download files
43
total views12
total downloads11
views this month4
downloads this month