Logo image
Open-Source Large Language Models Excel in Named Entity Recognition
Conference proceeding   Peer reviewed

Open-Source Large Language Models Excel in Named Entity Recognition

Dengya Zhu, Sirui Li, Nik Thompson and Kok Wai Wong
Neural Information Processing (ICONIP 2024), Vol.2295, pp.313-326
Communications in Computer and Information Science
Neural Information Processing 31st International Conference (ICONIP 2024) (Auckland, New Zealand, 02/12/2024–06/12/2024)
2025

Abstract

Close source software Evaluation Large language models Named entity recognition Natural language processing Open source software
Current state-of-the-art Named Entity Recognition (NER) typically involves fine-tuning transformer-based models like BERT or RoBERTa with annotated datasets, posing challenges in annotation cost, model robustness, and data privacy. An emerging approach uses pre-trained Large Language Models (LLMs) such as ChatGPT to extract entities directly with a few or zero examples, achieving performance comparable to fine-tuned models. However, reliance on the close-source commercial LLMs raises cost and privacy concerns. In this work, we investigate open-source LLMs like Llama2 for NER on local consumer-grade GPUs, aiming to significantly reduce costs compared to cloud solutions while ensuring data security. Experimental results demonstrate competitive NER performance, achieving F1 85.37% on the CoNLL03 dataset and can also be generalised to specific domains, such as scientific texts.

Details

Metrics

19 Record Views
Logo image