Optimizing diplomatic indexing: full-parameter vs low-rank adaptation for multi-label classification of diplomatic cables

Computer Science and Information Technologies

Optimizing diplomatic indexing: full-parameter vs low-rank adaptation for multi-label classification of diplomatic cables

Abstract

Accurate classification of diplomatic cables is crucial for Mission’s evaluation and policy formulation. However, these documents often cover multiple topics, hence a multi-label classification approach is necessary. This research explores the application of pre-trained language models (CahyaBERT, IndoBERT, and MBERT) for multi-label classification of diplomatic cable executive summaries, which align with the diplomatic representation index. The study compares full-parameter fine-tuning and low-rank adaptation (LoRA) techniques using cables from 2022-2023. Results demonstrate that Indonesian-specific models, particularly the IndoBERT, outperform multilingual models in classification accuracy. While LoRA showed slightly lower performance than full fine-tuning, it significantly reduced GPU memory usage by 48% and training time by 69.7%. These findings highlight LoRA’s potential for resource-constrained diplomatic institutions, advancing natural language processing in diplomacy and offering pathways for efficient, real-time multi-label classification to enhance diplomatic mission evaluation.

Discover Our Library

Embark on a journey through our expansive collection of articles and let curiosity lead your path to innovation.

Explore Now
Library 3D Ilustration