Optimizing diplomatic indexing: full-parameter vs low-rank adaptation for multi-label classification of diplomatic cables
Computer Science and Information Technologies

Abstract
Accurate classification of diplomatic cables is crucial for Mission’s evaluation and policy formulation. However, these documents often cover multiple topics, hence a multi-label classification approach is necessary. This research explores the application of pre-trained language models (CahyaBERT, IndoBERT, and MBERT) for multi-label classification of diplomatic cable executive summaries, which align with the diplomatic representation index. The study compares full-parameter fine-tuning and low-rank adaptation (LoRA) techniques using cables from 2022-2023. Results demonstrate that Indonesian-specific models, particularly the IndoBERT, outperform multilingual models in classification accuracy. While LoRA showed slightly lower performance than full fine-tuning, it significantly reduced GPU memory usage by 48% and training time by 69.7%. These findings highlight LoRA’s potential for resource-constrained diplomatic institutions, advancing natural language processing in diplomacy and offering pathways for efficient, real-time multi-label classification to enhance diplomatic mission evaluation.
Discover Our Library
Embark on a journey through our expansive collection of articles and let curiosity lead your path to innovation.
