Optimasi IndoBERT untuk Pengenalan Entitas Bernama Bahasa Indonesia pada Data Media Sosial dengan Penalaan Hiperparameter Optuna
DOI:
https://doi.org/10.30865/jurikom.v13i1.9545Keywords:
Named Entity Recognition, IndoBERT, Indonesian Language, Hyperparameter Optimization, OptunaAbstract
Named Entity Recognition (NER) merupakan salah satu tugas fundamental dalam pemrosesan bahasa alami yang berperan penting dalam ekstraksi informasi terstruktur dari teks tidak terstruktur. Pada Bahasa Indonesia, kinerja model NER berbasis pre-trained BERT sangat dipengaruhi oleh konfigurasi hiperparameter pada tahap fine-tuning. Namun, banyak penelitian masih menggunakan konfigurasi bawaan atau penyesuaian terbatas, sehingga potensi peningkatan kinerja dan stabilitas model belum sepenuhnya dimanfaatkan. Penelitian ini bertujuan untuk mengevaluasi dampak optimasi hiperparameter berbasis Optuna terhadap kinerja dan stabilitas pelatihan model pre-trained BERT untuk tugas NER Bahasa Indonesia. Model yang digunakan adalah IndoBERT (indobenchmark/indobert-base-p1) yang difine-tune untuk mengenali entitas Person (PER), Organization (ORG), dan Location (LOC) dengan skema pelabelan BIO. Metode optimasi hiperparameter dilakukan menggunakan pendekatan Bayesian berbasis
Named Entity Recognition (NER) is a fundamental task in natural language processing for extracting structured information from unstructured text. In Indonesian, particularly for informal and diverse social media text, the performance of NER models based on Bidirectional Encoder Representations from Transformers (BERT) is strongly influenced by hyperparameter configurations during fine-tuning. However, many studies still rely on default settings or limited adjustments, so the potential improvements in performance and training stability have not been fully exploited. This study evaluates the impact of hyperparameter tuning using Optuna with a Tree-structured Parzen Estimator (TPE) on the performance and training stability of IndoBERT (indobenchmark/indobert-base-p1) on Twitter/X data. The main contribution of this work is an empirical evaluation of how hyperparameter tuning improves IndoBERT’s performance and training stability, and the resulting recommendations of reliable configurations for reproducible experiments and practical deployment of Indonesian NER. The dataset is annotated using the Begin–Inside–Outside (BIO) labeling scheme for three entity types: person (PER), organization (ORG), and location (LOC). The optimization objective is defined as the F1-score on the validation set. The results show that the Optuna configuration achieves a precision of 0.9338, recall of 0.9312, F1-score of 0.9325, and accuracy of 0.9854 on the test set, outperforming the baseline with an F1-score of 0.9253 and accuracy of 0.9837. Multi-seed evaluation indicates consistent improvements, with an average F1 of 0.9302 ± 0.0016 compared to 0.9238 ± 0.0009 for the baseline. These findings confirm that Optuna-based hyperparameter tuning improves both the performance and reliability of IndoBERT for Indonesian NER on social media text.
References
[1] J. Li, A. Sun, J. Han, and C. Li, “A Survey on Deep Learning for Named Entity Recognition,” 2020.
[2] J. Dodge, G. Ilharco, R. Schwartz, A. Farhadi, H. Hajishirzi, and N. Smith, “Fine-Tuning Pretrained Language Models: Weight Initializations, Data Orders, and Early Stopping,” 2019.
[3] M. B. Shishehgarkhaneh, R. C. Moehler, Y. Fang, A. A. Hijazi, and H. Aboutorab, “Transformer-Based Named Entity Recognition in Construction Supply Chain Risk Management in Australia,” IEEE Access, vol. 12, no. March, pp. 41829–41851, 2024, doi: 10.1109/ACCESS.2024.3377232.
[4] S. O. Khairunnisa, “Dataset Enhancement and Multilingual Transfer for Named Entity Recognition in the Indonesian Language Dataset Enhancement and Multilingual Transfer for Named,” vol. 22, no. 6, 2026, doi: 10.1145/3592854.
[5] F. Koto and T. Baldwin, “IndoLEM and IndoBERT : A Benchmark Dataset and Pre-trained Language Model for Indonesian NLP,” pp. 757–770, 2020.
[6] T. Aprilah, D. R. Ignatius, M. Setiadi, and W. Herowati, “Enhancing Aspect-Based Sentiment Analysis via Hugging Face Fine- Tuned IndoBERT,” vol. 9, no. 6, pp. 3821–3830, 2025.
[7] T. Akiba, S. Sano, T. Yanase, T. Ohta, and M. Koyama, “Optuna: A Next-generation Hyperparameter Optimization Framework,” Proc. ACM SIGKDD Int. Conf. Knowl. Discov. Data Min., pp. 2623–2631, 2019, doi: 10.1145/3292500.3330701.
[8] M. Khaldi, Z. Alilat, H. Bendoubba, and N. Mahammed, “Hyperparameter Optimization for Malicious URL Detection : Leveraging Optuna and Random Search in Machine Learning and Deep Learning Models 1 Introduction Related research,” vol. 49, pp. 57–68, 2025.
[9] D. Klakow, “THE S TABILITY OF F INE - TUNING BERT : M ISCON -,” 2021.
[10] B. Jehangir, S. Radhakrishnan, and R. Agarwal, “A survey on Named Entity Recognition — datasets, tools, and methodologies,” Nat. Lang. Process. J., vol. 3, p. 100017, 2023, doi: https://doi.org/10.1016/j.nlp.2023.100017.
[11] “Bio Tagging Ner Bahasa Indonesia (from twitter/X).” Accessed: Dec. 01, 2025. [Online]. Available: https://www.kaggle.com/datasets/gibranfktian/bio-tagging-ner-bahasa-indonesia-crawl-twitterx/data
[12] H. Liu, G. Jun, and Y. Zheng, “Chinese named entity recognition model based on BERT,” vol. 06021, pp. 1–8, 2021.
[13] Y. Ling, Z. Ma, X. Dong, and X. Weng, “A deep learning approach for robust traffic accident information extraction from online chinese news,” no. February, pp. 1847–1862, 2024, doi: 10.1049/itr2.12493.
[14] E. T. Luthfi, Z. Izzah, M. Yusoh, and B. M. Aboobaider, “BERT based Named Entity Recognition for Automated Hadith Narrator Identification,” vol. 13, no. 1, pp. 604–611, 2022.
[15] B. S. Jati and M. N. Rizal, “Multilingual Named Entity Recognition Model for Indonesian Health Insurance Question Answering System,” pp. 180–184, doi: 10.1109/ICOIACT50329.2020.9332027.
[16] Z. Liu, F. Jiang, Y. Hu, C. Shi, P. Fung, and A. Group, “NER-BERT: A Pre-trained Model for Low-Resource Entity Tagging,” 2020.
[17] W. Liu and X. Cui, “applied sciences Improving Named Entity Recognition for Social Media with Data Augmentation,” 2023.
[18] K. Kamdan, M. P. Anugrah, M. J. Almutaali, R. Ramdani, and I. L. Kharisma, “Performance Analysis of IndoBERT for Detection of Online Gambling Promotion in YouTube Comments †,” 2025.
[19] B. Wilie et al., “IndoNLU : Benchmark and Resources for Evaluating Indonesian Natural Language Understanding,” pp. 843–857, 2020.
[20] M. C. Kenton, L. Kristina, and J. Devlin, “BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding,” no. Mlm, 1953.
[21] J. Bergstra and Y. Bengio, “Random search for hyper-parameter optimization,” J. Mach. Learn. Res., vol. 13, pp. 281–305, 2012.
[22] T. Wolf et al., “Transformers : State-of-the-Art Natural Language Processing,” pp. 38–45, 2020.
[23] S. Latisha, S. Favian, and D. Suhartono, “Criminal Court Judgment Prediction System Built on Modified BERT Models,” vol. 15, no. 2, 2024, doi: 10.12720/jait.15.2.288-298.
[24] V. Yadav and S. Bethard, “A Survey on Recent Advances in Named Entity Recognition from Deep Learning models,” pp. 2145–2158, 2018.



