Bayesian Tuning terhadap Model Pre-Trained PEGASUS untuk Peringkas Teks Informatif Berbahasa Indonesia

Bayesian Tuning of a Pre-Trained PEGASUS Model for Indonesian Informative Text Summarization

Authors

  • Kadek Artha Darma Pradnyana Universitas Udayana
  • I Nyoman Prayana Trisna Trisna Universitas Udayana
  • Wayan Oger Vihikan Universitas Udayana

DOI:

https://doi.org/10.24002/jbi.v17i1.12915

Keywords:

Peringkas Teks Abstraktif, PEGASUS, Optimasi Bayesian, Input Terformat, Input Informatif

Abstract

Penelitian ini mengeksplorasi peringkasan teks abstraktif untuk berita berbahasa Indonesia dengan melakukan fine-tuning pada model PEGASUS menggunakan Bayesian Optimization dan input kontekstual yang diperkaya. Dataset berisi 286.277 pasangan dokumen–ringkasan yang diambil dari JPNN.com, lengkap dengan judul dan kata kunci yang digunakan untuk membentuk input informatif. Evaluasi menggunakan ROUGE dan BERTScore menunjukkan peningkatan substansial dari informative input: +16.75% (ROUGE-1), +27.25% (ROUGE-2), +18.58% (ROUGE-L & ROUGE-LSUM), dan +2.7% (BERTScore-F1) dibandingkan dengan input reguler. Analisis saliency menunjukkan bobot kalimat kontekstual yang konsisten tinggi. Penerapan hyperparameter tuning Bayesian melalui Optuna memberikan kenaikan marginal (+1.21% ROUGE-1, +2.1% ROUGE-2, +1.38% ROUGE-L & ROUGE-LSUM, +0.23% BERTScore) yang dipengaruhi oleh jumlah trial terbatas (12) dan ruang pencarian yang sempit. Temuan ini menegaskan efektivitas desain input kontekstual dan potensi hyperparameter tuning untuk peringkasan berbasis Transformer pada bahasa dengan sumber daya terbatas.

 

This research explores abstractive text summarization of Indonesian news by fine-tuning the PEGASUS model using Bayesian optimization and enriched contextual inputs. The dataset contains 286,277 document-summary pairs scraped from JPNN.com, including titles and keyphrases used to construct informative input. Each section is marked with special tokens such as <TITLE>, <KEYPHRASES>, and <ARTICLE>. Evaluation using ROUGE and BERTScore shows that informative input substantially improves performance: +16.75% (ROUGE-1), +27.25% (ROUGE-2), +18.58% (ROUGE-L and ROUGE-Lsum), and +2.7% (BERTScore-F1) compared with regular input. Saliency analysis also shows consistently high sentence weights for contextual input components. Additionally, Bayesian hyperparameter tuning via Optuna yields marginal gains (+1.21% ROUGE-1, +2.1% ROUGE-2, +1.38% ROUGE-L & ROUGE-Lsum, +0.23% BERTScore) due to a limited number of trials (12) and a constrained hyperparameter search space. These findings demonstrate the effectiveness of contextual input design and the potential of Bayesian tuning to improve Transformer-based summarization for low-resource languages.

References

[1] W. Zarman, “Information Overload: Clarifying the Problem,” Indonesian Journal of Informatics Education, vol. 5, no. 2, pp. 1–5, 2021, doi: 10.20961/ijie.v5i2.56922.

[2] A. M. F. Yousef, A. Alshamy, A. Tlili, and A. H. S. Metwally, “Demystifying the New Dilemma of Brain Rot in the Digital Era: A Review,” Brain Sciences, vol. 15, no. 3, p. 283, 2025, doi: 10.3390/brainsci15030283.

[3] A. Eko Raharjo, “Profiling News Consumption on Social Media,” Jurnal Komunikasi Profesional, vol. 5, no. 4, pp. 320–334, 2021, doi: 10.25139/jkp.v5i4.3794.

[4] D. Yadav, J. Desai, and A. K. Yadav, “Automatic Text Summarization Methods: A Comprehensive Review,” SN Computer Science, vol. 4, no. 1, p. 33, 2022, doi: 10.1007/s42979-022-01446-w.

[5] Supriyono, A. P. Wibawa, Suyono, and F. Kurniawan, “A Survey of Text Summarization: Techniques, Evaluation and Challenges,” Natural Language Processing Journal, vol. 7, p. 100070, 2024, doi: 10.1016/j.nlp.2024.100070.

[6] H. Lucky and D. Suhartono, “Investigation of Pre-Trained Bidirectional Encoder Representations from Transformers Checkpoints for Indonesian Abstractive Text Summarization,” Journal of Information and Communication Technology, vol. 21, no. 1, pp. 71–94, 2022, doi: 10.32890/jict2022.21.1.4.

[7] Z. Alami Merrouni, B. Frikh, and B. Ouhbi, “EXABSUM: A New Text Summarization Approach for Generating Extractive and Abstractive Summaries,” Journal of Big Data, vol. 10, no. 1, p. 163, 2023, doi: 10.1186/s40537-023-00836-y.

[8] G. Tucudean, M. Bucos, B. Dragulescu, and C. D. Caleanu, “Natural Language Processing with Transformers: A Review,” PeerJ Computer Science, vol. 10, p. e2222, 2024, doi: 10.7717/peerj-cs.2222.

[9] F. F. Kartamanah, A. R. Atmadja, and I. Budiman, “Analyzing PEGASUS Model Performance with ROUGE on Indonesian News Summarization,” Jurnal dan Penelitian Teknik Informatika, vol. 9, no. 1, pp. 31–42, 2025, doi: 10.33395/sinkron.v9i1.14278.

[10] R. Alsultan et al., “PEGASUS-XL with Saliency-Guided Scoring and Long-Input Encoding for Multi-Document Abstractive Summarization,” Scientific Reports, vol. 15, no. 1, p. 26529, 2025, doi: 10.1038/s41598-025-11062-2.

[11] A. Bahari and K. E. Dewi, “Peringkasan Teks Otomatis Abstraktif Menggunakan Transformer pada Teks Bahasa Indonesia,” KOMPUTA: Jurnal Ilmiah Komputer dan Informatika, vol. 13, no. 1, pp. 83–91, 2024, doi: 10.34010/komputa.v13i1.11197.

[12] F. Koto, T. Baldwin, and J. H. Lau, “LipKey: A Large-Scale News Dataset for Absent Keyphrases Generation and Abstractive Summarization,” in Proc. 29th Int. Conf. Computational Linguistics (COLING), Gyeongju, Republic of Korea, 2022, pp. 3427–3437, doi: 10.18653/v1/2022.coling-1.303.

[13] B. Bischl et al., “Hyperparameter Optimization: Foundations, Algorithms, Best Practices, and Open Challenges,” Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, vol. 13, no. 2, p. e1484, 2023, doi: 10.1002/widm.1484.

[14] A. R. Lubis et al., “Enhancing Text Summarization with a T5 Model and Bayesian Optimization,” Revue d’Intelligence Artificielle, vol. 37, no. 5, pp. 1213–1219, 2023, doi: 10.18280/ria.370513.

[15] C. M. Muia, A. M. Oirere, and R. N. Ndung’u, “A Comparative Study of Transformer-Based Models for Text Summarization of News Articles,” International Journal of Advanced Trends in Computer Science and Engineering, vol. 13, no. 2, pp. 37–43, 2024, doi: 10.30534/ijatcse/2024/011322024.

[16] A. H. Victoria and G. Maragatham, “Automatic Tuning of Hyperparameters Using Bayesian Optimization,” Evolving Systems, vol. 12, no. 1, pp. 217–223, 2021, doi: 10.1007/s12530-020-09345-2.

[17] A. Dalal et al., “Text Summarization for Pharmaceutical Sciences Using Hierarchical Clustering with a Weighted Evaluation Methodology,” Scientific Reports, vol. 14, no. 1, p. 20149, 2024, doi: 10.1038/s41598-024-70618-w.

[18] Junadhi, Agustin, L. Efrizoni, F. Okmayura, D. R. Habibie, and Muslim, “Improving Evaluation Metrics for Text Summarization: A Comparative Study and Proposal of a Novel Metric,” Journal of Applied Data Sciences, vol. 6, no. 2, pp. 885–896, 2025, doi: 10.47738/jads.v6i2.547.

[19] R. H. Astuti, M. Muljono, and S. Sutriawan, “Indonesian News Text Summarization Using MBART Algorithm,” Scientific Journal of Informatics, vol. 11, no. 1, pp. 155–164, 2024, doi: 10.15294/sji.v11i1.49224.

Downloads

Published

2026-04-28