Exploiting Machine Learning Model Ensemble for AI-Generated Texts Detection

Authors

  • Chucheng Zhou

DOI:

https://doi.org/10.62051/dvxnw210

Keywords:

Large language models; AI-generated texts; ensemble model.

Abstract

In recent years, the development of Large Language Models (LLMs) has reached a state of maturity, resulting in texts that are becoming increasingly indistinguishable from those written by humans. This progress has sparked a growing need for precise and efficient methods to detect Artificial Intelligence (AI)-generated texts, as the blend of human and machine authorship becomes more seamless. This study utilizes data from student-written papers and articles generated by various LLMs to develop a machine learning model capable of accurately distinguishing whether an article was written by a student or an LLM. Four different classification models (MultinomialNB, SGDClassifier, LGBMClassifier, and CatBoostClassifier) and their ensemble models with weighted combinations were chosen to detect AI-generated texts. To evaluate the effectiveness of these models, the Area Under Curve (AUC) score metric was employed. The results indicate that the CatBoostClassifier model performs the best in mitigating overfitting, while the ensemble model demonstrates the optimal predictive performance. This discovery holds significant importance for enhancing the accuracy of detecting AI-generated articles.

Downloads

Download data is not yet available.

References

Chang, Yupeng, Xu Wang, Jindong Wang, Yuan Wu, Linyi Yang, Kaijie Zhu, Hao Chen et al. A survey on evaluation of large language models. ACM Transactions on Intelligent Systems and Technology, 2023: 1-43.

Sadasivan, Vinu Sankar, Aounon Kumar, Sriram Balasubramanian, Wenxiao Wang, and Soheil Feizi. Can AI-generated text be reliably detected?. ArXiv Preprint, 2023: 2303.11156.

Shah, Aditya, Prateek Ranka, Urmi Dedhia, Shruti Prasad, Siddhi Muni, and Kiran Bhowmick. Detecting and Unmasking AI-Generated Texts through Explainable Artificial Intelligence using Stylistic Features. International Journal of Advanced Computer Science and Applications, 2023, 14(10): 1043-1053.

Kibriya, Ashraf M., Eibe Frank, Bernhard Pfahringer, and Geoffrey Holmes. Multinomial naive bayes for text categorization revisited. In AI 2004: Advances in Artificial Intelligence: 17th Australian Joint Conference on Artificial Intelligence, 2005, 17: 488-499.

Kabir, Fasihul, Sabbir Siddique, Mohammed Rokibul Alam Kotwal, and Mohammad Nurul Huda. Bangla text document categorization using stochastic gradient descent (sgd) classifier. In 2015 international conference on cognitive computing and information processing, 2015: 1-4.

Si, Si, Huan Zhang, S. Sathiya Keerthi, Dhruv Mahajan, Inderjit S. Dhillon, and Cho-Jui Hsieh. Gradient boosted decision trees for high dimensional sparse output. In International conference on machine learning, 2017: 3182-3190.

Ke, Guolin, Qi Meng, Thomas Finley, Taifeng Wang, Wei Chen, Weidong Ma, Qiwei Ye, and Tie-Yan Liu. Lightgbm: A highly efficient gradient boosting decision tree. Advances in neural information processing systems, 2017, 30: 1-9.

Hancock, John T., and Taghi M. Khoshgoftaar. CatBoost for big data: an interdisciplinary review. Journal of big data, 2020, 7(1): 94.

DAIGT V2 Training Dataset. URL: https://www.kaggle.com/datasets/thedrcat/daigt-v2-train-dataset. Last Accessed 2024/03/13.

Christian, Hans, Mikhael Pramodana Agus, and Derwin Suhartono. Single document automatic text summarization using term frequency-inverse document frequency (TF-IDF). ComTech: Computer, Mathematics and Engineering Applications, 2016, 7(4): 285-294.

Downloads

Published

12-08-2024

How to Cite

Zhou, C. (2024) “Exploiting Machine Learning Model Ensemble for AI-Generated Texts Detection”, Transactions on Computer Science and Intelligent Systems Research, 5, pp. 196–203. doi:10.62051/dvxnw210.