Publication details

IMGTB: A Framework for Machine-Generated Text Detection Benchmarking

Authors

SPIEGEL Michal DOMINIK Macko

Year of publication 2024
Type Article in Proceedings
Conference Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 3: System Demonstrations)
MU Faculty or unit

Faculty of Informatics

Citation
web https://aclanthology.org/2024.acl-demos.17/
Doi http://dx.doi.org/10.18653/v1/2024.acl-demos.17
Keywords machine-generated text detection; natural language processing; large language models; evaluation; comparison; framework
Description In the era of large language models generating high quality texts, it is a necessity to develop methods for detection of machine-generated text to avoid their harmful use or simply for annotation purposes. It is, however, also important to properly evaluate and compare such developed methods. Recently, a few benchmarks have been proposed for this purpose; however, integration of newest detection methods is rather challenging, since new methods appear each month and provide slightly different evaluation pipelines.In this paper, we present the IMGTB framework, which simplifies the benchmarking of machine-generated text detection methods by easy integration of custom (new) methods and evaluation datasets. In comparison to existing frameworks, it enables to objectively compare statistical metric-based zero-shot detectors with classification-based detectors and with differently fine-tuned detectors. Its configurability and flexibility makes research and development of new detection methods easier, especially their comparison to the existing state-of-the-art detectors. The default set of analyses, metrics and visualizations offered by the tool follows the established practices of machine-generated text detection benchmarking found in state-of-the-art literature.
Related projects:

You are running an old browser version. We recommend updating your browser to its latest version.

More info