Existing approaches to multilingual text detoxification are hampered by the scarcity of parallel multilingual datasets. In this work we introduce a pipeline for cross-lingual parallel detoxification data generation. We also introduce SynthDetoxM, a manually collected and synthetically generated multilingual parallel text detoxification dataset comprising 16,000 high-quality detoxification sentence pairs across German, French, Spanish and Russian. The data was sourced from different toxicity evaluation datasets and then rewritten with nine modern open-source LLMs in few-shot setting. Our experiments demonstrate that models trained on our data achieve superior performance to those trained on the human-annotated MultiParaDetox dataset even in data limited setting. Models, trained on SynthDetoxM outperform all evaluated LLMs in few-shot setting. We release our dataset and code to help further research in multilingual text detoxification.
This work introduces a pipeline for cross-lingual parallel detoxification data generation. The approach leverages large language models (LLMs) to create synthetic data, addressing the lack of parallel multilingual detoxification datasets. The methodology involves the following steps:
The table below presents the Joint (J) scores from the automatic evaluation of different multilingual text detoxification approaches. The models were evaluated on Spanish, German, and Russian using the test set from MultiParaDetox.
Spanish | German | Russian | |
---|---|---|---|
Human References | 0.709 | 0.733 | 0.732 |
Baselines | |||
Duplicate | 0.090 | 0.287 | 0.048 |
Delete | 0.319 | 0.362 | 0.255 |
Backtranslation | 0.275 | 0.233 | 0.223 |
Supervised Approaches | |||
MultiParaDetox | 0.344 | 0.446 | 0.472 |
SynthDetoxM (Batch) | 0.402 | 0.460 | 0.475 |
SynthDetoxM (Full) | 0.470 | 0.482 | 0.546 |
LLM-based Approaches | |||
Gemma 2 | 0.380 | 0.353 | 0.404 |
Mistral Nemo | 0.290 | 0.286 | 0.258 |
Command R | 0.344 | 0.328 | 0.402 |
Qwen 2.5 | 0.443 | 0.402 | 0.428 |
Llama 3.1 8B | 0.341 | 0.394 | 0.357 |
Aya Expanse 8B | 0.246 | 0.305 | 0.225 |
Aya Expanse 32B | 0.320 | 0.399 | 0.323 |
Mistral Small | 0.308 | 0.371 | 0.273 |
Key findings:
@inproceedings{moskovskiy-etal-2025-synthdetoxm,
title = "{S}ynth{D}etox{M}: {M}odern {LLM}s are Few-Shot Parallel Detoxification Data Annotators",
author = "Moskovskiy, Daniil and
Sushko, Nikita and
Pletenev, Sergey and
Tutubalina, Elena and
Panchenko, Alexander",
editor = "Chiruzzo, Luis and
Ritter, Alan and
Wang, Lu",
booktitle = "Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)",
month = apr,
year = "2025",
address = "Albuquerque, New Mexico",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2025.naacl-long.294/",
pages = "5714--5733",
ISBN = "979-8-89176-189-6",
abstract = "Existing approaches to multilingual text detoxification are hampered by the scarcity of parallel multilingual datasets. In this work, we introduce a pipeline for the generation of multilingual parallel detoxification data. We also introduce SynthDetoxM, a manually collected and synthetically generated multilingual parallel text detoxification dataset comprising 16,000 high-quality detoxification sentence pairs across German, French, Spanish and Russian. The data was sourced from different toxicity evaluation datasets and then rewritten with nine modern open-source LLMs in few-shot setting. Our experiments demonstrate that models trained on the produced synthetic datasets have superior performance to those trained on the human-annotated MultiParaDetox dataset even in data limited setting. Models trained on SynthDetoxM outperform all evaluated LLMs in few-shot setting. We release our dataset and code to help further research in multilingual text detoxification."
}