Existing approaches to multilingual text detoxification are hampered by the scarcity of parallel multilingual datasets. In this work we introduce a pipeline for cross-lingual parallel detoxification data generation. We also introduce SynthDetoxM, a manually collected and synthetically generated multilingual parallel text detoxification dataset comprising 16,000 high-quality detoxification sentence pairs across German, French, Spanish and Russian. The data was sourced from different toxicity evaluation datasets and then rewritten with nine modern open-source LLMs in few-shot setting. Our experiments demonstrate that models trained on our data achieve superior performance to those trained on the human-annotated MultiParaDetox dataset even in data limited setting. Models, trained on SynthDetoxM outperform all evaluated LLMs in few-shot setting. We release our dataset and code to help further research in multilingual text detoxification.
This work introduces a pipeline for cross-lingual parallel detoxification data generation. The approach leverages large language models (LLMs) to create synthetic data, addressing the lack of parallel multilingual detoxification datasets. The methodology involves the following steps:
The table below presents the Joint (J) scores from the automatic evaluation of different multilingual text detoxification approaches. The models were evaluated on Spanish, German, and Russian using the test set from MultiParaDetox.
Spanish | German | Russian | |
---|---|---|---|
Human References | 0.709 | 0.733 | 0.732 |
Baselines | |||
Duplicate | 0.090 | 0.287 | 0.048 |
Delete | 0.319 | 0.362 | 0.255 |
Backtranslation | 0.275 | 0.233 | 0.223 |
Supervised Approaches | |||
MultiParaDetox | 0.344 | 0.446 | 0.472 |
SynthDetoxM (Batch) | 0.402 | 0.460 | 0.475 |
SynthDetoxM (Full) | 0.470 | 0.482 | 0.546 |
LLM-based Approaches | |||
Gemma 2 | 0.380 | 0.353 | 0.404 |
Mistral Nemo | 0.290 | 0.286 | 0.258 |
Command R | 0.344 | 0.328 | 0.402 |
Qwen 2.5 | 0.443 | 0.402 | 0.428 |
Llama 3.1 8B | 0.341 | 0.394 | 0.357 |
Aya Expanse 8B | 0.246 | 0.305 | 0.225 |
Aya Expanse 32B | 0.320 | 0.399 | 0.323 |
Mistral Small | 0.308 | 0.371 | 0.273 |
Key findings:
@misc{moskovskiy2025synthdetoxmmodernllmsfewshot,
title={SynthDetoxM: Modern LLMs are Few-Shot Parallel Detoxification Data Annotators},
author={Daniil Moskovskiy and Nikita Sushko and Sergey Pletenev and Elena Tutubalina and Alexander Panchenko},
year={2025},
eprint={2502.06394},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2502.06394},
}