Automatic Construction of a Korean Toxic Instruction Dataset for Ethical Tuning of Large Language Models
The paper presents KoTox, a dataset of 39K unethical instruction-output pairs, aimed at refining the training of Large Language Models (LLMs) and improving their ethical awareness. The dataset is designed…
Continue reading