Automatic Construction of a Korean Toxic Instruction Dataset for Ethical Tuning of Large Language Models
This study introduces KoTox, a Korean Toxic instruction dataset to improve the ethical robustness of Large Language Models (LLMs). The dataset comprises 39K unethical instruction-output pairs and focuses on three…
Continue reading