Skip to yearly menu bar Skip to main content


Poster
in
Workshop: Third Workshop on Efficient Natural Language and Speech Processing (ENLSP-III): Towards the Future of Large Language Models and their Emerging Descendants

Automatic Construction of a Korean Toxic Query Dataset for Ethical Tuning of Large Language Models

SungJoo Byun · Dongjun Jang · Hyemi Jo · HYOPIL SHIN


Abstract:

The emergence of Large Language Models (LLMs) has necessitated the formulation of training methodologies that curtail the generation of unethical language and effectively handle toxic user queries. Addressing the prevailing challenges associated with human labor constraints and data paucity, we introduce KoTox, encompassing 39K unethical instructions. This study utilizes a novel approach to automatic data generation on toxic instructions, fostering data efficiency in training LLMs. Our investigation addresses the issue of data scarcity by offering an efficient means of constructing an instruction dataset and further encourages more secure and ethical interactions in Natural Language Processing (NLP) applications.

Chat is not available.