Adapting Large Language Models for Content Moderation: Pitfalls in Data
Engineering and Supervised Fine-tuning
- URL: http://arxiv.org/abs/2310.03400v2
- Date: Thu, 7 Mar 2024 12:04:54 GMT
- Title: Adapting Large Language Models for Content Moderation: Pitfalls in Data
Engineering and Supervised Fine-tuning
- Authors: Huan Ma, Changqing Zhang, Huazhu Fu, Peilin Zhao, Bingzhe Wu
- Abstract summary: Large Language Models (LLMs) have become a feasible solution for handling tasks in various domains.
In this paper, we introduce how to fine-tune a LLM model that can be privately deployed for content moderation.
- Score: 79.53130089003986
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Nowadays, billions of people engage in communication and express their
opinions on the internet daily. Unfortunately, not all of these expressions are
friendly or compliant, making content moderation an indispensable task. A
common approach is to use a discriminative model to classify the content, but
this method often requires strict data engineering, otherwise it will face
unacceptable overfitting. With the successful development of Large Language
Models (LLMs) in recent years, LLM-based methods have become a feasible
solution for handling tasks in various domains. Thanks to the knowledge of the
foundation models, we can develop more robust privately deployed models with
limited data via fine-tuning these foundation models. Moreover, as a generative
model, it can provide detailed analysis of the review process, enhancing
interpretability. In this paper, we introduce how to fine-tune a LLM model that
can be privately deployed for content moderation. Specifically, we discuss the
differences between discriminative and generative models using content
moderation as an example. Additionally, we reveal that incorporating reasoning
processes during the fine-tuning of LLMs can effectively alleviate overfitting,
even if the model is not allowed to directly output reasoning processes during
deployment. We present a complete process, from data collection and
construction to model training and overfitting elimination, for fine-tuning
LLMs in vertical domain deployments. We report the entire research process and
the key findings in this paper, hoping to provide valuable experience for
researchers who are fine-tuning privately deployed models in their
domain-specific research.
Related papers
- Forewarned is Forearmed: Leveraging LLMs for Data Synthesis through Failure-Inducing Exploration [90.41908331897639]
Large language models (LLMs) have significantly benefited from training on diverse, high-quality task-specific data.
We present a novel approach, ReverseGen, designed to automatically generate effective training samples.
arXiv Detail & Related papers (2024-10-22T06:43:28Z) - Cross-Domain Content Generation with Domain-Specific Small Language Models [3.2772349789781616]
This study explores methods to enable a small language model to produce coherent and relevant outputs for two different domains.
We find that utilizing custom tokenizers tailored to each dataset significantly enhances generation quality.
Our findings demonstrate that knowledge expansion with frozen layers is an effective method for small language models to generate domain-specific content.
arXiv Detail & Related papers (2024-09-19T21:45:13Z) - Model Attribution in LLM-Generated Disinformation: A Domain Generalization Approach with Supervised Contrastive Learning [26.02988481241285]
Modern large language models (LLMs) produce disinformation with human-like quality.
diversity in prompting methods used to generate disinformation complicates accurate source attribution.
We introduce the concept of model attribution as a domain generalization problem.
arXiv Detail & Related papers (2024-07-31T00:56:09Z) - What matters when building vision-language models? [52.8539131958858]
We develop Idefics2, an efficient foundational vision-language model with 8 billion parameters.
Idefics2 achieves state-of-the-art performance within its size category across various multimodal benchmarks.
We release the model (base, instructed, and chat) along with the datasets created for its training.
arXiv Detail & Related papers (2024-05-03T17:00:00Z) - In-Context Unlearning: Language Models as Few Shot Unlearners [27.962361828354716]
We propose a new class of unlearning methods for Large Language Models (LLMs)
This method unlearns instances from the model by simply providing specific kinds of inputs in context, without the need to update model parameters.
Our experimental results demonstrate that in-context unlearning performs on par with, or in some cases outperforms other state-of-the-art methods that require access to model parameters.
arXiv Detail & Related papers (2023-10-11T15:19:31Z) - Matching Pairs: Attributing Fine-Tuned Models to their Pre-Trained Large
Language Models [11.57282859281814]
We consider different knowledge levels and attribution strategies, and find that we can correctly trace back 8 out of the 10 fine tuned models with our best method.
arXiv Detail & Related papers (2023-06-15T17:42:48Z) - Dataless Knowledge Fusion by Merging Weights of Language Models [51.8162883997512]
Fine-tuning pre-trained language models has become the prevalent paradigm for building downstream NLP models.
This creates a barrier to fusing knowledge across individual models to yield a better single model.
We propose a dataless knowledge fusion method that merges models in their parameter space.
arXiv Detail & Related papers (2022-12-19T20:46:43Z) - Large Language Models with Controllable Working Memory [64.71038763708161]
Large language models (LLMs) have led to a series of breakthroughs in natural language processing (NLP)
What further sets these models apart is the massive amounts of world knowledge they internalize during pretraining.
How the model's world knowledge interacts with the factual information presented in the context remains under explored.
arXiv Detail & Related papers (2022-11-09T18:58:29Z) - Model Reprogramming: Resource-Efficient Cross-Domain Machine Learning [65.268245109828]
In data-rich domains such as vision, language, and speech, deep learning prevails to deliver high-performance task-specific models.
Deep learning in resource-limited domains still faces multiple challenges including (i) limited data, (ii) constrained model development cost, and (iii) lack of adequate pre-trained models for effective finetuning.
Model reprogramming enables resource-efficient cross-domain machine learning by repurposing a well-developed pre-trained model from a source domain to solve tasks in a target domain without model finetuning.
arXiv Detail & Related papers (2022-02-22T02:33:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.