Adapting Large Language Models for Content Moderation: Pitfalls in Data
Engineering and Supervised Fine-tuning
- URL: http://arxiv.org/abs/2310.03400v2
- Date: Thu, 7 Mar 2024 12:04:54 GMT
- Title: Adapting Large Language Models for Content Moderation: Pitfalls in Data
Engineering and Supervised Fine-tuning
- Authors: Huan Ma, Changqing Zhang, Huazhu Fu, Peilin Zhao, Bingzhe Wu
- Abstract summary: Large Language Models (LLMs) have become a feasible solution for handling tasks in various domains.
In this paper, we introduce how to fine-tune a LLM model that can be privately deployed for content moderation.
- Score: 79.53130089003986
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Nowadays, billions of people engage in communication and express their
opinions on the internet daily. Unfortunately, not all of these expressions are
friendly or compliant, making content moderation an indispensable task. A
common approach is to use a discriminative model to classify the content, but
this method often requires strict data engineering, otherwise it will face
unacceptable overfitting. With the successful development of Large Language
Models (LLMs) in recent years, LLM-based methods have become a feasible
solution for handling tasks in various domains. Thanks to the knowledge of the
foundation models, we can develop more robust privately deployed models with
limited data via fine-tuning these foundation models. Moreover, as a generative
model, it can provide detailed analysis of the review process, enhancing
interpretability. In this paper, we introduce how to fine-tune a LLM model that
can be privately deployed for content moderation. Specifically, we discuss the
differences between discriminative and generative models using content
moderation as an example. Additionally, we reveal that incorporating reasoning
processes during the fine-tuning of LLMs can effectively alleviate overfitting,
even if the model is not allowed to directly output reasoning processes during
deployment. We present a complete process, from data collection and
construction to model training and overfitting elimination, for fine-tuning
LLMs in vertical domain deployments. We report the entire research process and
the key findings in this paper, hoping to provide valuable experience for
researchers who are fine-tuning privately deployed models in their
domain-specific research.
Related papers
- What matters when building vision-language models? [52.8539131958858]
We develop Idefics2, an efficient foundational vision-language model with 8 billion parameters.
Idefics2 achieves state-of-the-art performance within its size category across various multimodal benchmarks.
We release the model (base, instructed, and chat) along with the datasets created for its training.
arXiv Detail & Related papers (2024-05-03T17:00:00Z) - Entity Matching using Large Language Models [4.189643331553922]
This paper investigates using generative large language models (LLMs) as a less task-specific training data-dependent alternative to PLM-based matchers.
We show that GPT4 can generate structured explanations for matching decisions.
arXiv Detail & Related papers (2023-10-17T13:12:32Z) - In-Context Unlearning: Language Models as Few Shot Unlearners [27.962361828354716]
We propose a new class of unlearning methods for Large Language Models (LLMs)
This method unlearns instances from the model by simply providing specific kinds of inputs in context, without the need to update model parameters.
Our experimental results demonstrate that in-context unlearning performs on par with, or in some cases outperforms other state-of-the-art methods that require access to model parameters.
arXiv Detail & Related papers (2023-10-11T15:19:31Z) - Matching Pairs: Attributing Fine-Tuned Models to their Pre-Trained Large
Language Models [11.57282859281814]
We consider different knowledge levels and attribution strategies, and find that we can correctly trace back 8 out of the 10 fine tuned models with our best method.
arXiv Detail & Related papers (2023-06-15T17:42:48Z) - LLM2Loss: Leveraging Language Models for Explainable Model Diagnostics [5.33024001730262]
We propose an approach that can provide semantic insights into a model's patterns of failures and biases.
We show that an ensemble of such lightweight models can be used to generate insights on the performance of the black-box model.
arXiv Detail & Related papers (2023-05-04T23:54:37Z) - Scaling Vision-Language Models with Sparse Mixture of Experts [128.0882767889029]
We show that mixture-of-experts (MoE) techniques can achieve state-of-the-art performance on a range of benchmarks over dense models of equivalent computational cost.
Our research offers valuable insights into stabilizing the training of MoE models, understanding the impact of MoE on model interpretability, and balancing the trade-offs between compute performance when scaling vision-language models.
arXiv Detail & Related papers (2023-03-13T16:00:31Z) - Dataless Knowledge Fusion by Merging Weights of Language Models [51.8162883997512]
Fine-tuning pre-trained language models has become the prevalent paradigm for building downstream NLP models.
This creates a barrier to fusing knowledge across individual models to yield a better single model.
We propose a dataless knowledge fusion method that merges models in their parameter space.
arXiv Detail & Related papers (2022-12-19T20:46:43Z) - Large Language Models with Controllable Working Memory [64.71038763708161]
Large language models (LLMs) have led to a series of breakthroughs in natural language processing (NLP)
What further sets these models apart is the massive amounts of world knowledge they internalize during pretraining.
How the model's world knowledge interacts with the factual information presented in the context remains under explored.
arXiv Detail & Related papers (2022-11-09T18:58:29Z) - Model Reprogramming: Resource-Efficient Cross-Domain Machine Learning [65.268245109828]
In data-rich domains such as vision, language, and speech, deep learning prevails to deliver high-performance task-specific models.
Deep learning in resource-limited domains still faces multiple challenges including (i) limited data, (ii) constrained model development cost, and (iii) lack of adequate pre-trained models for effective finetuning.
Model reprogramming enables resource-efficient cross-domain machine learning by repurposing a well-developed pre-trained model from a source domain to solve tasks in a target domain without model finetuning.
arXiv Detail & Related papers (2022-02-22T02:33:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.