Google study finds LLMs are embedded at every stage of abuse detection
Online platforms are running large language models at every stage of LLM content moderation, from generating training data to auditing their own systems for bias. Researchers at Google mapped how this is happening across what the authors call the Abuse Detection Lifecycle, a four-stage framework covering labeling, detection, review and appeals, and auditing.

Earlier moderation systems, built on models like BERT and RoBERTa fine-tuned on static hate-speech datasets, could identify explicit slurs with reasonable accuracy. They struggled with sarcasm, coded language, and culturally specific abuse. LLMs address some of those gaps through contextual reasoning, but they introduce new operational and governance problems at each stage they enter.
Labeling: synthetic data at scale, with bias attached
Generating labeled training data has long been a bottleneck for LLM content moderation. Human annotators are slow, expensive, and inconsistent, particularly on implicit or context-dependent content. LLMs are used to produce synthetic labels at volumes that manual annotation cannot match.
One study cited in the survey used three LLMs as independent annotators, aggregated their labels through majority voting, and produced over 48,000 synthetic media-bias labels. Classifiers trained on that synthetic output performed comparably to models trained on expert-labeled data. A retrieval-augmented approach in financial text classification retrieved only 2.2% of available examples to match GPT-4 few-shot accuracy, significantly cutting inference costs.
Instruction-tuned models tend to under-predict abuse labels because of imbalanced training corpora. Models aligned through reinforcement learning from human feedback tend to over-predict, flagging content out of excessive caution. Different LLMs also carry distinct political or ideological leanings that surface in the labels they generate. Validation against human annotations remains necessary.
Detection: specialized models are outperforming generalists
At the detection stage, the survey distinguishes between general-purpose LLMs used as zero-shot classifiers and smaller models fine-tuned specifically for safety tasks. GPT-4 achieves F1 scores above 0.75 on standard toxicity benchmarks in zero-shot settings, which matches or exceeds non-expert human annotators. Few-shot prompting, providing three to five labeled examples in the prompt, closes much of the remaining gap with specialist models.
Meta’s Llama Guard family represents the fine-tuned specialist approach. It performs input-output safeguarding on both user prompts and model responses, and supports zero-shot policy adaptation, meaning administrators can pass a new safety policy directly in the prompt without retraining the model.
A persistent challenge in LLM content moderation is over-refusal. RLHF-aligned models used as classifiers tend to flag benign content that resembles unsafe content in surface features. Studies evaluating Llama-2 and GPT-4 found high false positive rates on prompts that merely touched sensitive topics without crossing policy lines.
Implicit abuse, including sarcasm and coded hate speech, remains difficult. Contrastive learning techniques applied to LLM embeddings have shown strong results on implicit hate detection, sometimes outperforming larger generative models in accuracy and computational cost. Coordinated inauthentic behavior requires a different approach: graph neural networks enhanced with LLM-generated semantic embeddings can identify networks of accounts that share both structural posting patterns and linguistically similar content. The FraudSquad framework, built for detecting LLM-generated spam reviews, reported a 44% precision improvement over prior baselines using this dual-view method.
Review and auditing: LLMs supporting and checking human decisions
At the review stage, LLM content moderation tools are used to generate policy-grounded explanations for moderation decisions, summarize evidence for human reviewers, and assist with the appeals process by translating policy violations into plain language. The survey cites research showing that this kind of reason-giving improves consistency and gives users a better basis for contesting decisions.
A known problem at this stage is that chain-of-thought explanations can be unfaithful. Models sometimes generate rationales that sound logically sound to reviewers but do not reflect the model’s actual decision process. Research has also found that the fluency of LLM-generated text leads human moderators to rate incorrect explanations as acceptable at higher rates.
At the auditing stage, LLMs are used to stress-test detection systems with adversarial prompts, identify demographic disparities in enforcement, and monitor for concept drift over time. One study analyzed toxicity elicitation across over 1,200 identity groups and found systematic disparities in how safety filters treated marginalized populations. Temporal instability is also documented: toxicity prediction scores from the same API have varied significantly across evaluation periods.
The production gap and the safety paradox
Running a large reasoning model on every piece of content on a high-volume platform is computationally impractical. The survey estimates that inference costs for frontier models are orders of magnitude higher per query than distilled baselines. Platforms are working around this by routing easy cases to smaller, faster models and reserving LLMs for ambiguous content. Research on the SafeRoute framework found that a significant portion of user traffic does not require the reasoning depth of a multi-billion parameter model.
The broader structural tension the survey identifies is that LLM content moderation improves defensive capabilities and lowers the barrier for attackers simultaneously. Generating unique, personalized abusive content at scale is now accessible to low-sophistication actors. Detection systems must now account for machine-generated disinformation and fake reviews alongside human-generated abuse.
The survey concludes that future architectures will need to combine smaller specialized guardrails with retrieval-augmented policy references, continuous red-teaming through autonomous adversarial agents, and sustained human oversight at multiple stages of the pipeline.

Guide: Breach and Attack Simulation & Automated Penetration Testing