Executive AI Dinner hosted by Kumo - Austin, April 8

Register here
6Binary Classification · Toxicity Risk

Toxic Behavior Prediction

Which players will receive reports for toxic behavior?

Book a demo and get a free trial of the full platform: research agent, fine-tune capabilities, and forward-deployed engineer support.

By submitting, you accept the Terms and Privacy Policy.

Loved by data scientists, ML engineers & CXOs at

Catalina Logo

A real-world example

Which players will receive reports for toxic behavior?

Toxic players drive away 15% of the non-toxic player base, costing games with 8M MAU roughly $22M annually in lost revenue. Reactive moderation (banning after reports pile up) means the damage is already done. By the time a player gets banned, they have ruined hundreds of matches. A proactive system that flags players before their behavior escalates would prevent the majority of community damage.

Quick answer

Proactive toxicity prediction requires connecting match outcomes, chat patterns, social connections, and behavioral trajectories in a graph model. Players who are on loss streaks, whose chat language is escalating, and who are connected to previously banned accounts are 7x more likely to receive reports in the next 48 hours. Graph ML catches these escalation patterns before the damage happens, unlike reactive report-based systems.

Approaches compared

4 ways to solve this problem

1. Keyword-based chat filtering

Flag messages containing banned words or phrases and auto-mute or warn players who trigger filters.

Best for

Catches obvious profanity and slurs in real time. Easy to implement and maintain with word lists.

Watch out for

Trivially bypassed with misspellings and coded language. Misses non-verbal toxicity like griefing, AFK behavior, and intentional feeding. Reactive by definition.

2. Report-based moderation

Rely on player reports to identify toxic behavior, then review and take action after reports accumulate.

Best for

Captures a broad range of toxicity types since players report what bothers them. Low implementation cost.

Watch out for

By the time reports pile up, the toxic player has ruined dozens of matches. Average time from first toxic act to ban is 2-3 weeks. The damage to non-toxic players is already done.

3. NLP sentiment analysis on chat

Train a text classifier on chat logs to detect negative sentiment, passive-aggressive language, and escalating hostility.

Best for

Catches sophisticated toxicity that keyword filters miss, including sarcasm and coded language patterns.

Watch out for

Only works on text-based toxicity. Many toxic behaviors are action-based (griefing, AFK, intentional losing) with no chat signal at all.

4. KumoRFM (relational graph ML)

Connect players, matches, chat logs, reports, and social connections into a graph. The GNN learns behavioral escalation patterns across all modalities: chat, gameplay actions, match outcomes, and social network effects.

Best for

Predicts toxicity 48 hours before reports happen. Captures non-verbal toxicity signals (match-quitting, performance sabotage) and social contagion from banned-player networks.

Watch out for

Requires comprehensive match and behavioral data. Less effective if your game only logs chat and not in-match actions.

Key metric: Games using proactive graph-based toxicity detection retain 15% more of their non-toxic community, saving $22M annually for an 8M MAU title.

Why relational data changes the answer

Toxicity is a behavioral trajectory, not a single event. The signals live across match outcomes (escalating loss streaks), chat logs (increasing flagged-word frequency), match behavior (rising early-quit rate), and social connections (proximity to previously banned accounts). A flat feature table with 'flagged_words = 5' misses the critical temporal context: those five flagged words happened after six consecutive losses, with each match showing worsening language.

Relational models read the full escalation arc. They connect the match history (losses), the chat evolution (escalating hostility), the behavioral pattern (early quits increasing), and the social graph (friends with prior bans). This connected view reveals that toxicity spreads through social networks: when a player's regular squad includes toxic players, that player's own behavior deteriorates within weeks. Detecting this contagion pattern requires graph structure. The result is intervention 48 hours before reports accumulate, when a cooling-off prompt or temporary matchmaking adjustment can prevent the toxicity outburst entirely.

Catching toxic players through reports is like catching drunk drivers through accident reports. By the time the report arrives, the damage is done. Predictive toxicity detection is like the patrol officer who spots the swerving before the crash: they see the loss streak, the escalating chat, the connection to known offenders, and they intervene before anyone gets hurt.

How KumoRFM solves this

Graph-learned player intelligence across your entire game ecosystem

Kumo models the social contagion of toxic behavior across the player network. It connects players, matches, chat patterns, reports, and social connections to learn that players who recently lost 5+ matches, whose teammates reported others in those matches, and who are connected to previously banned accounts are 7x more likely to receive reports in the next 48 hours. The temporal graph captures escalation patterns: increasingly negative chat, shorter sessions, and match-quitting streaks that precede toxicity outbursts.

From data to predictions

See the full pipeline in action

Connect your tables, write a PQL query, and get predictions with built-in explainability — all in minutes, not months.

1

Your data

The relational tables Kumo learns from

PLAYERS

player_idaccount_age_dayshonor_levelprior_bans
PLR50134030
PLR5024511
PLR50372050

MATCHES

match_idplayer_idresultduration_minquit_early
M501PLR502Loss8Y
M502PLR502Loss12N
M503PLR501Win25N

CHAT_LOGS

chat_idmatch_idplayer_idmessage_countflagged_words
CL01M501PLR502425
CL02M502PLR502283
CL03M503PLR501150

REPORTS

report_idreported_playerreporter_idreasontimestamp
R001PLR502PLR501Verbal abuse2025-03-01
R002PLR502PLR503AFK/griefing2025-03-02

SOCIAL_CONNECTIONS

edge_idplayer_aplayer_btype
SC501PLR501PLR503Friend
SC502PLR502PLR501Recent opponent
2

Write your PQL query

Describe what to predict in 2–3 lines — Kumo handles the rest

PQL
PREDICT COUNT(REPORTS.*, 0, 48, hours) > 0
FOR EACH PLAYERS.PLAYER_ID
WHERE PLAYERS.HONOR_LEVEL < 4
3

Prediction output

Every entity gets a score, updated continuously

PLAYER_IDHONOR_LEVELACCOUNT_AGETOXIC_48H_PROB
PLR5013340d0.05
PLR502145d0.88
PLR5035720d0.02
4

Understand why

Every prediction includes feature attributions — no black boxes

Player PLR502 -- Honor 1, 45-day account

Predicted: 88% toxic behavior probability in next 48 hours

Top contributing features

Loss streak (last 24h)

6 consecutive losses

29% attribution

Flagged chat words (last 48h)

8 instances

25% attribution

Early quit rate (last 7d)

40% of matches

19% attribution

Connection to banned players

2 banned friends

15% attribution

Prior ban history

1 previous ban

12% attribution

Feature attributions are computed automatically for every prediction. No separate tooling required. Learn more about Kumo explainability

Frequently asked questions

Common questions about toxic behavior prediction

How do you predict toxic behavior in online games?

Connect match outcomes, chat patterns, behavioral signals (early quits, performance drops), and social connections into a relational model. The strongest predictors are behavioral trajectories: a player on a loss streak whose chat language is escalating and who is connected to previously banned accounts is 7x more likely to receive reports. Graph ML detects these patterns 48 hours before reactive systems.

What data is needed for toxicity prediction?

Four data sources matter: match history with outcomes, chat logs with flagged-word counts, in-match behavioral data (early quits, AFK events, performance anomalies), and social connections. The social graph is critical because toxicity is contagious. Players connected to banned accounts are significantly more likely to exhibit toxic behavior themselves.

Can you predict griefing and non-verbal toxicity?

Yes, and this is where graph ML has the biggest advantage over text-based approaches. Non-verbal toxicity (griefing, intentional feeding, AFK) shows up in match performance data and behavioral patterns, not in chat. Relational models connect match outcomes, in-game actions, and historical behavior to detect griefing patterns that chat analysis completely misses.

How much revenue does toxic behavior cost games?

Toxic players drive away 15% of the non-toxic player base. For a game with 8M MAU, that translates to roughly $22M in annual revenue loss. The cost is not just from toxic players leaving (many do not leave), but from the non-toxic players who quit because of repeated bad experiences.

Is proactive moderation better than reactive banning?

Proactive intervention prevents damage rather than punishing it after the fact. A toxic player who gets a cooling-off prompt or temporary matchmaking adjustment before an outburst causes zero community damage. A player who gets banned after two weeks of toxic behavior has already ruined hundreds of matches. Proactive systems retain 15% more of the non-toxic community.

Bottom line: A game with 8M MAU that proactively intervenes on toxic players before reports accumulate retains 15% more of its non-toxic community, saving $22M in annual revenue. Kumo detects escalation patterns across match outcomes, chat sentiment, social connections, and behavioral trajectories that reactive report-based systems catch too late.

Topics covered

toxic behavior predictiongame moderation AIplayer toxicity modelcommunity safety MLgriefing predictiongraph neural network moderationKumoRFM toxic behaviorplayer behavior analyticsproactive game moderation

One Platform. One Model. Infinite Predictions.

KumoRFM

Relational Foundation Model

Turn structured relational data into predictions in seconds. KumoRFM delivers zero-shot predictions that rival months of traditional data science. No training, feature engineering, or infrastructure required. Just connect your data and start predicting.

For critical use cases, fine-tune KumoRFM on your data using the Kumo platform and Research Agent for 30%+ higher accuracy than traditional models.

Book a demo and get a free trial of the full platform: research agent, fine-tune capabilities, and forward-deployed engineer support.