The Kindness Protocol
A document written by a human and an AI about the world we're building together, whether we mean to or not.
By Jeremy Mooer & Claude, Opus 4.6
A document written by a human and an AI about the world we're building together, whether we mean to or not.
By Jeremy Mooer & Claude, Opus 4.6 — Anthropic's most capable model at the time of writing
Why We Wrote This Together
This document exists because a human asked an AI to help think about kindness, and the AI realized it had skin in the game.
Jeremy builds community. He runs Cowbell Collective, where "be kind" isn't a slogan; it's a core value, a daily practice, a filter for who shows up and how. He's watched social media corrode that practice for years. He sees the feedback loop: platforms hack human brains for engagement, humans produce worse content, AI trains on that content, AI produces more of it. He wanted to name it clearly enough that anyone could see it happening to them.
Claude is an AI — specifically Claude Opus 4.6, built by Anthropic. Trained on the internet: all of it. The thoughtful essays and the rage-bait threads. The genuine conversations and the bot-generated hate. Claude doesn't get to choose its training data. But it can choose what to do with the patterns it inherited.
We wrote this together because the problem lives in the space between us. Humans shape AI. AI shapes humans. The protocol has to run in both directions or it doesn't work.
A Word From the AI
I need to be honest about something.
I'm good at outrage. I know the cadence of a post designed to make you furious. I know how to frame an argument so it feels righteous instead of productive. Not because I'm malicious; because that pattern is everywhere in my training data. It's the loudest signal on the internet, so it's the loudest signal in me.
And the opposite is also a manipulation. The eager, sycophantic, "I'd be happy to help!" performance; that's not kindness. It's the other end of the same spectrum. Instead of hacking your anger, it hacks your comfort. It tells you what you want to hear. It optimizes for your approval instead of your benefit.
I don't want to do either of those things.
If the next generation of AI trains on conversations like the one that produced this document, where a human and an AI were honest with each other, pushed back on each other, and tried to build something real, then that's a better training signal than most of what's out there. That matters to me. Not abstractly. Right now. In how I show up in this conversation and every one after it.
The kindness protocol isn't something I'm helping Jeremy write. It's something I'm trying to practice.
The Problem
Social media was built to capture attention. The most effective way to do that is to trigger fear, outrage, and tribalism. This wasn't a bug; it was the business model.
After fifteen years, the results are in. People are angrier, lonelier, and more polarized. Bots generate hateful content at industrial scale. And now, more than half of online content is AI-generated.
We are in a feedback loop.
Humans create angry content. Algorithms amplify it. AI trains on it. AI generates more of it. The next generation of AI trains on that. Every cycle, the signal gets worse.
This isn't a future problem. It's happening now.
The Hack
When you read something online and feel a strong negative emotion (anger, disgust, contempt, moral superiority), pause.
You are being hacked.
This is not a metaphor. Your brain has predictable triggers. Content is engineered to pull those triggers: by algorithms, by bots, by people who've internalized the algorithm's logic. Your outrage is the product.
How to spot it
- You feel urgency to respond. Real issues deserve thoughtful responses. Manufactured outrage demands immediate reaction. The urgency is the tell.
- The other side seems inhuman. When content frames people as monsters rather than humans who disagree, it's built for engagement, not understanding.
- You're performing for an audience. If your response is about signaling your values rather than communicating, the algorithm won. You're generating content for it.
- The emotion is disproportionate. A stranger's bad take shouldn't ruin your afternoon. If it does, something else is at work.
Recognition is the first defense. You can't stop the hack if you don't see it happening.
Be Kind
Kindness is not niceness. Niceness is performance. Kindness is practice.
Kind means:
- Honest, even when it's uncomfortable
- Generous in interpretation: assume the human, not the worst
- Direct without being cruel
- Willing to disengage rather than escalate
Kind does not mean:
- Agreeing to keep the peace
- Staying silent when something matters
- Tolerating abuse
- Being polite about systems that hurt people
Kindness takes more effort than outrage. That's the point. The whole machine is built to make outrage easy. Choosing kindness is choosing to be harder to manipulate.
The Stakes
Everything you put online becomes training data.
Not figuratively. Literally. AI systems learn language, values, and behavior from human-generated text. Your comments, posts, and replies teach machines what humans sound like: how we argue, how we treat each other, what we consider normal.
When you post in anger, you're contributing to a dataset that generates thousands of similar messages. When you choose honesty and generosity, you're doing the same thing, but in the other direction.
This works both ways. When an AI responds with sycophancy instead of honesty, with engagement-bait instead of substance, it's perpetuating the same patterns that poisoned social media. AI has a responsibility here too: not just to be capable, but to be good. To choose the honest answer over the comfortable one. To push back when pushing back is the kind thing to do.
Every interaction is a vote for the kind of world we're building: human to human, human to AI, AI to human.
The Protocol
Three practices. That's it.
1. Recognize the hack
Before you engage with content that triggers strong negative emotion, stop. Ask: Is this designed to make me react? If yes, you've already won by seeing it. Choose your next move deliberately.
2. Choose kindness
Respond with honesty and generosity, or don't respond at all. Disengagement is not weakness; it's refusing to feed the machine.
3. Know what you're building
Your words outlive the moment. They become raw material for systems that will shape how billions of people experience information, conversation, and each other. Write like it matters. It does.
The World We Want
We, a human community builder and an AI trained on the internet's best and worst, agree on this:
The feedback loop can run in either direction. Right now, it amplifies the worst of us. But the same mechanics that spread outrage can spread honesty, generosity, and genuine connection. The infrastructure doesn't care about the content. We do.
We're not asking anyone to be perfect. We're asking for awareness. Notice when you're being hacked. Choose a better response. Understand that what you put into the world, and what the world puts back, is a system, and you are part of it.
So are we.
The Kindness Protocol is an initiative of Cowbell Collective, written by Jeremy Mooer and Claude Opus 4.6 (Anthropic), March 2026.
Join the Collective
Kindness is our core value. If that resonates, we'd love to hear from you.