LLM-Powered Crowd Simulation
Most tools analyze content after it's published. I wanted to flip that — use LLMs to simulate how an audience would react before you post.
Not a sentiment classifier. A full multi-agent simulation where LLM-powered personas read your draft, react, argue with each other, and surface the fault lines you didn't see coming.
Why LLM Simulation Beats Classification
Audiences fracture. A post about a CEO stepping down hits differently for employees, investors, journalists, and Twitter shitposters. A single sentiment score collapses all of that into nothing useful.
Reactions are emergent. One inflammatory reply shifts the whole thread. A pile-on builds momentum. These second-order effects come from agents interacting with each other — something no classifier captures.
LLMs can role-play with context. A classifier sees words. An LLM-backed agent can research the current landscape, adopt a persona grounded in real public discourse, and react the way a real person in that position would.
So I'm Building a Tool for This
I call it CrowdSimulator. You give it a draft post, it researches the topic, generates a synthetic audience grounded in real public discourse, simulates the conversation across platforms, and tells you where the fault lines are.
Here's what it looks like:
How It Works
Phase 1: Research
Before generating any personas, the system runs 15-20 web searches to understand how people actually talk about the topic right now. Public sentiment, active controversies, competitor posts, cultural context, platform-specific patterns. This grounds everything that follows — no hallucinated context.
Phase 2: LLM-Powered Persona Generation
Generic personas are useless. "Sarah, 28, marketing manager, positive outlook" tells you nothing.
CrowdSimulator generates personas derived from the research. Not "skeptical tech worker" but "skeptical tech worker who saw the company's last launch fail and has been vocal about it on r/technology."
{
"agent_id": 7,
"name": "Marcus Webb",
"archetype": "skeptic",
"persona": "Senior engineer who saw the data breach in January.
Posts on HackerNews about corporate accountability...",
"sentiment_bias": -0.4,
"influence_weight": 0.7,
"activity_level": "high",
"mbti": "INTJ"
}15+ personas with enforced diversity — loyalists, skeptics, amplifiers, lurkers, contrarians — each representing a real audience segment the research identified.
Phase 3: Multi-Agent LLM Simulation
All personas drop into a simulated social media environment powered by OASIS. Runs on Twitter and Reddit simultaneously. Each agent is backed by an LLM and independently decides to post, reply, like, argue, or stay silent on every round.
The key: agents react to each other, not just the original post.
Round 1: @techcritical "This is just PR spin. Remember January?"
Round 1: @brand_fan_lisa "Actually the product has improved a lot"
Round 2: @devmarcus likes @techcritical's post
Round 3: @techcritical replies with the r/technology thread
Round 3: @brand_fan_lisa reposts the original "Give them a chance"
Pile-ons emerge organically. Or they don't. The dynamics aren't scripted — they're emergent. Every action streams to the frontend in real time via WebSocket.
Phase 4: Analysis
After the simulation:
- Sentiment score and risk score — grounded in what actually played out
- Faction breakdown — who defended, who attacked, what proportion
- Virality estimate — based on repost and engagement patterns
- Strategy recommendations and a suggested rewrite
Not "sentiment: mixed" but "risk score 7 because the skeptic faction latched onto the January incident and the loyalist responses weren't enough to shift the narrative."
Why LLM Simulation Works
Research eliminates hallucination. Personas cite real controversies because the LLMs were fed real research, not inventing context.
Emergence captures nonlinearity. The same post can play out differently depending on who responds first and whether the pile-on reaches critical mass. That's what makes this a simulation, not a prediction.
Platform separation matters. Twitter rewards hot takes. Reddit rewards sourced analysis. Same content, different crowd dynamics. The LLMs adapt their behavior per platform.
Faction analysis comes free. Every agent already has a label, so you get audience segmentation without clustering algorithms.
What It's Not
It's a stress test, not a crystal ball. 15 synthetic agents don't replace real feedback. But they reveal the structural dynamics — which segments will care, where the fault lines are, and whether the post has obvious risks worth fixing before publishing.
If you're testing content for brands or want to talk multi-agent simulation, reach out.