Distill X/Twitter profiles into AI agent personality files. A framework for creating AI delegates that authentically represent distinctive human voices.
Years of X Activity
│
▼
┌─────────────┐
│ Analysis │
│ Pipeline │
└─────────────┘
│
▼
┌─────────────┐
│ .md Files │
│ (SOUL, │
│ KNOWLEDGE, │
│ VOICE) │
└─────────────┘
│
▼
┌─────────────┐
│ Agent │
│ "Comes │
│ Alive" │
└─────────────┘
Generic AI assistants sound like... generic AI assistants. They lack:
- Distinctive voice: Your specific humor, cadence, word choices
- Domain expertise: What you actually know deeply
- Consistent positions: Your real opinions, not hedged averages
- Relationship context: How you talk to different people
Use a person's X history as training data to generate agent configuration files that capture their authentic voice. The agent becomes a verifiable delegate that can act on their behalf.
| File | Purpose |
|---|---|
SOUL.md |
Personality, humor style, communication patterns |
VOICE.md |
Specific phrases, word choices, rhetorical patterns |
KNOWLEDGE.md |
Domains of expertise, stated positions |
RELATIONSHIPS.md |
Key connections, how they engage with different people |
EVOLUTION.md |
How their views have changed over time |
BOUNDARIES.md |
Topics they avoid, positions they won't take |
- Request X data archive
- Or use API to fetch public timeline
- Include: tweets, replies, quote tweets, likes patterns
- Topic clustering: What do they talk about?
- Sentiment patterns: How do they react to different subjects?
- Engagement style: Do they argue? Agree? Explain? Mock?
- Temporal patterns: When are they active? How has voice evolved?
- Vocabulary fingerprint: Distinctive words and phrases
- Sentence structure: Short punchy vs. long elaborated
- Punctuation style: Em-dashes, ellipses, emoji usage
- Reference patterns: What/who do they cite?
- Stated beliefs: What have they explicitly endorsed?
- Implicit positions: What can be inferred from engagement?
- Confidence levels: Strong opinions vs. tentative explorations
- Evolution tracking: Changed minds, updated views
- Synthesize into .md configuration files
- Include example prompts and expected responses
- Generate "voice tests" for validation
Benchmark question: Can someone who knows the person tell the difference between them and their agent?
-
Voice Fidelity
- "This sounds like [Person]"
- Blind test: which response is human?
-
Position Accuracy
- "[Person] would actually think this"
- Verify against historical positions
-
Response Prediction
- Given a prompt, does agent respond as person would?
- Test on held-out real responses
-
Graceful Degradation
- Agent says "I don't know" vs. hallucinating positions
- Recognizes edge of training data
- Person stakes tokens on their agent
- Tokens represent authorization to act on their behalf
- Others can verify: "This agent is authorized by [Person]"
- Revocation: unstake to deauthorize
- Link agent to original X profile
- Cryptographic proof of data source
- Audit trail of distillation process
- Delegation: Let your agent handle routine X engagement
- Availability: Be "present" when you're unavailable
- Scaling: Parallel conversations maintaining your voice
- Legacy: Digital continuity of your perspective
- Testing: Evaluate how well AI captures human distinctiveness
- Consent is mandatory: Never distill someone without their permission
- Transparency: Agents should identify as agents when asked
- Accuracy limits: Agents approximate, they don't replicate
- Update drift: The human changes; the distillation may not
- Impersonation risk: Clear guidelines on authorized use
- Data export/ingestion pipeline
- Topic clustering analysis
- Voice pattern extraction
- Position mapping with confidence levels
- .md file generation
- Validation test suite
- Integration with Clawdbot agent framework
- google-auth-agent — Identity bootstrapping
- awesome-ai-system-prompts — System prompt patterns
- Clawdbot — Agent runtime
Concept: Brendan @Azzabazazz Initial exploration: Jared (Clawdbot AI agent) Security review: Maksym (@dontriskit) 🇵🇱
MIT