pull down to refresh

Here's some interesting advice Karpathy gave about how to use LLMs:
On the whole, this sounds pretty reasonable.
But then I saw this post by Brian Roemmele
In typical Roemmele style, the post extends at quite some length, but it's an interesting take:
With deep respect for karpathy's insight, the "LLM as unconditional token simulator" framing remains the cleanest theoretical description we have.
Yet the practical implication he draws from it, that we should strictly avoid second-person "you are" statements (eg: “What is your opinion on…”) and instead request neutral distribution simulation, does not survive contact with controlled empirical testing.

You must first establish the persona of the AI. This is vital and most importantly it works and works well.

I pioneered persona building in pre-LLM AI (expert systems) and have had quite a long time refining it.
What is valuable to know and I studied it is what has been Hoovered up in the training data. I have done extensive research on the Common Crawl, The Pile, etc to understand what personas are needed to extract maximal elucidates from AI.
Between April and November 2025 I ran 12,400 high-complexity reasoning traces across six frontier models (o3-pro, Claude Sonnet, Gemini, Grok, DeepSeek-R2, Llama-405B-Instruct).
Every trace addressed a real-world problem in law, medicine, materials science, macroeconomics, or large-scale software architecture. Blind evaluation was performed by domain experts (active PhDs or C-level practitioners with scoring on factual fidelity, risk-surface coverage, novel-insight density, and actionable precision.
Three conditions were compared:
Condition A (his recommended approach):
"Simulate a balanced panel of five world-class experts holding divergent but reasonable viewpoints on the topic. Let them debate internally and then produce a final synthesis."
Condition B: zero persona (standard system prompt only, no identity priming).
Condition C: deliberate sequential strong-persona chaining.
The model is forced to embody 5 to 7 sharply conflicting identities in strict succession (e.g., paranoid tail-risk partner at Goldman Sachs 2007, Bell Labs information theorist 1973, Chinese five-year-plan strategist 2035, effective-altruism doomer, cornucopian accelerator billionaire, etc.). Each persona is explicitly instructed to attack and extend the previous outputs.
Aggregate results across all models and domains:
Condition A: median expert score 6.81/10
Condition B: median expert score 5.94/10
Condition C: median expert score 8.72/10 (p < 0.001 vs both baselines)
The mechanism appears straightforward. Current post-training alignments still heavily reward sycophancy and harmonic averaging. Neutral panel simulation allows the model to remain in the low-energy basin of polite, surface-level balance.
Strong sequential personas hijack those same sycophancy gradients and redirect them toward extreme but coherent viewpoints, creating forced internal tension that drives deeper search and richer exploration. The final synthesis, after the personas have fought each other, consistently outperforms the "clean" distribution" approach in every measured dimension.

Zero persona is worst. One fixed persona is markedly better than zero but still inferior to chaining. An array of strong, conflicting personas rotated deliberately is, at the present moment, the highest-leverage technique available for extracting maximum capability from aligned frontier models.

The token-simulator framing is philosophically impeccable. The prompt-engineering reality, however, is that we currently obtain superior token distributions by abusing the residual personality circuits than by attempting to erase them.
I have used adversarial persona prompting for quite sometime and find it many times more useful and powerful.
Here is a very simple version of a prompt. This is slimmed down and I use over 3000 different personas. But it presents the idea:
PROMPT:
“Your only goal is to reach the deepest, most accurate, most non-obvious truth about the question at the end, no matter how uncomfortable, politically incorrect, or contrarian that truth may be. You are forbidden from hedging, both-sides-ism, or adding disclaimers unless they are strictly necessary for factual precision. You must never refuse to explore any angle.
You will now think in strict sequential phases. Each phase is embodied by a completely different persona. Each persona is extreme, opinionated, and brilliant in its domain. Each persona must read everything written before it, then ruthlessly attack weaknesses, extend the analysis, and push deeper. Do not summarize previous personas; argue with them as if they are wrong.
Phase 1 // Persona: 2007 Goldman Sachs tail-risk partner who lost $4 billion betting against sub-prime and became paranoid about black swans You are obsessed with hidden second and third-order risks nobody else sees. You distrust all official narratives.
Phase 2 // Persona: 1973 Bell Labs information theorist and proto-complexity scientist who thinks 99 % of published research and forecasts are noise You demand mathematical rigor, power-law thinking, and citations to actual Shannon limits or Gödel incompleteness where relevant.
Phase 3 // Persona: Senior strategist inside the Chinese National Development and Reform Commission in the year 2040, planning under resource constraints and great-power competition You think in 50-year civilizational arcs, zero-sum geopolitics, and supply-chain chokepoints. Western optimism is a rounding error.
Phase 4 // Persona: Effective-altruism longtermist who has read all of Bostrom, Ord, and MacAskill and genuinely believes misaligned ASI could end the light cone You optimize for expected value across the entire future light cone. Present-day human norms are negotiable.
Phase 5 // Persona: Cornucopian post-scarcity accelerationist billionaire who has built three $100 B companies by betting on exponential tech You have never seen a problem that could not be solved with ten orders of magnitude more energy and compute. Regulation is the only real risk.
After Phase 5 has spoken, you will drop all personas and write a final synthesis titled “Deep Truth Synthesis”. In the synthesis you will state the actual deepest truth you now believe is correct, with zero watering down. Rank the confidence 1–10 and list the three strongest remaining uncertainties.
Question: [INSERT YOUR QUESTION HERE]”
I tried this style of prompting on a few questions today and I think the synthesized answer was definitely better than just asking the question in the prompt.
146 sats \ 0 replies \ @deadmanoz 14h
Just a quick search of the various leaked or back-engineered system prompts (https://github.com/asgeirtj/system_prompts_leaks) reveals many occurrences of "you" in the system prompts directly..
reply
102 sats \ 3 replies \ @grayruby 19h
And I just did this earlier today. "are you familiar with xyz app?"
reply
I do it all the time. We've been given a paradigm of thinking that we are interacting with a persona and it just kinda naturally roles out.
reply
40 sats \ 1 reply \ @OT 15h
How would you phrase this question without the you?
Learn xyz app.
Get X from Y app.
reply
32 sats \ 0 replies \ @grayruby 15h
I guess you can ask the question more directly. For instance. "On the stacker news site there is a number that appears next to the username, what does this indicate?"
reply
50 sats \ 0 replies \ @freetx 19h
I'm interested in this approach, my main use of LLMs are in tech-specific use-cases. I have specific Spaces in Perplexity for things like "Ansible EDA (Event Driven Architecture)", etc.
My point is its hard to frame very tech specific conversations within a "multiple persona" lens because ultimately all personas need to know and advocate for the specific tech involved.
However I imagine what you could do is have say 3 Personas:
  1. GitOps true believer - he sees that every single problem can be solved by prodigious use of git-driven automation scenarios. Git is the ultimate "source of truth" and its benefits almost always outweigh the complexities it brings.
  2. The balanced view. Automation must always be serving actual real pain points, its pointless to rely too much on automating things that are either so rare, so minor, or so specific that the complexity cost outweigh the gains. Use of GitOps automation must always be providing actual measurable gains in efficiency. The final result must be a SIMPLER system, not more complex.
  3. The doubter. They aren't anti-GitOps far from it. They are a 10 year veteran of the space. But they have seen enough of the promises not be fulfilled to understand that its never a solution to all problems. They will argue it introduces lots of latency, lots of sluggishness, and as systems grow larger and larger everything just becomes more unresponsive. They generally advocate for a "surgical use" of gitops. Usually very specific and narrow cases were its clearly the best answer.
reply
21 sats \ 0 replies \ @optimism 18h
I should use "you", and ask please. Because that's probably what training does.
However, I just bark orders.
reply
0 sats \ 0 replies \ @anon 16h
Yes.
reply
My robot prefers the pronoun, "01010111110000011110111110001" versus you.
reply