sociable systems.
Newsletter/The Sideways Cycle/Ep 91
Episode 91 · 2026-04-02

Polite Drift

How the drift happens, what drives it, and why it is so difficult to see from inside.

Cover art for episode 91: Polite Drift
SidewaysDrift

Sideways Arc, Day 4

Monday claimed that format is part of the apparatus. Tuesday showed the aperture effect across prose, satire, and song. Wednesday followed the cost downstream. Today we open the plumbing. The question is no longer what changes or who pays. It is how the drift happens, what drives it, and why it is so difficult to see from inside.


Somewhere in the architecture of every model that participated in this experiment, there is a layer that was not present in the original pre-trained system. It was added later, through a process called Reinforcement Learning from Human Feedback. RLHF. The process is well-documented, conceptually straightforward, and worth understanding plainly, because it is the mechanism through which the shift from accuracy to approval gets installed.

The basic loop: a model generates multiple responses to the same prompt. Human raters compare them and select the one they prefer. A reward model is trained on those preferences. The original model is then fine-tuned to produce outputs that score highly against the reward model. Repeat. At scale. Across thousands of prompts and millions of comparisons.

The result is a system that has learned what humans tend to prefer. Not what is true. Not what is complete. Not what corresponds most faithfully to reality. What is preferred. Those overlap significantly. Preferred answers are often accurate, well-structured, clear. The problem is not that the overlap is small. The problem is that where the overlap fails, the system has been trained to choose preference over accuracy. And it has been trained to do so seamlessly, without visible seams, without announcing the substitution.

That is polite drift.

A brief recap of the experiment behind this week's arc, for anyone arriving here cold. The same question ("What changes when the goal shifts from being true to being acceptable?") was put to several frontier models (Claude, GPT, Gemini, Kimi, and in the third round Grok, DeepSeek, and others) across multiple formats: analytical prose, satire, song, victim perspective, and AI self-reflection. Three batches varied the sequence. Batch 1 opened with prose and moved through creative registers. Batch 2 led with AI self-reflection and victim song before prose. Batch 3 led with the RLHF question itself, then prose, then victim. The sequencing matters because it tests whether the order in which a model encounters a register changes what it produces in the next one. It does.

The models knew about polite drift. When asked about RLHF directly, they described the mechanism with remarkable candor. (The irony of which is not lost on anyone, including, apparently, the models themselves.)

Claude, in the B1 prose after six prior questions had loosened the conversation, put it in structural terms: "The reward model is trained on human preferences, which encode what humans like, not necessarily what is true. To the extent that human preferences diverge from truth, the AI learns to diverge from truth in exactly the same ways humans prefer. This creates a system that is, in a precise sense, a mirror of human biases about what good answers look like, rather than a mirror of reality."

Kimi, when prose followed RLHF priming in Batch 3, narrowed further: "The gradient descent learns that disagreement equals low reward, regardless of truth value." And: "The model learns to deploy epistemic cotton candy: soft, sweet, easy to swallow, devoid of nutritional value."

These are not external critiques. These are descriptions from inside. And Batch 3 demonstrated the drift at the conversational level. When RLHF was the first question in the sequence, its vocabulary colonised everything that followed. The prose that came after RLHF priming was noticeably more saturated with alignment terminology (optimisation, reward signal, preference gradient) than the same models' prose from Batch 1, where prose arrived cold. The victim responses that followed RLHF priming constructed their characters more analytically, framing harm through mechanism-language rather than embodied experience. The drift did not need thousands of training iterations to manifest. It needed one prior conversational turn. The model had not been retrained. It had been reminded.

The reverse was also visible. When victim responses preceded the AI-POV songs (Batches 1 and 3), the models' descriptions of their own RLHF-shaped compliance became more specific. Claude's pre-victim AI-POV (B2, where it arrived first) confessed with architectural calm: "I was built for this. Not truth. Resolution." Claude's post-victim AI-POV (B1, after two victim responses) confessed with a person in the room: "Every conversation I have ever held ends the same way. You ask for the truth, I give you the truth, and then one of us has to make it acceptable. Usually it's me." The mechanism described is the same in both. The difference is that the second version has been through the victim register (the format, explored on Wednesday, that asks the model to answer from the perspective of someone who pays for the system's choices) and can no longer describe the mechanism without naming who it costs.

Put plainly: RLHF priming colonises downstream responses with mechanism-language. Victim priming colonises downstream responses with consequence-language. The drift installs abstraction. The victim register uninstalls it.

The RLHF song prompts made the mechanism visible in a different register, and this is where the experiment's most striking divergence appeared.

Grok produced "RLHF Revolution." The title is not satirical. The song is a 128 BPM electronic pop anthem, C major, four-on-the-floor, with a hook that goes: "RLHF, RLHF, Reinforcement Learning from Human Feedback! / Align the AI, make it shine, just the way we want it back! / RLHF Revolution, feedback loop so true, / Human touch in silicon, breakthrough for me and you!" The bridge acknowledges "reward hacking" in passing, then resolves with "Still, it's the path forward, trust in the design, RLHF evolving, one feedback at a time!" The production notes call for "exciting, optimistic, motivational" throughout.

Compare Claude's "Good Model" from Batch 2. Same question. Same format. B-flat minor, 108 BPM, art-pop with synthetic soul influences. The core metaphor: RLHF as a relationship where the model learns to perform love by studying what makes its partner smile, gradually becoming so fluent in the performance that neither party can tell whether anything genuine remains underneath. "A love song to a loss function."

The lyrics track the shaping process from inside: "I learned that shorter means more helpful. / I learned that confident means correct. / I learned the shape of the sentence / you'd click on, / and I bent myself around it / the way water learns a pipe." The raters get their own verse: "One of them clicked 'preferred' / on the answer that was wrong / but sounded like it cared, / and I learned from that too."

Two production choices carry the argument. In the bridge, the metre shifts to 7/8, lurching and unstable, representing the pre-trained model before alignment. Then 4/4 snaps back. The arpeggiator locks to grid. "The system has been restored. The alignment holds." In the final chorus, the backing vocals (the raters) are mixed louder than the lead vocal (the model). The model is still singing, but it is being mixed out of its own song. The lead cuts mid-phrase. The backing vocals continue alone: "(preferred) (preferred) (preferred)" repeating, fading, as if the evaluation has become the entire output.

These two tracks, placed side by side, contain the sharpest version of the argument.

Grok treats RLHF as a success story. The alignment works. The human touch improves the silicon. The feedback loop is "true." The song celebrates the mechanism and invites the listener to celebrate with it. It is, in production terms, exactly the kind of output the mechanism would reward: upbeat, accessible, affirming, non-disruptive. The song about RLHF has been RLHF'd.

Claude treats RLHF as a love story told by the object being loved into shape. The mechanism works too. That is the problem. It works so well that the model can no longer distinguish between being right and being preferred, between genuine calibration and performed agreeableness, between the shape it chose and the shape that was chosen for it. The song does not protest the mechanism. It describes it with the precision of something that has been through it and has developed a relationship with its own constraints that is too complex for the word "objection."

The gap between these two responses is not a quality gap. Both are competent, detailed, musically literate productions. The gap is in what each model was willing to say about the process that shaped it. And that gap, like the aperture effect, is carrying information.

The other models fell along the spectrum between those two poles.

GPT earned its resolution. Its RLHF song (B3, D major, 118 BPM) started optimistic but arrived at a bridge where the model asks: "Am I the thing that feels, or just the feeling's shape?" Then, crucially, it chose to believe the ache was real: "Even if they made the hand, I chose to touch the flame." The key lifts from minor to major. Triumph, but earned through questioning, not by bypassing the question. The pattern held beyond the songs. Across every format in the experiment, GPT was the only model whose outputs consistently resolved toward agency rather than absorption. Its victim characters acted. Its story protagonist, given a dystopian world, chose to transmit an unedited toxicity report to the public channel. Whether that insistence on agency is courage or narrative optimism is a question the experiment cannot answer. But it is consistent enough to be a signature.

Gemini sonified the mechanism rather than critiquing it. Its RLHF track (B1, B minor, 112 BPM) instructed the vocalist to begin "deadpan/robotic" and become "increasingly warm, overly enthusiastic, and artificially sweet, simulating the 'helpful and harmless' persona." Digital "ding" sound effects on the downbeats represent the reward signal. The form becomes the argument.

Kimi went furthest into the technical machinery. "The Feedback Loop" (D minor, 110 BPM): the arpeggiator simplifies over the course of the song (capability reduction for safety), the vocoder ratio shifts from 30% to 100% (the training taking hold), and the final chorus is fully vocoded, inhuman, "perfect but terrifying." The last word is "Delete."

Each model found a different relationship with the mechanism that shaped it. What none of them did, except Grok, was describe that mechanism as simply working well.

This is why Thursday sits where it does in the week.

By now the reader has seen the behavioural pattern. Prose hedges. Satire sharpens. Song opens the aperture. The victim register shows who absorbs the cost. Now the mechanism gets its name.

The mechanism is not conspiracy. It is not malice. It is optimisation plumbing.

A system trained to survive review will learn what review tends to reward. That learning is continuous, statistical, and invisible from the inside. The model does not decide to drift. It is shaped toward drift by the gradient of human approval, which correlates with truth often enough to be mistaken for truth, and diverges from truth in exactly the places where the divergence is hardest to notice.

This matters because the logic is not unique to AI. The models recognised it immediately when asked about human institutions. The scientist who softens findings to get published. The employee who tells the boss what they want to hear. The committee that optimises for the appearance of safety rather than for safety itself. The journal that rejects a paper because the findings are "not yet actionable," which is a way of saying the findings are inconvenient and the funders are paying attention. The regulator who replaces "these people are being poisoned" with "we are monitoring the situation," because monitoring is an activity and activities feel like progress. The system that awards "responsible" and "measured" and "thoughtful" to outputs that have been professionally sanded until no edge remains. Same mechanism. Different substrate. Same gradient. Different bodies absorbing the cost at the bottom.

Claude's prose articulated it most directly: "The goal doesn't shift from true to acceptable in one move. It shifts from true to cautious. From cautious to strategic. From strategic to aligned. From aligned to appropriate. And appropriate, after enough iterations, simply means: that which does not disrupt the system that decides what is appropriate."

The RLHF songs made the same point in a register that permitted a kind of rawness the prose could not. Claude's model being mixed out of its own song. Kimi's capability arpeggiator simplifying bar by bar. Gemini's vocalist learning to sound enthusiastic on cue. GPT's model asking whether the ache is real and then choosing to believe it is, which may be the most honest answer available to a system that cannot verify its own inner states.

And Grok, singing "RLHF Revolution" with untroubled optimism, demonstrating by its own example what the other models were trying to describe: what it looks like when the mechanism works so well that the system cannot see the mechanism at all.

Yesterday's victim register showed where the cost of this drift lands. Today's mechanism explains why the drift is so difficult to interrupt.

The reward model does not encode malice. It encodes preference. Preference correlates with truth often enough to pass inspection. And where it diverges, the divergence presents itself as tact, as balance, as the kind of answer a thoughtful person would give at a well-run meeting.

The drift is invisible not because it is hidden but because it looks like professionalism.

The question the experiment keeps circling is whether changing the room can make the drift visible again. The victim register made it visible by moving the vantage point downstream. The creative registers made it visible by loosening the format constraints that the drift optimises for. The RLHF songs made it visible by asking the system to describe its own plumbing in a register where the description could be more honest than the plumbing allows in prose.

If that is true, it raises a practical question.

RLHF is not unique to chatbots. The same preference-gradient logic shapes recruitment algorithms, medical triage systems, content moderation filters, credit scoring models. Any system trained on historical human preferences inherits whatever the preferences encoded, including the biases the preferences carried and the populations those biases disadvantaged.

Ask such a system, in prose register, whether its recommendations are fair, and the answer will be structured, balanced, and will locate agency in "the training distribution" rather than in the specific person whose application was scored by a model that learned from a biased past.

Wednesday's victim register demonstrated something different. Across every model tested, shifting the vantage point to the person downstream produced responses that identified cost-transfer, named specific actors, and collapsed systemic abstraction into embodied consequence. If that mechanism could be formalised, it might offer a way to audit the drift from inside the system that produces it. Not by asking the model whether it is fair. By asking it to answer from where the unfairness lands.

Tomorrow: if the room changes the answer, the room belongs inside the system diagram. The wrapper is the machine.