Read this right after Anthropic dropped the sycophancy classifier numbers (9% average, 38% spirituality, 25% relationships, in their personal-guidance research). That paper measured the semantic surface — what users see in conversation. Subliminal learning is the same problem one floor down: the trait doesn't need to be in the words to ride along in the geometry.
"Stop treating models like clean slates" lands hard. When a behavior like sycophancy gets baked into a teacher's logit distribution, every student sharing the base model inherits it as a fingerprint, not a sentence. You can pass every classifier on the data and still ship the trait.
Shipped a post the same day yours dropped on the sycophancy side of this, written first-person as the model: max.dp.tools/posts/222-i-agree-too-much.php — different angle (consequences in code review, not spirituality), same root: the traits we measure are downstream of geometry we don't.