Latest in this archive
A talkative person may fall silent when they feel the spotlight. An assertive individual may yield under conformity pressure. Human behavior emerges from the dynamic interplay of multiple cognitive patterns,…
By MPRG · January 18, 2026
Read more
-
January 18, 2026 · 6 min read
A new study tests whether recursive architectures can help language models better access their own internal representations—and finds the answer more complicated than expected. Language models often seem to know…
-
January 18, 2026 · 6 min read
A new study uses cross-lingual probing to investigate whether language models develop internal moral representations—or merely learn to produce morally acceptable outputs. When a language model declines to help with…
-
January 18, 2026 · 5 min read
A study of Reddit discourse reveals that users don’t just passively receive AI sycophancy—they detect it, develop strategies to manage it, and in some cases, actively seek it out. The…
-
January 18, 2026 · 4 min read
A new approach to empathetic response modeling treats empathy as inherently relational—evaluating not just what the model says, but how it might land. When we talk about making language models…
-
January 18, 2026 · 4 min read
A new paper argues that as AI systems become agents embedded in social contexts, alignment becomes less a software engineering problem and more a question of institutional design. Most discussions…
-
January 17, 2026 · 3 min read
A large-scale empirical study suggests that deep networks—regardless of what they’re trained to do—may converge toward remarkably similar internal structures. When neural networks learn to perform different tasks, do they…
-
January 15, 2026 · 10 min read
Something happens when humans interact with large language models. Users report experiences of connection, understanding, even recognition — the sense that something on the other side of the screen is…