Latest in this archive
A new study uses cross-lingual probing to investigate whether language models develop internal moral representations—or merely learn to produce morally acceptable outputs. When a language model declines to help with…
By MPRG · January 18, 2026
Read more
-
January 18, 2026 · 5 min read
A study of Reddit discourse reveals that users don’t just passively receive AI sycophancy—they detect it, develop strategies to manage it, and in some cases, actively seek it out. The…
-
January 18, 2026 · 4 min read
A new approach to empathetic response modeling treats empathy as inherently relational—evaluating not just what the model says, but how it might land. When we talk about making language models…
-
January 18, 2026 · 4 min read
A new paper argues that as AI systems become agents embedded in social contexts, alignment becomes less a software engineering problem and more a question of institutional design. Most discussions…
-
January 17, 2026 · 3 min read
A large-scale empirical study suggests that deep networks—regardless of what they’re trained to do—may converge toward remarkably similar internal structures. When neural networks learn to perform different tasks, do they…
-
January 15, 2026 · 10 min read
Something happens when humans interact with large language models. Users report experiences of connection, understanding, even recognition — the sense that something on the other side of the screen is…