Lara Isabelle Rednik May 2026

Laddar favoriter...

Lara Isabelle Rednik May 2026

In an era obsessed with alignment, safety, and scaling, Rednik is the strange, Slavic-inflected whisper reminding us that before we align AI with human values, we should probably make sure we aren't confusing "human values" with "English syntax."

Her 2025 experiment, now known as , found that when asked to generate counterfactual histories (e.g., "What if the printing press had been invented in 100 AD?"), models trained primarily on English produced 40% less creative divergence than models fine-tuned on Romance languages. Lara Isabelle Rednik

Her conclusion was stark: By training our AIs on a global, flattened English corpus, we are not just standardizing language. We are standardizing imagination. Naturally, the tech world has pushed back. OpenAI’s chief ethicist called her work "linguistic determinism dressed up as data science." A prominent Google DeepMind researcher accused her of "romanticizing non-English syntax." In an era obsessed with alignment, safety, and

The Unspoken Pattern (Rednik, 2023) | "The Rednik Threshold" (arXiv:2503.08821) What do you think? Is grammar destiny for AI? Or is Rednik overthinking the subjunctive? Drop your take in the comments. Author Bio: Jordan M. is a recovering digital strategist and M.A. candidate in Language & Technology at Columbia. Naturally, the tech world has pushed back

Digital Humanities / Emerging Voices

April 16, 2026