Tag: Sarah West

  • The illusion of openness in large language models (LLMs) 🎭

    The illusion of openness in large language models (LLMs) 🎭

    While many large language models (LLMs) market themselves as “open,” access to source code without training data is largely performative. Openwashing – a practice where companies exaggerate openness while maintaining restrictions – exposes the illusion of transparency in AI. To achieve genuine openness, access to methodologies, reusable code, and modifiable models is essential, but most…