Summary
The episode discusses a new study revealing that large language models, including GPT‑5 and open‑source alternatives, systematically assign negative stereotypes to speakers of German regional dialects compared to Standard German, influencing decisions in hiring, education, and other contexts. Researchers from Johannes Gutenberg University Mainz and partner institutions demonstrated that bias intensifies when dialects are explicitly identified and that larger models can exhibit stronger prejudice. The guests, doctoral researcher Minh Duc Bui and Professor Katharina von der Wense, explain that these biases stem from training data reflecting societal stereotypes and stress the need for fairer model design to respect linguistic diversity.

Comments
Want to join the conversation?
Loading comments...