GPT-3 has a startling ability to produce text that sounds like it was written by a human. But a new study shows it also has a worrying proclivity for anti-Muslim bias. Researchers from Stanford and McMaster universities probed the neural network on tasks including prompt completion, analogical reasoning, and story generation. They discovered that a Muslim-violence bias appears consistently — and creatively — in various uses of the model: While these associations between Muslims and violence are learned during pretraining, they do not seem to be memorized; rather, GPT-3 manifests the underlying biases quite creatively, demonstrating the powerful ability of language models to mutate biases in different… This story continues at The Next Web