In recent months I have demonstrated repeatedly that ChatGPT frequently generates Holocaust "hallucinations", which I define as new, false information. Examples:
- Me calling out a hallucination which reversed the truth. https://www.codohforum.com/viewtopic.php?p=9736#p9736
- Me calling out (conservatively) 5 hallucinations. https://www.codohforum.com/viewtopic.php?p=10690#p10690
- Me calling out (conservatively) 11 hallucinations. https://www.codohforum.com/viewtopic.php?p=12956#p12956
- Me calling out a fabricated quote. https://www.codohforum.com/viewtopic.php?p=13146#p13146
- Me calling out continued usage of previous hallucinations. https://www.codohforum.com/viewtopic.php?p=13331#p13331
- ConfusedJew admits to a hallucination, going further to say, "The frustrating thing is that it isn't honest about why it makes mistakes". https://www.codohforum.com/viewtopic.php?p=13322#p13322
https://www.gcedclearinghouse.org/sites ... 178eng.pdf
1. AI automated content may invent facts about the Holocaust.
AI models have produced misleading or false narratives about the Holocaust. Data voids and biases have led to “hallucinations” in generative AI systems, producing incorrect or invented content that never occurred. Without AI literacy and research skills, users may not know how to verify AI-produced texts, or recognize the unreliability of the data.
AI has invented events, personalities and even whole phenomena relating to the Holocaust; one example of such hallucination on ChatGPT is the concept of the ‘Holocaust by drowning,’ which assumes that mass murder campaigns took place in which Nazis and their collaborators drowned Jews in rivers and lakes. Although historically no such campaigns took place, AI invented them based on the concept of the Holocaust by bullets – i.e. large-scale murder by shooting. Google’s Bard has, on other occasions, hallucinated fake quotes from witnesses to support distorted narratives of Holocaust massacres (Makhortykh et al., 2023c).
While I'm not an AI expert, my understanding is that, when the AI takes a side on an issue but has insufficient justification for it, it's forced to invent and misinterpret evidence to support its beliefs. Since AIs are trained on a wealth of Holocaust-affirming material, and since Holocaust affirmation may even be baked into their secretive core prompts, it is not surprising that hallucination occurs frequently with this topic.Text-generative AI systems (e.g. Google Bard) may generate false information about lesser-known events during the Holocaust, basing instead their data sets on better-known histories of places and events for which information is more widely available on the internet. For example, Bard’s responses to prompts concerning the 1941 massacres in Liubar in Ukraine reiterate details from the more well-known narrative of Babyn Yar, resulting in historically inaccurate claims, both concerning how victims were killed and whether Ukrainian collaborators were involved in the killing.
Among the problems with allowing AI posts to run wild on the forum:
- Users avoid accountability for the claims and accusations they make.
- Pollution of forum search results with false information.
- Pollution of web search results.
- Pollution of future AI training material.
- Time wasting repetition.
- Degeneration of discussion to instead talk about AI.
- Require all AI posts to be very explicitly designated as such.
- Only permit AI posts in threads that are tagged for AI, perhaps in the topic title.
- Only permit AI posts in a subforum designated for that purpose.
The question could be asked, how is a moderator to determine if a post is or is not using AI? After all, ConfusedJew was making AI posts without acknowledgement from his very first post. To me, AI posts are fairly obvious, so I don't think any complicated guidelines are needed here. Judgement can be used on a case-by-case basis. Otherwise, I'm sure there are objective ways of checking for AI usage, like the common trick of looking for em dashes (—).