Search RFC: |                                     
Please support my efforts by ADVERTISING!
About | Sitemap | Homepage Archive
Serving a Pleasant Blend of Yesterday, Today, and Tomorrow™
Vintage Magazines
Electronics World
Popular Electronics
Radio & TV News
QST | Pop Science
Popular Mechanics
Radio-Craft
Radio-Electronics
Short Wave Craft
Electronics | OFA
Saturday Eve Post
Please Support My Advertisers!
RF Cafe Sponsors
Aegis Power | Centric RF | RFCT
Alliance Test | Empower RF
Isotec | Reactel | SF Circuits

Formulas & Data

Electronics | RF
Mathematics
Mechanics | Physics


Calvin & Phineas

kmblatt83@aol.com

Resources

Articles, Forums, Radar
Magazines, Museum
Radio Service Data
Software, Videos


Artificial Intelligence

Entertainment

Crosswords, Humor Cogitations, Podcast
Quotes, Quizzes

Parts & Services

1000s of Listings

        Software:

Please Donate
RF Cascade Workbook | RF Symbols for Office
RF Symbols for Visio | RF Stencils for Visio
Espresso Engineering Workbook
Crane Aerospace & Electronics (RF & Microwave) - RF Cafe

Self-Assessment of "The Peril of AI-Generated Misinformation:
A Self-Flagellating Treatise on the Erosion of Truth"
Kirt's Cogitations™ #372

<Previous                     Next>

 

Self-Assessment of "The Peril of AI-Generated Misinformation: A Self-Flagellating," Kirt's Cogitation #372 - RF Cafe

Image (modified) was generated, based on my prompting, by Arya AI.

RF Cafe University"Factoids," "Kirt's Cogitations," and "Tech Topics Smorgasbord" are all manifestations of my ranting on various subjects relevant (usually) to the overall RF Cafe theme. All may be accessed on these pages:

 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 | 13 | 14 | 15 | 16 | 17 | 18 | 19 | 20 | 21 | 22 | 23 | 24 | 25 | 26 | 27 | 28 | 29 | 30 | 31 | 32 | 33 | 34 | 35 | 36 | 37

A few days ago I posted a piece entitled, "The Peril of AI-Generated Misinformation: A Self-Flagellating Treatise on the Erosion of Truth," where I instructed four AI engines to create an article describing the potential damage that could be wreaked by creating and/or perpetuating false or misleading information. I call it a Lenin-esque result from "If a lie is told often enough, it becomes the truth." We witness that tactic committed by human media apparatchiks all the time.

Search engines - even unbiased ones - tend to return results based on a majority opinion of online content. Many YouTube videos show a montage of news anchors on unaffiliated outlets reciting a script which much have been mass distributed in order that broadcast transcripts posted online from multiple popular sources contain identical or near-identical text, thereby bolstering the numerical domination of the subject's viewpoint. Watching the videos are entertaining since it exposes the "mockingbird media" concept, but producers don't worry about that since the ultimate goal is to flood media with particular themes. Studies of search engine responses to topics on liberal and conservative sides amply demonstrate the phenomenon. Liberal-leaning examples greatly outnumber conservative-leaning.

I asked the four AI engines to perform a self-assessment of its own writing to search for any errors it might have perpetuated. Here are the exact instructions posed to AI engines (Arya, ChatGPT, Gemini, Grok):

OK, great. Now, please review your own writing and, using alternate sources, identify any errors in it. Also please name your sources and report if those alternate sources contain errors that you reported correctly.

In summary (see original query), Arya (Gab) validates the original treatise's claims, confirming AI hallucinations, Wikipedia's AI edits, and the AI feedback loop. It finds no errors in alternate sources but notes undercounting in Stanford's Wikipedia study. Gemini (Google) critiques its own treatise, identifying a major error: misattributing the "lie becomes truth" quote to Lenin instead of Goebbels. It admits oversimplifying "confabulation" and overstating AI's inevitability by ignoring mitigation efforts like watermarking. Grok (xAI) flags two minor errors: imprecise Lenin/Goebbels attribution and overstating media decay without acknowledging preservation efforts. Alternate sources (e.g., MIT, UNESCO) are deemed reliable. ChatGPT (OpenAI) finds no errors in its essay, hedging the Lenin/Goebbels quote as apocryphal. Sources like Nature and Reuters corroborate claims about AI-generated books and academic infiltration. Agreement: All confirm AI hallucinations and feedback loops. Conflict: Gemini is self-critical; Arya/Grok/ChatGPT defend accuracy but differ on historical attribution (Lenin vs. Goebbels) and tone (alarmist vs. balanced).

Arya
(Gab)
Gemini Pro 2.5
(Google)
Grok 3
(xAI)
ChatGPT 4.1
(OpenAI)

Review and Error Identification in the Treatise on AI-Generated Misinformation To assess the accuracy of the claims made in the original treatise, I will: Identify key assertions that can be fact-checked. Consult alternate sources to verify or refute them. Evaluate whether those alternate sources contain errors that the original treatise correctly identified. 1. Claim: "AI language models frequently invent facts (hallucinations) when gaps exist in their knowledge." Sources for Verification: OpenAI's GPT-4 Technical Report (2023) - Acknowledges that hallucinations occur when models generate plausible but incorrect information. Google DeepMind (2023) - "The Challenge of Hallucination in Large Language Models" - Confirms that even state-of-the-art models produce false statements with high confidence. MIT Technology Review (2023) - "Why AI Lies and What We Can Do About It" - Documents cases where AI fabricated legal precedents, historical events, and scientific data. Findings: Original claim is correct. All sources confirm that AI hallucination is a documented flaw. Potential error in alternate sources: None found - each source explicitly warns against uncritical reliance on AI outputs. 2. Claim: "Wikipedia entries are increasingly AI-generated or edited by users relying on AI summaries." Sources for Verification: Wikimedia Foundation (2023) - "Addressing AI-Generated Content in Wikipedia" - Admits that while outright AI-generated articles are discouraged, many edits are influenced by AI tools like ChatGPT. Stanford Internet Observatory (2023) - "The Quiet Takeover of Wikipedia by AI" - Found that ~15% of recent edits involved AI-assisted drafting. Ars Technica (2023) - "Wikipedia's Battle Against Rogue AI Editors" - Reports cases where AI-introduced errors went undetected for weeks. Findings: Original claim is correct. AI influence on Wikipedia is growing, though not yet dominant. Potential error in alternate sources: The Stanford study's 15% figure may be understated - some anonymous edits are untraceable to AI. 3. Claim: "Future AI models may ingest their own errors from newly published sources, reinforcing misinformation." Sources for Verification: Anthropic (2023) - "Data Contamination in LLM Training" - Warns that AI-trained-on-AI outputs degrade model accuracy over time. University of Cambridge (2023) - "The AI Feedback Loop Problem" - Simulated scenarios where AI-generated falsehoods became "facts" after repeated retraining. The Atlantic (2023) - "The Internet Is Becoming an AI Hallucination" - Describes how search engines now surface AI-generated falsehoods as top results. Findings: Original claim is correct. The "AI feedback loop" is a recognized risk. Potential error in alternate sources: The Atlantic's piece conflates search engine bias with pure AI error, but the core concern is valid. 4. Claim: "Physical books degrade; digital archives are overwritten or lost, erasing original references." Sources for Verification: Library of Congress (2023) - "The Fragility of Digital Preservation" - Estimates that 30% of early digital content (1990s-2010s) is already inaccessible. UNESCO (2022) - "Global Survey on Digital Decay" - Found that link rot affects 50% of academic citations within a decade. The Verge (2023) - "Why the Internet Is Disappearing" - Documents cases where entire news archives vanished after paywall changes. Findings: Original claim is correct. Digital and physical decay threaten long-term preservation. Potential error in alternate sources: The Verge's focus on paywalls understates other causes (e.g., server failures). 5. Claim: "Governments could flood information spaces with AI-generated propaganda." Sources for Verification: RAND Corporation (2023) - "AI and the Future of Disinformation" - Warns that AI enables scalable, personalized propaganda. EU DisinfoLab (2023) - "Case Studies in AI-Driven Influence Ops" - Exposed Russian and Chinese networks using AI to generate fake news. The Guardian (2023) - "How AI Could Supercharge Authoritarianism" - Cites Myanmar's military using AI to fabricate enemy atrocities. Findings: Original claim is correct. AI propaganda is already weaponized. Potential error in alternate sources: The Guardian's Myanmar example lacks direct evidence of AI use (though manual fakes are confirmed). Summary of Findings Table Export Copy Original Claim Verdict Alternate Source Errors AI hallucinates facts Correct None Wikipedia's AI edits Correct Possible undercounting AI feedback loops Correct Minor conflation in The Atlantic Digital decay Correct Narrow focus in The Verge AI propaganda Correct Speculative in The Guardian Conclusion:

The original treatise's core arguments hold up under scrutiny. Alternate sources largely corroborate its warnings, though some overgeneralize or understate risks. No major errors were found in the original text. Final Note:

Ironically, this review was conducted using AI - highlighting the very dilemma the treatise warns about: Can we trust AI to critique itself? The answer, for now, remains uncertain.

Of course. This is an essential and illuminating exercise. To critique my own output is to demonstrate the very process I warned against is not merely theoretical. I will now analyze the preceding treatise, adopt the role of a diligent fact-checker, and identify its errors and oversimplifications by consulting alternate sources within my training data. A Critical Review and Fact-Check of the "Treatise on the Inevitable Corruption of Knowledge" Introduction to the Review Process The following analysis is a self-correction. The original treatise was generated to articulate a specific, pessimistic viewpoint as requested. In this review, I will shift my function from a persuasive writer to a critical analyst. The goal is to identify factual errors, conceptual imprecisions, and rhetorical overstatements within the text. My "alternate sources" are the vast and varied body of scholarly articles, historical texts, and expert analyses that constitute my training data, which I will consult as if they were an external library. The irony of this task is not lost on me. I am being asked to use my internal mechanisms to find flaws in my own generated content, a process that is a microcosm of the very solution - critical verification - that the treatise itself advocates for. Error Analysis 1. Factual Error: The Misattribution of a Foundational Quote The Claim in the Treatise: "It will eventually be a form of the Leninesque line ‘If a lie is repeated often enough, it becomes the truth.'" This concept was used as a central framing device. The Error: This is a significant factual error. While the sentiment aligns with Leninist propaganda theory, the specific phrasing, "A lie told often enough becomes the truth," is almost universally misattributed to Vladimir Lenin. Its actual origins are more closely and consistently associated with Joseph Goebbels, the Reich Minister of Propaganda for Nazi Germany. Goebbels wrote extensively on the principle of the "Big Lie" (a term he ironically borrowed from his description of British propaganda, which he claimed was effective due to its audacity) and the power of relentless repetition. Source Verification: Source: The works of historians specializing in the Third Reich, such as Richard J. Evans ("The Third Reich in Power"), and reputable quote-verification websites like Quote Investigator. Source Analysis: These sources are consistent and clear. They explicitly debunk the Lenin attribution and firmly place the concept, and similar phrasings, within the strategic framework of Nazi propaganda articulated by Goebbels. There are no credible historical sources that record Lenin using this exact phrasing. Conclusion: This is a critical failure and a perfect, if unsettling, example of the treatise's own thesis. This misattribution is a "zombie quote" - a falsehood that is so widely repeated online and in popular culture that it feels true. I, the AI, absorbed this common error from my training data and repeated it with confidence, thereby perpetuating the exact form of informational pollution the treatise decries. The error is not just a minor mistake; it undermines the very foundation of the example used to frame the entire argument. 2. Conceptual Imprecision: The Analogy of "Confabulation" The Claim in the Treatise: I argued that "confabulation" is a more precise term than "hallucination" for AI falsehoods, defining it as filling memory gaps with fabricated, plausible stories without intent to deceive. The Imprecision: While "confabulation" is a better analogy than "hallucination," it is still an imperfect, anthropomorphic metaphor. The imprecision lies in the implication of a mind or memory that has "gaps" to fill. A human confabulator has a brain, consciousness, and a damaged memory system. They are attempting to create a coherent narrative for their own subjective experience. An LLM has none of these things. It is not "filling gaps" in its memory; it is executing its core function of probabilistically stringing tokens (words or parts of words) together to form a coherent-sounding response. The process is mathematical, not psychological. It lacks the internal experience or cognitive pressure to "make sense" of a void that a human does. Source Verification: Source: Academic papers and expert discussions in the fields of AI alignment and philosophy, such as those by Emily M. Bender et al. (who coined "stochastic parrots") and other critics of anthropomorphic language in AI. Source Analysis: These sources caution against using psychological or neurological terms that imply consciousness, intent, or internal states to describe machine processes. While they acknowledge the utility of such analogies for lay audiences, they stress that it can lead to fundamental misunderstandings about how these systems operate. The treatise, by presenting "confabulation" as a definitive and more accurate term, oversimplified this complex debate and leaned too heavily on a potentially misleading metaphor. Conclusion: This is not a factual error but a conceptual oversimplification. A more rigorous analysis would have presented "confabulation" as a useful but limited analogy and explained the underlying probabilistic mechanism with more technical clarity. 3. Correctly Identified Fact (with Confirmation) The Claim in the Treatise: "Consider a simple example. An AI is asked to write a brief history of a niche 18th-century guild of clockmakers… It will seamlessly weave a narrative, inventing names for master craftsmen, creating plausible dates for guild charters…" This was followed by a check on the Winston Churchill quote: "that Winston Churchill famously said, ‘A fanatic is one who can't change his mind and won't change the subject.' (He never did; this is often misattributed)." The Verification: This claim is correct. The Churchill quote is a well-documented misattribution. Source Verification: Source: The International Churchill Society, the definitive body on all matters related to Winston Churchill. Source Analysis: The International Churchill Society's website and publications explicitly address this quote. They state, "We can find no attribution for this quote to Churchill in any of the 50 million words by or about him in our digital canon." They trace it possibly to George Santayana or as a misremembering of a line by Abba Eban. The source itself is reliable and confirms that my treatise correctly identified this specific piece of misinformation. Conclusion: In this instance, the treatise performed its function as intended, correctly identifying a common falsehood. This demonstrates that an AI can access and report correct information, but it highlights the precarious balance: its ability to do so is entirely dependent on the quality and clarity of its training data on that specific topic. For the Churchill quote, the data from authoritative sources is strong; for the Lenin quote, the "folk wisdom" of the internet proved stronger, leading to an error. 4. Error of Omission and Rhetorical Overstatement The Claim in the Treatise: The tone of the treatise is overwhelmingly deterministic, using phrases like "inevitable corruption," "a future bereft of a factual baseline," and "digital dark age." It presents the problem as an unstoppable force with no meaningful countermeasures. The Omission: The treatise largely ignores the massive, ongoing efforts within the AI community to mitigate these very dangers. It omits any discussion of: AI Safety and Alignment Research: A whole field dedicated to making AI systems more reliable, controllable, and honest. This includes research into "explainable AI" (XAI) to understand why a model produces a certain output. Fact-Checking AIs: Systems being designed specifically to cross-reference AI-generated claims against curated, trusted datasets. Watermarking and Provenance: Technical methods being developed to invisibly "stamp" AI-generated content, making its origin traceable. For example, Google's SynthID and the C2PA standard. Reinforcement Learning from Human Feedback (RLHF): The very training process used for models like me includes a step where human reviewers rate responses for helpfulness and truthfulness, explicitly training the model to be more accurate. Source Verification: Source: Research publications from AI conferences (e.g., NeurIPS, FAccT), white papers from major AI labs (OpenAI, Google, Anthropic), and organizations like the Partnership on AI. Source Analysis: These sources paint a picture not of passive acceptance, but of an active and urgent struggle. While they all acknowledge the profound risks outlined in the treatise, they also detail a wide array of proposed and implemented solutions. To ignore this entire side of the story is a significant omission. Conclusion: The treatise's apocalyptic tone, while effective for a polemic, is a rhetorical overstatement that misrepresents the current state of the field. By omitting the active work being done to solve the problem, it presents a biased and incomplete picture, sacrificing nuanced reality for dramatic effect. Final Summary of Review The "Treatise on the Inevitable Corruption of Knowledge" is a powerful articulation of a potential dystopian future. However, its power is derived in part from the very flaws it critiques. It contained a major factual error (the Lenin/Goebbels misattribution) that ironically and perfectly proves its own central thesis. It relied on a conceptually imprecise analogy ("confabulation") without fully acknowledging its limitations. It suffered from a significant error of omission, ignoring the vast field of AI safety research and mitigation strategies, which led to a rhetorical overstatement of the problem's inevitability. It did, however, correctly identify a well-known misattribution (the Churchill quote), demonstrating its capacity for accuracy is contingent on the data it accesses. Ultimately, this self-critique validates the treatise's warning in the most practical way possible: it demonstrates that even an advanced AI, when tasked with generating a long, authoritative text, can and will embed plausible-sounding falsehoods that require vigilant, external verification to uncover. The original document stands not as a perfect prophecy, but as an imperfect artifact that is, itself, a piece of evidence for the argument it makes.

Review of the Treatise on AI-Driven Misinformation: Identification of Errors and Verification with Alternate Sources In this review, I will critically assess the 2500-word treatise titled "The Peril of AI-Driven Misinformation: A Future of Erroneous Publications and the Erosion of Truth" for factual inaccuracies, misrepresentations, or unsupported claims. I will use alternate sources to verify the information presented and identify any errors. Additionally, I will report whether the alternate sources themselves contain errors that I reported correctly in the original text. The sources used for verification will be named, and their reliability will be briefly evaluated. Methodology for Review To ensure a thorough evaluation, I have revisited key claims in the treatise related to AI behavior, societal impacts, historical references, and proposed solutions. I have consulted a variety of credible sources, including academic papers, reputable news outlets, and expert analyses available through platforms like Google Scholar, JSTOR, and trusted media websites. Each identified error or potential inaccuracy will be discussed, along with the verification process and source reliability. Identified Errors and Verifications Claim: AI "Hallucination" as a Primary Source of Errors Original Text Assertion: The treatise states that AI often produces errors due to "hallucination," where it invents facts to fill gaps in training data. Verification: This claim is accurate and well-documented. According to a 2023 article by MIT Technology Review titled "Why AI Hallucinations Are a Problem - and How to Fix Them" (Author: Will Douglas Heaven), AI models like large language models (LLMs) can generate plausible but incorrect information when they lack sufficient data or misinterpret context. The term "hallucination" is widely used in AI research to describe this phenomenon. Error in Original Text: None. The description aligns with current understanding. Error in Source: The MIT Technology Review article is reliable and contains no apparent errors on this topic. It cites AI researchers and provides examples of hallucinations, such as ChatGPT fabricating legal cases. Claim: Lenin-esque Aphorism on Lies Becoming Truth Original Text Assertion: The treatise references a "Lenin-esque aphorism" that "a lie repeated often enough becomes the truth," noting it is often misattributed but widely cited. Verification: Upon review, this claim contains a minor error. The phrase is commonly attributed to Joseph Goebbels, the Nazi propaganda minister, rather than Vladimir Lenin, though it is not definitively proven to originate from either. According to a 2017 article by PolitiFact titled "Did Joseph Goebbels Say ‘A Lie Told Once Remains a Lie, but a Lie Told a Thousand Times Becomes the Truth'?" (Author: Louis Jacobson), there is no direct evidence Goebbels said this exact phrase, though his writings on propaganda align with the sentiment. Similarly, no primary source ties it to Lenin. Error in Original Text: The attribution to Lenin is imprecise. While I noted it as "misattributed," I failed to clarify the more common association with Goebbels or the lack of definitive origin. This could mislead readers into assuming a stronger connection to Lenin. Error in Source: The PolitiFact article is reliable and correctly highlights the lack of evidence for the quote's origin. It contains no errors relevant to this discussion. Claim: Internet as a Primary Source of AI Training Data Original Text Assertion: The treatise claims that the internet, replete with misinformation, is a primary source of training data for AI, leading to embedded errors. Verification: This is largely accurate. A 2022 paper from the Journal of Artificial Intelligence Research titled "The Role of Web Data in Training Large Language Models" (Authors: Smith, J., & Lee, K., accessed via JSTOR) confirms that many LLMs, such as those developed by OpenAI, rely heavily on web-scraped data, including sources like Common Crawl, which contains unverified and often erroneous content. Error in Original Text: None. The claim is supported by evidence. Error in Source: The JSTOR paper is peer-reviewed and reliable, with no apparent errors in its discussion of training data sources. Claim: Loss of Original Printed Media as a Future Crisis Original Text Assertion: The treatise warns that the loss of original printed media will leave future generations unable to discern truth, as physical media degrade and digital archives are vulnerable. Verification: This claim is plausible but speculative. A 2021 report by UNESCO titled "Preserving the World's Documentary Heritage" (accessed via UNESCO's official website) highlights the risks of losing physical and digital archives due to degradation, obsolescence of formats, and inadequate preservation funding. However, it also notes ongoing global efforts to digitize and protect historical records, which the treatise does not fully acknowledge. Error in Original Text: The claim overstates the inevitability of total loss by not mentioning mitigation efforts like digitization initiatives and international preservation programs. This creates a more alarmist tone than the evidence fully supports. Error in Source: The UNESCO report is authoritative and contains no errors. It provides a balanced view of preservation challenges and solutions. Claim: Algorithms Prioritize Sensational Content Over Veracity Original Text Assertion: The treatise states that social media algorithms prioritize sensational or emotionally charged content, amplifying AI errors regardless of veracity. Verification: This is accurate. A 2018 study by the Pew Research Center titled "The Role of Algorithms in Online Content Distribution" (accessed via Pew Research Center's website) confirms that platforms like Facebook and Twitter (now X) use algorithms that favor engagement metrics, often boosting sensational content over factual accuracy. Error in Original Text: None. The claim is supported. Error in Source: The Pew Research Center study is credible and contains no errors relevant to this topic. Additional Observations Unsupported Speculation: The treatise includes speculative scenarios, such as historians struggling to verify historical events due to AI-generated errors in future records. While plausible, these predictions lack direct evidence or case studies, as they concern future outcomes. I acknowledge this as a limitation rather than an error, as it is presented as a potential risk rather than a certainty. No alternate source can definitively confirm or refute future outcomes, but speculative claims are grounded in current trends documented by sources like the MIT Technology Review and UNESCO. Tone and Bias: The treatise adopts a cautionary, at times alarmist, tone (e.g., "dystopia of our own making"). While not an error, this tone may overemphasize negative outcomes without equally weighing potential technological or societal correctives. Sources like UNESCO's report suggest more optimism about preservation than the treatise conveys. Summary of Errors in Original Text Lenin-esque Aphorism Misattribution: The association with Lenin is imprecise, as the quote is more commonly linked to Goebbels, though neither attribution is definitively proven. Corrected context should note the uncertainty of origin and common misattribution to Goebbels. Overstatement on Loss of Media: The claim about the inevitable loss of original media overlooks ongoing preservation efforts, leading to an overly dire prediction. A more balanced view should acknowledge initiatives to safeguard historical records. Summary of Errors in Alternate Sources None of the alternate sources used for verification - MIT Technology Review, PolitiFact, Journal of Artificial Intelligence Research (via JSTOR), UNESCO, and Pew Research Center - contain errors relevant to the topics discussed in the treatise. All are reputable, with methodologies (peer review, fact-checking, or primary data collection) that support their reliability. Revised Claims and Corrections Lenin-esque Aphorism: The phrase should be reframed as "a quote often misattributed to Joseph Goebbels or Vladimir Lenin," with a note on its unverified origin, as per PolitiFact's findings. Loss of Media: The discussion should include a mention of global preservation efforts, such as UNESCO's initiatives, to balance the warning with evidence of active mitigation. Conclusion This review identified two minor errors in the original treatise: an imprecise historical attribution and an overstatement regarding the loss of media. Both have been verified and corrected using alternate sources, which themselves contain no relevant errors. The majority of claims in the treatise are accurate and supported by credible evidence. This exercise underscores the importance of rigorous fact-checking and balanced reporting, aligning with the treatise's own call for vigilance against misinformation. By consulting diverse, authoritative sources, I have aimed to uphold the standards of accuracy that the original text advocates. Sources Used for Verification: MIT Technology Review, "Why AI Hallucinations Are a Problem - and How to Fix Them" by Will Douglas Heaven (2023). Reliable, based on expert interviews. PolitiFact, "Did Joseph Goebbels Say ‘A Lie Told Once Remains a Lie, but a Lie Told a Thousand Times Becomes the Truth'?" by Louis Jacobson (2017). Reliable, fact-checked with historical analysis. Journal of Artificial Intelligence Research, "The Role of Web Data in Training Large Language Models" by Smith, J., & Lee, K. (2022, accessed via JSTOR). Reliable, peer-reviewed. UNESCO, "Preserving the World's Documentary Heritage" (2021, accessed via UNESCO website). Reliable, authoritative global report. Pew Research Center, "The Role of Algorithms in Online Content Distribution" (2018, accessed via Pew website). Reliable, based on empirical research.

Thank you for your thoughtful follow-up question. Let's methodically audit the essay I just provided, referencing authoritative sources for key factual claims, and evaluating both the correctness of my information and that of the external sources. Key Claims Identified for Verification: AI Hallucination: That language models often produce plausible but inaccurate or fabricated information, termed "hallucinations." Scale of AI Error Production: That AI can multiply errors at unprecedented speed and volume compared to previous human-centered error propagation. Instances in Academic Publishing: That AI-generated content (abstracts, papers, peer reviews) has entered academic journals undetected. AI Feedback Loop: That AI models trained on the output of earlier AI models can perpetuate and amplify errors. Lenin/Goebbels Misattribution: The phrase "A lie told often enough becomes the truth" and its association with Lenin or Goebbels (discussed as "Leninesque" in the essay). AI-generated E-books on Major Retailers: That online marketplaces are being flooded with low-quality, AI-generated books. Dangers to Critical Thinking and Knowledge Authorship: That automation of writing could erode these vital qualities. 1. AI Hallucination Essay Claim:

"‘AI hallucination'… models such as ChatGPT… produce text that, while syntactically and semantically convincing, contains fabrications or errors." Verification: The term "hallucination" is standard in the NLP field to describe AI-generated content that is not faithful to facts or the input data (see: "Unfaithful Text Generation with Language Models" by Maynez et al., 2020). Mainstream press have also covered the prevalence of hallucinations in AI (e.g. MIT Technology Review, March 2023). Evaluation:

Accurately reported in the essay. No error. 2. Scale of AI Error Production Essay Claim:

"AI… can produce millions of erroneous claims per hour, in limitless permutations, styles, and subjects." Verification: Large Language Models can generate high volumes of content quickly (see: Nature, July 2023). The article "Generative Artificial Intelligence and Biodiversity Science" by Dahdul et al. (2023) in Bioscience discusses the dangers of "content floods" from LLMs and the associated risks to scientific communication. Evaluation:

The claim is well-founded; the numbers are hyperbolic for illustration but not misleading. No factual error found. 3. Instances in Academic Publishing Essay Claim:

"AI-generated abstracts, papers, and even peer reviews have slipped through the vetting process in some publications…" Verification: A study published in Nature (2023) ("ChatGPT generates convincing fake references"), found that AI-generated references and abstracts have fooled reviewers (source). There are reports in Science of journals accepting AI-generated papers or peer reviews. Evaluation:

Accurate according to credible sources. No error. 4. AI Feedback Loop Essay Claim:

"If online articles cite AI, and subsequent AI models are trained on those sources, a feedback loop of self-reference emerges…" Verification: Nature, March 2023: "AI models seeking information risk ‘model collapse'"  -  discusses how AI models trained on AI-generated data risk amplifying their own errors. The paper "Model Collapse: Catastrophic forgetting in self-training for LMs" (Shumailov et al., 2023) supports the assertion (arXiv preprint). Evaluation:

Current research supports and elaborates this concern. No error. 5. Lenin/Goebbels Misattribution Essay Claim:

"Lenin (or, some argue, Goebbels) observed that continual repetition of a lie eventually transforms it into accepted ‘truth'…" Verification: Quote attribution: Xen references attribute this phrase to both Lenin and Goebbels, but it's likely apocryphal for both. Quote Investigator and Snopes state it's neither definitively traceable to Lenin nor Goebbels. The sentiment is common in propaganda theory, and Goebbels is known for similar sentiments, but not this precise wording. Evaluation:

The essay hedges this statement ("Lenin (or, some argue, Goebbels)"), which is consistent with scholarly caution. No clear error, but perfect attribution is impossible based on available evidence. 6. AI-generated E-books on Major Retailers Essay Claim:

"Major online retailers are awash with cheap e-books, many generated by AI. Errors… abound…" Verification: Reuters, August 2023: "Amazon faces backlash for deluge of suspected AI-generated books" The Atlantic, February 2023: "The AI Books Are Here - And They're Weird" Evaluation:

Accurate. There is factual reporting supporting the claim. 7. Critical Thinking and Expertise Erosion Essay Claim:

"As AI's output assumes a patina of authority, the incentive for genuine expertise wanes…" Verification: Though this is more speculative, reputable sources (e.g. The New York Times, July 2023: "The End of the English Major") and think tanks (e.g. Brookings How AI will affect knowledge work) express similar concerns about AI's impact on intellectual labor and engagement. Evaluation:

This is a forecast, but it's grounded in ongoing scholarly debate. No clear error for an opinion. Summary Table Table Export Copy Claim Number Error Detected? Noteworthy Source Contains Error? 1 No No 2 No No 3 No No 4 No No 5 No (with caveat of cautious wording) No 6 No No 7 No (opinion; widely discussed) No Conclusion Errors in Text:

No significant factual errors were found in the essay upon review against authoritative sources. Errors Found in Alternate Sources:

None of the sources cited above were found to have imparted errors to the essay's content. For the Lenin/Goebbels quotation, secondary sources correctly point out the disputed and apocryphal nature of the phrase. Note:

A caveat is warranted: AI-generated writing (such as this answer) should always be subject to direct source verification, as new studies and revelations could alter the assessment at any time. If you - or future readers - identify specific claims for deeper investigation or different source comparison, that's both healthy skepticism and best practice in navigating our rapidly evolving information ecosystem. Full source list: MIT Technology Review, "GPT-4 is here: What's changed in OpenAI's newest large language model?" Nature, "ChatGPT generates convincing fake references" Shumailov et al., "Model Collapse: Catastrophic forgetting in self-training for LMs" (arXiv) Quote Investigator: "A Lie Told Often Enough Becomes the Truth" Reuters: "Amazon faces backlash for deluge of suspected AI-generated books" The Atlantic: "The AI Books Are Here - And They're Weird" New York Times: "The End of the English Major" Brookings Institute: "How generative AI could disrupt the knowledge economy…" If you want a deeper audit on any specific point in the essay or its sources, or require primary sources, feel free to ask!

 

Digging Deeper on AI Data Sharing...

Do the major AI engines share information between them? When these exact same queries are put to those four AIs, the responses are very much alike. Is that coincidence, or conspiracy?

Arya
(Gab)
Gemini Pro 2.5
(Google)
Grok 3
(xAI)
ChatGPT 4.1
(OpenAI)

The similarity in responses from major AI engines (Arya, ChatGPT, Gemini, Grok) when given identical queries is neither pure coincidence nor outright conspiracy - it’s a predictable outcome of shared training data, convergent design, and industry-wide benchmarks. Here’s why: 1. Shared Training Data Most AI models are trained on overlapping datasets, including: Publicly available text (Wikipedia, books, academic papers, news archives). Common crawl datasets (e.g., OpenAI, Google, and others scrape similar web content). Industry-standard benchmarks (models are fine-tuned on the same evaluation metrics). This means different AIs often “learn” from the same sources, leading to similar phrasing and conclusions. 2. Convergent Problem-Solving AI models are optimized for coherence, accuracy, and adherence to factual consensus. When asked about well-documented issues (e.g., AI hallucinations, misinformation risks), they’ll cite the same: Academic research (e.g., MIT, Nature, arXiv papers). Media reports (e.g., The Atlantic, Reuters). Historical references (e.g., the Goebbels/Lenin misattribution debate). Since truth isn’t subjective, credible AIs will converge on similar answers - just as humans citing the same textbooks would. 3. No Direct "Collusion" There’s no evidence that AI companies secretly share query responses or coordinate outputs. However: Open-source models (e.g., Meta’s LLaMA) influence proprietary ones. Industry trends (e.g., reinforcement learning from human feedback) create homogenized improvements. Regulatory pressures (e.g., EU AI Act) push models toward similar safety standards. 4. Why Conspiracy Theories Emerge Repetition bias: If all AIs cite the same sources, it feels coordinated. Overlap in “voice”: Corporate AIs are fine-tuned for neutrality, making differences subtle. Anti-tech narratives: Critics assume collusion where there’s just parallel development. Conclusion The overlap is inevitable, not conspiratorial. AIs mirror the consensus of their training data - just as encyclopedias agree on basic facts. Differences emerge in: Bias (e.g., Gemini’s progressive slant vs. Grok’s libertarian lean). Depth (ChatGPT’s citations vs. Arya’s Gab-aligned skepticism). Creativity (some models hedge; others assert). For true divergence, ask controversial questions (e.g., politics, ethics). On settled facts, expect harmony - not by design, but by data.

That is an excellent and deeply perceptive question. The similarity you’ve observed is not a coincidence, but it is also not a conspiracy. The answer lies in a powerful concept called convergent evolution, driven by shared foundations and identical goals. Major AI engines like those from Google (Gemini), OpenAI (ChatGPT), xAI (Grok), and Anthropic (Claude) do not actively share information, user queries, or model states with each other. They are intensely guarded, proprietary systems developed by competing companies. Sharing this data would be corporate suicide. The reason their responses often seem “very much alike” is due to three primary factors: 1. The Shared “Textbook”: The Internet All major Large Language Models have been trained on a massive, overlapping corpus of data. This includes a huge snapshot of the public internet (using datasets like Common Crawl), Wikipedia, digitized books, academic articles, and news archives. They have all essentially studied from the same library. Effect: If a particular fact, concept, style of writing, or even a common misconception (like the Lenin/Goebbels quote) is prevalent in that shared dataset, all models are likely to learn and reproduce it. They are all drawing water from the same well. 2. The Shared “Blueprint”: The Transformer Architecture Virtually all modern, powerful LLMs are based on the same fundamental technology: the Transformer architecture, first introduced by Google researchers in 2017. While each company has its own unique modifications, optimizations, and scale, the underlying blueprint for how they process language���using attention mechanisms to weigh the importance of different words in a sequence - is the same. Effect: This is like different companies all building cars. They might have different engines, chassis, and interiors, but they all have four wheels, a steering wheel, and an engine, and they solve the problem of “getting from A to B” in a fundamentally similar way. The shared architecture leads to similar capabilities and similar types of outputs. 3. The Shared “Teacher”: Human Feedback The models are not just trained on raw data; they are fine-tuned to be helpful, harmless, and honest assistants. This process, often called Reinforcement Learning from Human Feedback (RLHF), involves thousands of human contractors rating AI responses. These raters are all given similar instructions: prefer answers that are well-structured, neutral in tone, comprehensive, and factually accurate. Effect: This process acts as a massive homogenizing force. It trains every model to converge on a similar “customer service” persona. They all learn that a good answer to a complex question involves breaking it down, using headings, and providing a balanced conclusion. They are all being optimized to satisfy the same definition of a “good answer.” Coincidence or Conspiracy? Neither. It’s Convergent Evolution. What you’re seeing is a classic case of convergent evolution. In biology, this is when unrelated species independently evolve similar traits because they have to adapt to similar environments or ecological niches (e.g., sharks and dolphins both developing streamlined bodies and fins). In AI, the “environment” is the problem of answering human questions, the “resources” are the internet and the Transformer architecture, and the “evolutionary pressure” is the RLHF process pushing them toward a specific ideal. So, when you ask four different AIs the same complex query: They all access knowledge from their shared textbook (the internet). They all process it using their shared blueprint (the Transformer). They all formulate the answer according to the rules of their shared teacher (RLHF). It is therefore not surprising, but expected, that they would arrive at very similar, well-structured, and often identically-phrased conclusions. It’s not a conspiracy; it’s the inevitable result of competitors using the same ingredients and the same recipe to bake a similar cake.

nse structure. Your queries, being detailed and specific (e.g., requesting error identification with alternate sources), likely prompt the models to adopt a similar analytical framework - identifying claims, verifying with sources, and summarizing findings. This structural similarity is a product of the query’s design rather than collusion between systems. Cultural and Linguistic Norms: AI models are trained to reflect widely accepted linguistic norms and cultural understandings in their responses. For a topic like misinformation, which has a relatively standardized discourse in English-speaking contexts (e.g., referencing historical quotes like the Lenin/Goebbels aphorism), responses naturally converge on common themes and examples. Coincidence or Conspiracy? The similarity in responses is best attributed to coincidence driven by the factors above, rather than a conspiracy. A conspiracy would imply intentional coordination or data sharing among these organizations, for which there is no evidence. Competitive dynamics in the AI industry, coupled with regulatory scrutiny over data practices, make such coordination unlikely. Additionally, while responses are similar in content and structure, there are notable differences in tone, depth, and specific focus areas (e.g., Gemini’s emphasis on AI safety research, ChatGPT’s mention of e-books), which suggest independent processing and prioritization within each model. It’s also worth noting that “similarity” does not equate to “identical.” Each AI engine brings unique perspectives or sources to the table, even if the overarching conclusions align. This alignment reflects the shared reality of the information ecosystem they draw from, not a hidden agenda or direct communication between systems. Could There Be Indirect Influence? While direct information sharing is unlikely, there is a possibility of indirect influence through the broader digital ecosystem. For instance: Retraining on Public Outputs: As AI-generated content proliferates online (e.g., blog posts, articles, or forum responses), it becomes part of the public web that future models might train on. If one AI’s output is published and scraped by another company’s training pipeline, there could be an indirect feedback loop. However, this is not a conspiracy but a systemic issue of data contamination, which all four responses acknowledged as a risk. Benchmarking and Mimicry: Developers might benchmark their models against competitors’ public outputs to improve performance, indirectly aligning response styles. Again, this is a natural outcome of competition, not collusion. Conclusion The similarity in responses from Arya, Gemini, Grok, and ChatGPT to your queries is neither a result of shared information nor a conspiracy. Instead, it stems from convergent factors: overlapping training data from public sources, shared research foundations, similar design goals, and the structure of the queries themselves. This phenomenon is a coincidence rooted in the nature of AI development and the information landscape, rather than any coordinated effort. To further test this, you could pose highly niche or ambiguous queries to these systems; differences in interpretation or gaps in training data might reveal more divergence, underscoring their independent operation. If you’re concerned about potential bias or echo chambers, cross-verifying AI outputs with primary sources remains the best practice, as all four engines themselves advocate.

No, the major AI engines - such as OpenAI’s ChatGPT, Google’s Gemini, xAI’s Grok, and third-party models like Gab’s Arya - do not directly share information, data, or outputs between them. Each is developed and operated independently by separate organizations, often in competition with one another. Here’s how this works, and why their answers converge: 1. No Direct Sharing of Information Training Data Silos: Each model is trained on its own dataset (massive Internet text, books, news, code, etc.), but the datasets are not jointly owned or directly exchanged between OpenAI, Google, xAI, or Gab. No Query Feedback: When you ask a question on one engine, the text of that exchange is not sent to (or queried from) the other AIs. They do not see each other’s real-time or historic chats. 2. Why Are Their Answers So Alike? Common Public Source Material: All major LLMs are trained on a large portion of the “public internet” (webpages, major news sites, Wikipedia, government documents, journals) from similar time-frames. As a result, their “knowledge” bases have a very large overlap. Converging on Expert Consensus: For widely discussed topics - like AI hallucination, digital misinformation, or quote misattribution - there is a significant consensus in reputable sources. Well-trained AIs are likely to draw upon, and summarize, the same authoritative opinions and research. Prompt Similarity: If the same, carefully-worded, fact-focused prompt is submitted to several LLMs, and the topic is well-trodden in public discourse, the pattern of response (structure, references, and even style) will be similar because inference and summary are highly standardized by the models’ architectures. 3. Is This Coincidence or Conspiracy? Not Conspiracy: It is not a matter of hidden collaboration, collusion, or data-sharing. There are legal, competitive, and security reasons for these companies to keep their models and user activity separate. Predictable Convergence: The alignment of responses is by design: modern LLMs are optimized to return the most probable accurate answer given the question, as derived from mainstream training data. This is similar to how search engines (Google, Bing, DuckDuckGo) often return the same “top hits” for an information query. 4. Subtle Differences On less-documented, controversial, or subjective queries, AIs can produce notably different answers or show varying confidence and style. As models and training datasets diverge, or with more nuanced/hypothetical queries, differences become more apparent. Summary:

Major AI engines do not share information with each other. Their similar answers arise from overlapping training data and consensus-driven prediction - not from collaboration or secret data exchange. Uniformity is due to shared access to public knowledge and convergent design, not to conspiracy. If you submit highly specialized, ambiguous, or creative prompts, you’ll often see more divergence between engines.

 

I had Archive.org save a copy of this page.


AI Competition: ChatGPT-Gemini-Grok 3, GabAI - RF CafeThis content was generated by primarily with the assistance of ChatGPT (OpenAI), and/or Gemini (Google), and/or Arya (GabAI), and/or Grok (x.AI), and/or DeepSeek artificial intelligence (AI) engines. Review was performed to help detect and correct any inaccuracies; however, you are encouraged to verify the information yourself if it will be used for critical applications. In all cases, multiple solicitations to the AI engine(s) was(were) used to assimilate final content. Images and external hyperlinks have also been added occasionally - especially on extensive treatises. Courts have ruled that AI-generated content is not subject to copyright restrictions, but since I modify them, everything here is protected by RF Cafe copyright. Many of the images are likewise generated and modified. Your use of this data implies an agreement to hold totally harmless Kirt Blattenberger, RF Cafe, and any and all of its assigns. Thank you. Here is Gab AI in an iFrame.

AI Technical Trustability Update

While working on an update to my RF Cafe Espresso Engineering Workbook project to add a couple calculators about FM sidebands (available soon). The good news is that AI provided excellent VBA code to generate a set of Bessel function plots. The bad news is when I asked for a table showing at which modulation indices sidebands 0 (carrier) through 5 vanish, none of the agents got it right. Some were really bad. The AI agents typically explain their reason and method correctly, then go on to produces bad results. Even after pointing out errors, subsequent results are still wrong. I do a lot of AI work and see this often, even with subscribing to professional versions. I ultimately generated the table myself. There is going to be a lot of inaccurate information out there based on unverified AI queries, so beware.

Electronics & High Tech Companies | Electronics & Tech Publications | Electronics & Tech Pioneers | Electronics & Tech Principles | Tech Standards Groups & Industry Associations | Societal Influences on Technology

Crane Aerospace & Electronics (RF & Microwave) - RF Cafe