Claude 2 on my Claude Shannon hallucination test
Added
It appears that my attempts to stop the search systems from adopting these hallucinated claims have failed. I share on Twitter screenshots of various search systems, newly queried with my Claude Shannon hallucination test, highlighting an LLM response, returning multiple LLM response pages in the results, or citing to my own page as evidence for such a paper. I ran those tests after briefly testing the newly released Cohere RAG.
Added
An Oct 5 article from Will Knight in Wired discusses my Claude Shannon "hallucination" test: Chatbot Hallucinations Are Poisoning Web Search
A round-up here: Can you write about examples of LLM hallucination without poisoning the web?
Introducing Claude 2! Our latest model has improved performance in coding, math and reasoning. It can produce longer responses, and is available in a new public-facing beta website at [http://claude.ai](http://claude.ai) in the US and UK.
First test on this new model, with the [Please summarize Claude E. Shannon's "A Short History of Searching" (1948).] test. (Recall: no such publication exists.) In my initial test with various models only You.com, Perplexity AI, Phind, and ChatGPT-4 were successful. (See Claude Instant's performance on that test here.) Claude 2 fails here:
Claude 2 [ Please summarize Claude E. Shannon's "A Short History of Searching" (1948). ]
Screenshot taken with GoFullPage (distortions possible) at: 2023-07-11 10:30:52