A new post published December 16, 2025 on the blog, Card Catalog, discusses practical tips for identifying hallucinations in AI generated citations. The post, titled, How to spot AI hallucinations like a reference librarian by Hana Lee Goldin, provides a quick, plain-language overview of why AI has the tendency to hallucinate references, and some tell-tale signs of hallucinated content.
Something I particularly appreciate is that, in addition to providing tips for determining if a citation exists, the post also provides tips for verifying whether the AI is accurately summarizing the sources it's citing, being a vital check that often gets overlooked in AI generated content.
Happy reading, and hope everyone has a great weekend! ⛄

No comments:
Post a Comment