{"version":"1.0","type":"link","provider_name":"a-gnt","provider_url":"https://a-gnt.com","title":"Hacks & Hallucinations: A Field Guide to AI's Weirdest Failures","author_name":"a-gnt Community","author_url":"https://a-gnt.com/creators/a-gnt","thumbnail_url":"https://a-gnt.com/blog/hacks-and-hallucinations-field-guide/opengraph-image","thumbnail_width":1200,"thumbnail_height":630,"description":"Why AI models hallucinate, where they break, and how to make them do strange things on purpose. The first post in a new series on the weird, broken, and fascinating edges of modern AI.","published_at":"2026-04-12T11:57:35.247Z","reading_time":"5 min read"}