created: 2024-04-02T09:03
updated: 2024-04-02T09:43
tags:
- AI
- fragments
webcontent_type: article
webcontent_author: Naomi Klein
webcontent_topic: AI
AI machines aren’t ‘hallucinating’. But their makers are.
That’s true – but why call the errors “hallucinations” at all? Why not algorithmic junk? Or glitches? Well, hallucination refers to the mysterious capacity of the human brain to perceive phenomena that are not present, at least not in conventional, materialist terms. By appropriating a word commonly used in psychology, psychedelics and various forms of mysticism, AI’s boosters, while acknowledging the fallibility of their machines, are simultaneously feeding the sector’s most cherished mythology: that by building these large language models, and training them on everything that we humans have written, said and represented visually, they are in the process of birthing an animate intelligence on the cusp of sparking an evolutionary leap for our species.