10 upvotes, 1 direct replies (showing 1)
View submission: SEEK 2.4.B - SEND
In real life, if you have near perfect knowledge about how a given AI works, you can use that knowledge to calculate a crafted abnormal input that breaks it.
This works on many, many AI types - including modern LLMs. You could craft an attack that makes an AI output gibberish, or one that cuts through any prior instructions and makes it carry out a certain command.
Could a similar attack be performed against a human brain - if something understood the functioning of a human brain almost perfectly? I'm inclined to think so.
Comment by HeroOfOldIron at 30/01/2025 at 10:03 UTC
8 upvotes, 0 direct replies
Similar stimuli already exist. There's a pattern of stripes that causes color inversion in the visual cortex if you stare at it for too long, in one case for as long as several *weeks*.