Comment by ACCount82 on 29/01/2025 at 22:01 UTC

10 upvotes, 1 direct replies (showing 1)

View submission: SEEK 2.4.B - SEND

View parent comment

In real life, if you have near perfect knowledge about how a given AI works, you can use that knowledge to calculate a crafted abnormal input that breaks it.

This works on many, many AI types - including modern LLMs. You could craft an attack that makes an AI output gibberish, or one that cuts through any prior instructions and makes it carry out a certain command.

Could a similar attack be performed against a human brain - if something understood the functioning of a human brain almost perfectly? I'm inclined to think so.

Replies

Comment by HeroOfOldIron at 30/01/2025 at 10:03 UTC

8 upvotes, 0 direct replies

Similar stimuli already exist. There's a pattern of stripes that causes color inversion in the visual cortex if you stare at it for too long, in one case for as long as several *weeks*.