đŸ Archived View for gemini.ctrl-c.club âș ~ssb22 âș misread.gmi captured on 2024-08-31 at 13:07:45. Gemini links have been rewritten to link to archived content
âŹ ïž Previous capture (2024-08-18)
-=-=-=-=-=-=-
Sometimes Iâve come across a bit of real-life interpersonal relationships drama in which one person says of another, âthat person gave me a funny lookâthey must hate meâ.
Although a person saying this may have been destabilised by other factors (high-stress medical conditions or whatever) and may have immediate emotional needs that are more for sympathy than logic, there is a part of me that wants to avoid the possibility of reinforcing a potentially harmful misunderstanding, and respond along the lines of:
What if you misread them?âWhat if what they were thinking of wasnât you at all?âCanât we at least apply Hanlonâs Razor here?
to which the frustrated individual on the receiving end of the âlookâ might reply something like:
What do *you* know about facial expressions?âYou have a brain condition that stops you from being able to process them!âI *know* Iâve read that person correctly, and youâre in no position to suggest otherwise.
Itâs correct that I basically canât process facial expressions myself, but I *do* happen to have scientific evidence that ânormalâ people arenât always right about them.
When I was a PhD student, I sat two desks away from a specialist who knows *everything* about facial expressions.âShe invented a computer system to read them, and went on to turn that into a successful AI company.âYou might have read her book Girl Decoded.âAnd while Rana el Kaliouby was primarily focused on *computer* readings of emotions, she necessarily had to look at the way *humans* read them tooâwhich brings me to a very interesting number she found in one of her experiments.
Rana has improved her emotion-reading system over the years: her early versions were not as good as the modern one.â(I say this because I want it to be clear that the numbers Iâm about to drag up from her early days are *not* representative of the modern commercial system.)âI distinctly remember one time when Rana was testing an *early* version of her system on a certain set of videos, and its accuracy turned out to be 77.4%. Although getting it right nearly four times out of five was definitely impressive, it didnât seem *super* impressive just yet, and Rana wanted to compare it with something else to get some idea of where she stood.
Well there *wasnât* another computer system that could try to read the complex emotions in those videosâthere might have been a couple of basic systems that could tell âhappinessâ from âsadnessâ, but Rana was the first one to try to distinguish between things like âconfusionâ and âconcentrationâ.âSo the only way sheâd be able to get a meaningful comparison figure with which to evaluate her system, was to run an experiment asking *humans* to name the emotions and see how well *they* did.
Of course Rana picked fully-sighted neurotypical subjects with no known medical issues that affected the reading of facial expressions, and there was me sitting in the corner thinking âoh, those sighties will get 98% for sureâ.
On the basic test, they got 71%.
They were wrong nearly three times in ten, just looking at *basic* emotions.
That person who gave you a funny look?âCould be a 30% chance your brain was misreading it.
And then, when those volunteers were tested on more *complex* emotional states, their accuracy fell to 51.5%.
This experiment says there could be a 50-50 chance your brain has misread that subtly funny look the person gave you.
(Iâm not leaking unpublished results here.âItâs all in Ranaâs thesis.âNaturally she doesnât make it a *major* point, because her thesis is focused on the automatic system, but the human results *are* in there.âDownload Ranaâs thesis and search for Figure 3.4.)
We *cannot* jump to conclusions about a personâs relationship with us from the way our brain reacted to just one look, even if we know that look was reacting to *us* (which we donât know for sure).âWe need more data than that.âSure, we might not want to hang around to collect said more data if weâre likely to be in a one-off confrontational situation with a stranger, but if itâs a person youâll meet again, then please donât write them off *just* because of a lookâyou might have read it correctly, but you also might not have, so it might be worth giving your hypothesis an extra test or two, just to make sure.
Now, we *could* argue about these percentages.âAlthough Rana conducted the experiment rigorously, it *was* only one experiment and could probably do with being repeated at a larger scaleâafter all, Rana wasnât exactly focusing her PhD on how good *humans* are at reading emotions; that was just an incidental figure she needed as a baseline for evaluating her computer system.âNevertheless, it is strong evidence that there exist at least some circumstances in which ânormalâ people can fail to read emotions correctly between 30% and 50% of the time.
Can I take a âcheap shotâ at Kate Crawfordâs book *Atlas of AI*âCrawford was right to point out the huge hidden resource consumption of some âBig Techâ (not just limited to âAIâ) and the dangers of people having âautomation biasâ (thinking the computer is better than it really isâIâm not sure âAIâ should have been called âintelligenceâ unless you limit that to pure âcrystallisedâ intelligence in Cattellâs model), but when she had a go at Ranaâs brainchild she went too far.âYes I may be biased: I shared an office with Rana and encouraged her in her quest, saying her system could be a great help to people who have trouble reading emotions themselves.âBut I do have an actual argument here; Iâm not *just* being pro-Rana because I was one of her âlab matesâ.
In fairness to Kate Crawford, I donât think she *meant* to do a âhatchet jobâ on Rana.âI mean, if youâre writing a book called âAtlas ofâ some problem, then it can understandably make you feel pressure to dig up a bit of âdirtâ on as many aspects of it as possible.âSomething exists called affective computing?âQuick, find something shaky about itâOK, got something, so letâs dismiss that field and move on to the next because this publisher deadline wonât meet itself.âThat being the case, Iâm only having a go at the book, not the author who might have done better if she hadnât been under pressure at the time.
So the objection to the whole of affective computing that was raised in *Atlas of AI* went back to Ranaâs early experiments testing both her system and her colleagues on a set of videos that was originally compiled by Professor Simon Baron-Cohen (the one who came up with the âempathisingâsystemisingâ theory, with men tending to lean a little toward the âsystemisingâ solve-the-problem side and women tending to lean a little toward the âempathisingâ understand-the-person side, and autism being an arrangement of the brain thatâs extremely good at systemising but has more trouble with empathising, although that *doesnât* mean autistic people donât *have* empathy, they just might need a clearer explanation of the situation first)âBaron-Cohen compiled a set of videos for an experiment he wanted to run to find out if autistic people can be trained to recognise emotions (which is a fair enough question, whatever its answer turned out to be), and, in compiling that set of videos (which was also the one used by Rana in her tests), Baron-Cohen did employ some actors.
Actors!âOh no!âEverybody panic and run because the whole field that Rana invented is resting on the wobbly foundation of a set of videos made by *actors*!
Except, of course, thatâs not the whole story.
Firstly, those were early days.âRana and her company have done *much* more work since then, working with *millions* of natural spontaneous emotions recorded from people who gave their consent.âThe state of Ranaâs prototype in 2003 is not at all relevant to the modern system.âEverybodyâs got to start *somewhere* and itâs been effectively remade since then.
Secondlyâand more relevant to our interpretation of that small-scale experiment with human emotion readersâalthough some actors can be bad, other actors can be good, and the *best* actors have learned the skill of adopting a temporary mindset that they *are* the character, facing the situation *for real*âor at least have the ability to re-live real incidents from their own lives that trigger the required emotions for real, which is one reason why being a top actor can be such a high-stress job.
True, itâs safe to say Baron-Cohen wouldnât have been able to afford to employ actors with top track records like Patrick Stewart, nor did he clip pre-confirmed footage from existing blockbuster films (there *might* be legal arguments about copyright exceptions for certain uses in certain situations, but he probably preferred to avoid the possibility of getting that wrong and simply create new material from scratch), but if you employ enough early-career actors you should eventually find unrecognised âtalentâ among themâand each of the videos in that set was run past a panel of ten judges, and kept only if at least eight of the judges said it was a good performance of the emotion it was supposed to be.
Which basically means that if you take a set of performances that is misread at most 20% of the time by one set of people, then it can be misread 30% to 50% of the time by another set of people, when the two reading sessions were done under slightly different circumstances.
As I said, this was an early-days baseline test for getting Ranaâs automatic system off the ground: it could do with a larger scale, more controlled repeat.âBut as preliminary data itâs still enough to cast doubt on the idea that the ânormalâ human brain is always reliable at reading the emotions of others.âIt is not.
All material © Silas S. Brown unless otherwise stated.