💾 Archived View for idiomdrottning.org › ml captured on 2023-04-26 at 12:55:18. Gemini links have been rewritten to link to archived content
⬅️ Previous capture (2023-04-19)
➡️ Next capture (2023-05-24)
-=-=-=-=-=-=-
I hate machine learning so much, to the point that I’m almost willing to consider bad arguments against them! But ideally let’s stick to good arguments against them, and see how far we can get.
I started hating ML from a gut feeling first, and had to come up with arguments and reasons later. Feels over reals. I think most people work this way without being aware of it. What the thinker thinks, the prover will prove.
Bad argument. I thought we hated copyright. Extending the definition of copyright to not only including distributing works themselves but also learning from looking at the work is beyond twisted.
Short-term argument. The models of today are not fully hooked up yet. They’re UI demos. I mean, yes, they do suck, all AI art is really ugly and all AI texts is really dumb and all AI-programmed apps are really bad and buggy and AI is hallucinating more than Wikipedia.
Presently.
But if they stay this bad, we won’t need any arguments because they’ll collapse by themselves in dot-com 2.0, and if they become good, this argument isn’t gonna last us far.
Super bad argument. Education adapted to cameras and calculators. If it weren’t for these other arguments against ML, education could adapt to it as a new tool, too.
They make us question what being human really is. So did cameras and movie cameras, but, that’s not necessarily a particularly fun fire to play with. As dismissive as I am of concerns about their use in education, I take this part seriously. The art of humanity struggled with those inventions and barely made it out on the other side OK. What is our purpose once we’ve automated every little dream away?
Pretty good argument, but we eventually did manage to survive cameras, and artists could use them, leverage them. Arguably the birth of realist painting and the enlightenment was due to the invention of optical devices and their adaptation by painters like da Vinci and Vermeer.
Personally, yeah, it felt like a super gut punch. Compared to other artists I felt like my journey towards becoming good took a lot longer than most, but I didn’t give up and I finally got there and became able to draw and paint well. I spent 20 years on what would’ve taken many other artists a few years (they’d either get good or give up). And then within one year of me finally feeling like I know how to do it, along comes Dall-E. FML.
A.k.a. bots will take our jobs.
Obviously, research regarding technological unemployment is as vital today as further refinement or production of labor-saving and comfort-giving devices. Unless we radically and fundamentally transform distribution of resources and labor, ML is gonna make it so that the owner class is gonna own even more and the wealth gaps are gonna widen. This is a good argument when we speak among each other and planning to storm the palace. As far as the owner class themselves are concerned, I’m pretty sure they’ve chalked this one up in the “pro” column rather than as a true danger of ML.
This is an issue I’ve been thinking about for all my life. Since I was like six years old. It’s what steered me leftward, too. I was like “OK if bots take over our jobs that’s gonna be good because we won’t have to work but it’s gonna be bad because how will we get money? Some people will own the bots but not everyone can do that and then people will starve and it’ll be a huge disaster.” Damn it. I wish I had been wrong about that.
It’s an argument that’s not entirely clear-cut because if it weren’t for market capitalism’s little exploitation issue, this would’ve been a good thing rather than a bad. Y’all know how much I hate the human compiler at work, hate having to do boring chores by hand that a bot could do.
Surveillance, exploitation, a powerful tool in the hands of the mighty. Yeah. A good argument. New technology has historically facilitated theretofore unimaginable horrors:
The fake news is gonna get faker, and that includes fake news by state actors like Trump and Putin, and their dismissal of real news by labeling it as “fake”. Good argument, but just like the previous two arguments, talking isn’t gonna do much in the face of malicious actors. These tools are a wild chainsaw swung chaotically and they’re a precise scalpel wielded ruthlessly. Once they’re in the wrong hands it might be too late to talk about it.
If the argument is “the models are gonna come alive and feel bad”, then no. Not good argument. If the argument is “fools are going to mistakenly believe that the models are alive and give them too much responsibility or respect” then maybe. Yeah. Yes, I am afraid of that. That was my first fear; although arguably many of the other arguments I list here are even bigger issues.
Very good argument. Training these models use climate-wreckingly large amount of resources. Probably the best argument. The most legitimate reason to regulate them (even Hayek realizes that externalities need regulation) and the biggest drawback with them. It’s a good argument because it appeals to all quarters, and it’s true, and it’s the one of the biggest reasons to hate them.
This is something we absolutely need to fix before going forward. Climate change is urgent, real, and it’s here now. People who dismiss this issue need to sober up or they’ll have the friendliest chatbot on the cinder.
IBM and the Holocaust - Wikipedia
Computing Power—I guess I was saying these same things six years ago