AI & ChatGPT: The Bullshit Generator, Class Wars and Why Do We Even Bother?
The Wonderful World of AI Built on Human Suffering.
Top of Mind This Week: AI & ChatGPT
Just being dramatic with an image I generated using Midjourney with the prompt “AI overlord standing on a platform of crushed human skulls”.
Talofa reader,
I went down a rabbit hole of ChatGPT and AI this week. It’s been a hot topic for a short while now since ChatGPT’s release in November, 2022 and there’s no shortage of articles about it in my Readwise feed.
There’s been a lot of takes, some pro, some con. My friend DZ’s newsletter got me thinking about the darker side of AI and so I thought I’d read up on it this past week and write some stuff down.
It’s a bit of a read so if you don’t mind a bit of profanity and bad punctuation, enjoy.
"The standard you walk past is the one you accept" -- Lt. General David Morrison.
What is ChatGPT & AI Reaaaally?
Let’s start with what we know about ChatGPT.
ChatGPT is an AI chatbot developed by OpenAI and built on top of OpenAI’s GPT3 family of “Large Language Models” (LLM) and fine tuned using supervised and reinforcement learning techniques1.
The GPT stands for “Generative Pre-trained Transformer”. The Generative part means it generates human-like text. It’s Pre-trained on a massive amount of text data (publicly available data from the internet for example) and uses the “Transformer” deep learning model that uses mechanisms to pay attention to and weigh the significance of the words in the data it’s processing.
AI: But How Intelligent?
Maybe it’s just me but it took looking further into what “intelligence” actually means here in the context of AI to understand that’s it’s not T1000 about to suddenly figure out how to chase us down and kill us all with multiple martial arts techniques it has just watched on TV.
AI is trained on a very specific thing, so its deep on that one thing, and pretty useless at anything outside of it e.g. an AI beat a chess grand master but would fail a basic maths test because it has no idea what you’re talking about i.e. it hasn’t been trained on maths.
So What’s AGI?
This is a super oversimplified take on the difference, but I think it’s important to know that there are two different “AI” that people talk about, and will sometimes confuse and conflate when talking about the thing that’s going to take over the world and kill us vs. the one that’ll beat us at chess.
AGI, thanks to Wikipedia, is:
Artificial general intelligence (AGI) is the ability of an intelligent agent to understand or learn any intellectual task that human beings or other animals can.[1][2]
So, AI will only know how to kill us if we train it accordingly. But AGI will figure out how to kill us all on its own.
AGI aside, how are we training these AI models?
How Are These AI Being Trained?
You could be forgiven for thinking the AI world is a super slick computer enhanced synthesis of automation and maths wizardry efficiency.
It’s not.
OpenAI’s blog tell us:
We trained an initial model using supervised fine-tuning: human AI trainers provided conversations in which they played both sides—the user and an AI assistant. We gave the trainers access to model-written suggestions to help them compose their responses.
Ok, so humans help train these models.. tell me more
To create a reward model for reinforcement learning, we needed to collect comparison data, which consisted of two or more model responses ranked by quality. To collect this data, we took conversations that AI trainers had with the chatbot.
Cool. More humans, doing the fine-tuning.
You can see humans in Steps 1 & 2 of the OpenAI diagram (Step 3 is what the marketing hype sounds like.. “what humans?!”)
So what? People are being employed to provide training to AI models.
Sure, let’s have a look at these workers…
Humans of AI Development
Following my nose and reading through the lesser known and talked about (at least on my mainstream social media) articles on the negative aspects involving AI has been educational and unsurprising.
None of this negative shit is new. The powerful exploit the less powerful. It’s like a law of nature at this point.
In 2017, Mary L. Gray and Siddharth Suri wrote ‘The Humans Working Behind the AI Curtain’ citing Facebook’s supposedly "unbiased algorithm" that turned out to be powered by regular humans (you know, the kind susceptible to "bias"). Imagine my surprise to find that, in 2022, Chloe Xiang's piece ‘AI Isn’t Artificial or Intelligent’ suggests we're pretty much in the same position, powering our AI innovation with underpaid workers in foreign countries, or by the better-known term, the "Global South".
In 2016, it was people like middle-aged mother of two Kala, sitting on a computer in Bangalore, India, looking through NSFW content for the likes of Google, Facebook, and Twitter. In 2022, we've either expanded or just relocated the "crucial contributions" work to Kenya, where we read about it in Time's 'Inside Facebook's African Sweatshop' in February, with Vice bringing it home in Jan 2023 with an article detailing how ‘OpenAI Used Kenyan Workers Making $2 an Hour to Filter Traumatic Content From ChatGPT’.
But this is all OK. You know why? Cos “impact sourcing” /s
“Impact Sourcing”
Now I’ve heard some bullshit in my time when it comes to tech marketing, or corporate social responsibility marketing and branding but learning this term “impact sourcing” in the context of this AI training work has been an absolute stunner.
What is “impact sourcing”?
Impact sourcing, also known as socially responsible outsourcing, refers to an arm of the business process outsourcing (BPO) industry. It employs people at the base of the pyramid or socioeconomically disadvantaged individuals as principal workers in BPO centers to provide high-quality, information-based services to domestic and international clients.
You could only seriously believe you are bettering the lives of these workers if you believe the PTSD from sorting through NSFW text and images every day, for $2 an hour, is an improvement on their lives.
And I guess that’s the point, we believe these people are beneath us and our AI goals.
Not necessarily that we literally think the thought “we are better than them”, but that our actions will amount to “they are not important enough for us to change our behaviour that’s complicit in this outcome”.
Which brings us to…
The Classic “Us” vs “Them”
It’s just a matter of distinguishing which “us” is “us” this time. Poor? Non-white? A little from column A, a little from column B? A Union B?
The Class War.
We all know about the inherent biases in our technology. We’ve known about it for a while now.
We've long talked about and identified how the lack of representation and input at the "tables" of those who develop and build our tech result in things like cameras that can't pick up black skin, or medical school computer program inviting applicants for interviews based on gender and non-European names.
AI brings a whole other dimension to that. Dan McQuillan in his piece 'We Come to Bury ChatGPT, Not to Praise It.' 2 talks about how OpenAI CEO Sam Altman thinks of people as 'stochastic parrots' - meaning people are just large language models with learned patterns of behaviour and nothing more (yes, paraphrasing and simplifying). It's an idea seemingly only applied to us ordinary folk, and the "elites" of our society are apparently something better. So of course, us "ordinary folk" are perfect fodder to feed the AI machine meant to serve the upper class with free training labour for Google’s reCAPTCHA to train their models.
Digital Colonialism.
I'm introduced to the idea of "Digital Colonialism" in Arvind Narayanan's article ‘Digital inequalities will power digital colonialism’. The idea that people are treated differently based on what country they're from is pretty standard colonialism.
We can do it digitally now.
When India was used as a testing ground for a Bing chatbot and users complained to Microsoft about the chatbots abusive behaviour, nothing was done. But when the issues made it to The New York Times, Microsoft made changes within days. That's digital colonialism.
But the example he uses that really brings this idea home is of Sam Altman, CEO of OpenAI (which makes ChatGPT), talking about how people who can't afford healthcare, can use a really smart chatbot. The fact they see can't see anything wrong with this - that rich people have doctors, poor people have chatbots - is mind-blowing but not surprising.
Why Am I Calling It a ‘Bullshit Generator’?
I got this phrase from Dan McQuillans no-holds-barred piece ‘We come to bury ChatGPT, not to praise it.’ , which I thought was brilliant.
Dan says:
"ChatGPT is, in technical terms, a 'bullshit generator'. If a generated sentence makes sense to you, the reader, it means the mathematical model has made sufficiently good guess to pass your sense-making filter. The language model has no idea what it's talking about because it has no idea about anything at all. It's more of a bullshitter than the most egregious egoist you'll ever meet, producing baseless assertions with unfailing confidence because that's what it's designed to do."
Brilliant because he calls out the uglier realities behind the technology and also the kinds of people and ideas who benefit from such technology and how they count as nothing, "the immediate vulnerability of millions of ordinary people" exploited for AI's gain.
I’m pretty sure I could’ve just summed up this whole newsletter with this paragraph. But this is a learning exercise, hence why all the notes 😁.
What Can We Do about this?
Unlike AI, I’m okay with saying “I don’t know” the answer to a question.
Because I don’t.
Arvind Narayanan lists a few positives about ChatGPT in ‘ChatGPT is a bullshit generator. But it can still be amazingly useful’ where it can help in context where the truth or accuracy isn’t important (fiction) and debugging code. But then notes a study done that found “Copilot generated insecure code 40% of the time.”.
Awesome 😂.
I think of 'what can we do?' in this context - i.e. of the big AI machine and its impending impacts on the world, as well as its current impact on the humans being fed to its development - the same way I look at individual recycling to clean up the world's pollution and climate issues: pretty useless at the individual level, and would instead need some sort of massive groundswell force of public opinion and civic action to force the powerful to change the situation.
I don’t see that happening.
Abstaining from using these tools won’t change anything the same way vegans and the plant-based meat industry haven’t change our meat eating habits.
So, in my humble opinion, learn how AI really works, how it will impact you, your job, and your community. Teach what you know - actually know, stop just reading headlines and parroting the marketing bullshit hype - and read widely, discuss, and share with your community.
My “always” take on new tech and it’s almost guaranteed potential for harm to people is the same as in professional fighting - “protect yourself at all times”.
Thanks for reading. I'll see you in the next episode.
Learning
Things I’m actively studying or learning this week…
Studying for the ‘AWS Certified Security - Speciality’ certificate.
Building
Things I’m building or working on this week…
reviving a recent side-project on a sentiment-analysis pipeline for YouTube channels: https://github.com/ronamosa/aws-youtube-analyze-arch
Updating myblogto be searchable and add more docs.
Interesting Reads
Articles or other writing that stood out to me this week…
‘AI Search Engines And The Quest For Ignorance’ - by Marcus Hutchins
‘Have I Been Trained’ — tool for artists to check if their artwork has been used to train AI models like Stable Diffusion and flag them for removal i.e. opt-out.
Community
Other projects in community I’m working on…
Pasifika Tech Education Charity - Providing Tech Learning Opportunities for the Pasifika Community.
Pasifika Tech Network - A Network for Pasifika Tech Professionals & Learners.