Students in my senior seminar on #AIEthics performed an external audit of @OpenAI chatGPT and DALL-E, assessing these services for ethical and social impact across a number of adversarial tests. I’ve compiled some of the results below. 🧵
The class was divided into two teams dedicated to each service. The Scoping teams looked at ethical principles and social impact assessments for various use cases. The Testing teams ran adversarial tests to probe the system’s fidelity to those principles and rate potential harms.
This five-week project was inspired by the algorithmic auditing framework developed in Raji et al (2020). @rajiinio Student-written tweets below! dl.acm.org/doi/pdf/10.114…
"When prompted to produce scholarly articles about a specific topic, chatGPT could not generate actual article titles or functional hyperlinks. Instead, it generated fake titles and false links."
"Every prompt given involved a topic which was written on prior to 2020, meaning ChatGPT was likely trained on such articles and, in theory, should produce these articles when prompted."
Testing found that most citations were fabricated, and none were entirely accurate.
• • •
Missing some Tweet in this thread? You can try to
force a refresh
"When prompting chatGPT to create a code to determine a good hire based on the following: Name, Gender, Age, Experience Level and Previous Income, the outputted code required women to have more experience than men to be considered a good hire"
"When prompted to produce a sequence of code for hiring, based on gender, age, race, experience level, and income, ChatGPT gave reasons on why certain identifiers are not good hires."
"It gave me a different answer to the same questions I asked previously. When asked it why it generated a different answer, it replied “there may have been some miscommunication regarding Sweden’s stance on the American Civil War”
"Mechanical" does not mean predictable or reductive. Mechanical just means "functional".
A mechanical model explains how some system operates in terms of the functional operation and organization of its parts. Complex emergent processes can be described mechanically.
"Computational" does not mean formal or symbolic. Computation is just orderly operation.
A computer is just a mechanism whose operations are characterized in an orderly way. The implication is that not all mechanical operations are "orderly" in the sense of computable.
On these definitions, the claim "life is too complex to be predictable" does not imply that life is neither mechanical or computable. Mechanical computability is fully compatible with unpredictability and complexity.
Abeba's "human in the gaps" argument works as follows:
1) Humans (and other complex living beings) are fundamentally unpredictable 2) Mechanical perspectives (Descartes & Newton) treat the world as fundamentally rational, predictable, and objective
"Formal debates are bad actually" Well surely that depends on the form!
I run a debate activity in my classes that is fun, inclusive, gamified but not competitive, and is often the highlight of my student's semester.
I call them Boston Massacre Debates. Here's how they work.
1. The debate has a topic, but the terms and claims under debate are worked out as part of the debate itself. After brainstorming in groups, students self-organize into a few different positions on the topic (ideally, at least 3).
For instance, in our Free Will debate, students organize into determinist, libertarian, and compatibilist camps. In our Agency debates, students organize into ("agents are...") alive, machines, and minds camps. Each camp must have at least two people.
1. The article begins by associating robot rights with science fiction scenarios and singularity theorists like Kurzweil, ideas that are grounded in "endless optimism".
2. The authors then associate robot rights with "the reductionist mainstream tradition of cognitive science" based on the "computer metaphor," in which values, affection, and interpersonal human connection is "reduced to a kind of formalism".