Class 14 (English 197 – Spring 2024)

This is the main course website. There is also a course Canvas site for uploading assignments.

Class Business

Epigraphs to Frame the Discussion

Nicholas Thompson & Geoffrey Hinton (An AI Pioneer Explains the Evolution of Neural Networks,” 2019)

Nicholas Thompson: Explain what neural networks are. Explain the original insight.

Geoffrey Hinton: You have relatively simple processing elements that are very loosely models of neurons. They have connections coming in, each connection has a weight on it, and that weight can be changed through learning. And what a neuron does is take the activities on the connections times the weights, adds them all up, and then decides whether to send an output. If it gets a big enough sum, it sends an output. If the sum is negative, it doesn’t send anything. That’s about it. And all you have to do is just wire up a gazillion of those with a gazillion squared weights, and just figure out how to change the weights, and it’ll do anything. It’s just a question of how you change the weights.

NT: When did you come to understand that this was an approximate representation of how the brain works?

GH: Oh, it was always designed as that. It was designed to be like how the brain works.

Fleuron icon (small)

Emily M. Bender, et al., “On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?” (2021)

Contrary to how it may seem when we observe its output, an LM is a system for haphazardly stitching together sequences of linguistic forms it has observed in its vast training data, according to probabilistic information about how they combine, but without any reference to meaning: a stochastic parrot. (¶ 6.1)

Fleuron icon (small)

Albert Einstein (responding in 1926 to quantum mechanics)

The theory produces a good deal but hardly brings us closer to the secret of the Old One. I am at all events convinced that He [God] does not play dice.

From Word Embedding to Sentence Embedding

Neural Networks (& Word Embedding)

Deep Learning Neural Networks

Large Language Models (LLMs)

Text-to-Image Large Models (LLMs)

Interpretability & Explainabiiity of LLMs

Thinking With / Thinking about Large Language Models

Minh Hua and Rita Raley“Playing With Unicorns: AI Dungeon and Citizen NLP” (2020)

If a complete mode of understanding is as-yet unachievable, then evaluation is the next best thing…. (link)

In this endeavor, the General Language Understanding Evaluation benchmark (GLUE), a widely-adopted collection of nine datasets designed to assess a language model’s skills on elementary language operations, remains the standard for the evaluation of GPT-2 and similar transfer learning models…. Especially striking, and central to our analysis, are two points: a model’s performance on GLUE is binary (it either succeeds in the task or it does not)…. But if the training corpus is not univocal — if there is no single voice or style, which is to say no single benchmark — because of its massive size, it is as yet unclear how best to score the model. (link)

Our research questions, then, are these: by what means, with what critical toolbox or with which metrics, can AID [AI Dungeon], as a paradigmatic computational artifact, be qualitatively assessed, and which communities of evaluators ought to be involved in the process? (link)

AID, as an experiment with GPT-2, provides a model for how humanists might more meaningfully and synergistically contribute to the project of qualitative assessment going forward…. (link)

Our presupposition … is that it is not by itself sufficient to bring to bear on the textual output of a machine learning system the apparatus of critical judgment as it has been honed over centuries in relation to language art as a putatively human practice. What is striking even now is the extent to which humanistic evaluation in the domain of language generation is situated as a Turing decision: bot or not(link)

Fleuron icon (small)

This is the main course website. There is also a course Canvas site for uploading assignments.