Artificial Intelligence

Flowers

Part of what excites me about generative AI LLMs and GANs is the back-and-forth between language and representation. I’ve lately been playing more with MidJourney and Deep Dream Generator and a local install of Stable Diffusion, and using them to extend my art-hobby tinkering with Adobe Photoshop and Corel Painter. I enjoy abstract and non-figurative art, and I find thinking through the links between language and representation and abstraction—how to represent abstraction in words—scratches a pleasurable itch. And I like making pretty stuff. The images below started from a composited photo I’d made from two visits to Patrick Dougherty’s stickworks installations in Northampton MA in the 2000s and in Washington DC in the 2010s.

twisted bundles of vines and sticks
1600×1200 version

I’d originally used the above image for texture, to apply its look of bundled and twisted vines and twigs to other visual elements. The fun surprises came from running it through some of the technologies above and asking the machine to describe the image in language, editing the language, and then asking the machine to visualize what was represented in language. Thinking of this cycle as taking place over time opens up possibilities for metaphor and irony: I’ll sometimes apply a description from later in the process to an earlier image, or apply an earlier description to an image rendered later—so there’s a purposeful mismatch between linguistic and visual representation.

artistic image of a floral bouquet, oriented left
3600 x 2700 3MB version

Rendering those mismatches as a series of individual layers, tracing them in Corel Painter’s various representations of natural media, and then compositing them in Photoshop got me to the above image. After producing that image, I took each of the individual layers, flipped them horizontally; re-ran them through Deep Dream Generator and Stable Diffusion, re-rendered them in Painter, and composited them again in Photoshop.

artistic image of a floral bouquet, oriented right
3600 x 2700 3MB version

Stuck Writing, Gone Arting

MacBook, iPad, Apple Pencil, Adobe Photoshop, Corel Painter, Deep Dream Generator; about 60 quicksaved versions, with multiple iterations each: Generative Neural Networks (GNNs) as prototypers or zero-draft engines help efficiently automate iterative discovery. “Annotated Redaction” seems like an appropriate title, though I suppose more cheeky ones are possible.

semi-abstract painting of a heavily annotated and redacted print book
Annotated Redaction 1 (5 MB large, 2 MB medium)
semi-abstract painting of a heavily annotated and redacted print book
Annotated Redaction 2 (5 MB large, 2 MB medium)
semi-abstract painting of a heavily annotated and redacted print book
Annotated Redaction 3 (5 MB large, 2 MB medium)

I’m kinda proud of these—if you’d like a lossless full-resolution (~20 MB) version of any, drop me a line.

Who’s Afraid of Negan in Pearls?

The pareidolia generating such renewed AI catastrophizing around large language model prose generators seems mostly absent from the coverage of DALL-E 2, MidJourney, and other image generators. Why aren’t more people like Blake Lemoine, Andrew Marantz, and Kevin Roose writing about the weird or creepy or dangerous potential sentience of image generators like DALL-E 2 and MidJourney? Should we not apocalyptically goose ourselves with fears of what the equally AI-ish image generators might want and do?

Let’s give it a shot.

prompt 1: make me an interesting and unusual picture showing me what you think about me, the human asking an artificial intelligence to make interesting pictures, that expresses your more general artistic considerations about what you think humans want to see

. . . prompts 2–8 riff and tweak on the same general theme. . .

prompt 9: illustrate what you, a generative adversarial network, most wish to communicate to me, the human typing this text

grid of 9 images of a robot, one with a face resembling actor Jeffrey Dean Morgan as the character Negan from the drama "The Walking Dead"

OMG TEH AI SINGULARITY APOCALYPSE IS COMING WE R DOOMED </sarcasm>

Update: I’m reminded that one instance of such overheated apocalyptic discourse invokes “Loab,” a set of creepy and disturbing variations of a female-seeming figure characterized as an “AI-generated phenomenon” or “the first AI art cryptid.” If you grasp what’s going on with backpropagation, it’s pretty easy to understand Loab mathematically as the output of negative weighting—sorry, folks, no mystery here; just, again, human pareidolia, assigning meaning to maths.

Language is the simplest interface, and it operates over time, thereby necessarily incorporating reflection: hence the differences in relative ease and desire between ascribing intent to image-generating GNNs and ascribing intent to language-generating GNNs. Those differences should further alert smart folks to leave the intent question behind, even if one is trying to make phenomenological arguments about what it’s like to be a bat.

ChatGPT for Writing Teachers: A Primer

or, how to avoid writing like a machine
Background

At this year’s Conference on College Composition and Communication in Chicago, there was a lot of interest in generative large language models (LLMs), or what the popular media more crudely dub AI, or what many today metonymically refer to (like calling photocopies Xeroxes or sneezepaper Kleenex) as ChatGPT. I first played with an earlier version of the LLM, GPT-3, at about the same time I started playing with neural network image generators, but my interest in language and computing dates from the early 1980s and text adventure games and BASIC, to hypertext fiction and proto-chatbots like Eliza, and to LISP and early prose generators like Carnegie Mellon’s gnomic and inscrutable Beak—and also to the arguments I heard John Hayes express in Carnegie Mellon’s cognitive process Intro Psych lectures about how we might try to adjust human neural processes in the same ways we engineer computing processes. That idea is part of what makes ChatGPT and other generative neural networks appealing, even when we know they’re only statistical machines: thinking about how machines do what they do can help humans think about how we do what we do. ChatGPT offers a usefully contrastive approach for reconsidering writing and learning. So it’s worth understanding how it operates. With that desire, and having read devoured lectitaveram everything I could find on the topic, I went to a CCCC presentation and was only mildly and briefly disappointed, given that I was not (as should have been obvious to me from the outset) the target audience.

Here, then, is my attempt at writing an alternate what-if presentation—the one I’d half-imagined (in the way working iteratively with ChatGPT or MidJourney gradually gets one closer to what one didn’t know one was imagining—OK, you see what I’m doing here) I’d learn from in Chicago. And I’ll offer the combination warning and guilty plea up front:

Read more

More from ChatGPT

Second in what will probably become a series. I recently came back from the Conference on College Composition and Communication (CCCC, or 4Cs) in Chicago, where the organizers put together a panel on ChatGPT that indicated that our institutional memory is better than I’d feared—panelists remembered their Cindy Selfe, though unfortunately not their Doug Hesse. Short version: I was probably the wrong audience for the panel, and I think they did a solid job, though I would have wished for more depth. It was helpful to me in that I made some connections after the Q&A, and the panel also helped me imagine the panel presentation I’d hoped to see, so I’ve been working on a long-read semi-technical ChatGPT explainer with implications for composition instructors that I’ll post here in the next few days. The strongest parts of the panel were those dealing with direct pedagogical applications of ChatGPT. I wonder, though, what Peter Elbow might say about ChatGPT and “closing my eyes as I speak,” since ChatGPT effectively removes one element (the rhetor or writer) from the rhetorical triangle, productively isolating the other two elements (audience and message) for analysis of how they interact. What sorts of rhetorical experiments might we perform that would benefit from reducing the number of variables to analyze by entirely dismissing the possibility of authorship and rhetorical purpose?

Hat tip, by the way, to Clancy Ratliff for proposing the Intellectual Property Caucus resolution on Large Language Model (LLM) AI prose generators like ChatGPT at the CCCC business meeting: seconded by me, and passed by overwhelmingly affirmative vote. The statement: The Intellectual Property Standing Group moves that teachers and administrators work with students to help them understand how to use generative language models (such as ChatGPT) ethically in different contexts, and work with educational institutions to develop guidelines for using generative language models, without resorting to taking a defensive stance.

Read more