Discussion about this post

User's avatar
Ken Muldrew's avatar

"LLMs are, as Ted Chiang pithily put it, blurry JPEGs of human knowledge - that is, they are tools for generating lossy summarizations of large bodies of human knowledge."

A better metaphor might be that LLMs are microscopic images at, and below, the diffraction limit. Strictly speaking, one is skating on very thin ice if they take any object in a microscopic image that is below the diffraction limit as real (as opposed to an artifact). However, as video microscopy gained purchase in the 80s and 90s, it became clear that one could use motion to discriminate real objects from artifacts. Much as analyticity allows you to assume a type of continuity for a function, observed motion that obeys (or appears to be consistent with) physical law allows you to assume some smudge in the field of view is a vesicle on a kinesin motor running along a microtubule rather than just a typical smudge-artifact on a static field of view. I'm grossly simplifying, but it is possible to see objects that are below the diffraction limit using video microscopy that would be dismissed as mere artifacts on a static image. It's blurry like an overly compressed jpeg but there is also data that can be extracted by using clever technique even though it appears not to be there at all.

The trick with deep conv nets is that the multiple layers allow sequential correlations, rather than static correlations, to give one an opportunity to infer a causal process (like physical motion in the microscopic field). Unfortunately, the "one" who is given that opportunity is a computer program, so it doesn't really have the experience of cause-and-effect, built up through years of manipulating the environment. Thus the program will often infer causation through sequential correlations that are "unphysical" and thus spit out an extrapolation that is ridiculous to a human (so-called hallucinations). If there was some way for an LLM to back out the series of correlations that led to a particular result, when asked to explain itself, we might learn something about the kind of "unphysical" correlations that cause these errors.

Expand full comment
Christine M. Nowik, Ph.D.'s avatar

Not to be That Guy, but on the matter of the syllabus, I've found good luck with starting the term with instruction on annotation and then applying our skills to the syllabus. Annotation and close reading are skills for any discipline. I just so happen to teach writing, so I layer in genre discussion. Part of the "read the syllabus" problem is that the syllabus is an incredibly complex rhetorical situation because of the multiple audiences (students, bureaucrats, other institutions, courts). I require students to annotate with both a comment/reaction as a reader (another layered lesson since we focus on audience in our work together) and a question a reasonable person might have (my effort to not only clarify matters on the syllabus but also to normalize questioning/not knowing). We engage in this activity via a shared doc in any course I teach.

Results of this effort include a steep reduction in one-off questions and a 5/5 on the end-of-term course survey metric, "The syllabus clearly outlines the policies and procedures of the course."

Expand full comment
7 more comments...

No posts