Discussion about this post

User's avatar
Henry Farrell's avatar

So this is in part a spin-off from the Dan Davies Extended Universe as the namecheck in passing suggests. I don't talk about accountability sinks, since I'm trying to get away from the LLMs-as-agents discussion, but there are I think some interesting further questions about dealing with variety. I mention in passing the problem that they tend to select against variety, but Dan has an interesting podcast interview with Patrick McKenzie where he suggests in passing that LLMs could be handy in making organizations better able to deal with complex environments. I suspect that neither of us has thought this through in huge depth, but since I'll be meeting up with him this afternoon, I may ask him ...

I'm guessing that the hallucinations problem is less of a problem for ideology, so long as you can be sure that you are not actually peddling fake quotes (as mentioned, NotebookLM has workarounds for this, but not knowing the Chinese system I can't say for sure). Explicating Chairman Xi thought is a ritual performance, where the maximally unsurprising outcome for some extrapolation of it is often going to be a useful thing to know.

Expand full comment
Gerben Wierda's avatar

I think 'summarisation' is not the best concept (though I can understand seeing these systems as a lossy compression of language or culture). Approximation is the best concept. The token-statistics (based on output from human understanding) can be used to approximate what the result of human understanding could be. These systems do not really summarise (see https://ea.rna.nl/2024/05/27/when-chatgpt-summarises-it-actually-does-nothing-of-the-kind/) even if you can use them to do so. They do a mix of 'ignoring what is to be summarised but generate mostly from parameters' and 'shortening', which isn't summarising (see link). Approximation also covers the behaviour of LLMs better outside the summarising use case.

GenAI also does not hallucinate, that is labeling it from the human perspective. The systems approximate, and some approximations are wrong (but still valid *approximations*). The errors aren't errors, they are fundamental features of the systems.

Thirdly, they don't need to be good to be disruptive. Innovation can be either "doing something heretofore impossible" (AlphaFold) or "doing something more cheaply" (either better, good, or 'good enough'). Most of LLM use is 'doing something cheaply' (both in terms of money as in result). Klarna replacing graphic artists with GenAI is an example of(see https://ea.rna.nl/2024/07/27/generative-ai-doesnt-copy-art-it-clones-the-artisans-cheaply/).

Lastly, *the* issue everyone is still ignoring is the fact that automation offers us a speed/efficiency gain, but the price paid is less agility. All organisations these day suffer that they become less agile because they have been welded to 'large landscapes of brittle machine logic'. Such landscapes are ever harder to change because logic is brittle, and as a result they have 'inertia'. This is an overall automation/IT issue. We may potentially expect landscapes (or even language itself) becoming welded to LLMs to slow doen in variation and change. Fun observation: human intelligence is also mostly 'mental automation' and this delivers (from an evolutionary perspective) necessary speed and efficiency, but shows the same price paid in agility. Our convictions, beliefs, assumptions are our mental automation, and they provide speed and efficiency, but they do not change easily. (See https://www.youtube.com/watch?v=3riSN5TCuoE). We're not moving towards a singularity point but a complexity crunch.

The providers tune their systems such that there is not so much randomness ('temperature') that even grammar fails, because we humans have a quick evaluation of intelligence that uses 'good language' as a proxy for 'intelligence'. So, you cannot push these beyond a certain temperature as they become more creative, but less convincing. Grammar, it turns out, is easier to approximate with tokens than meaning. See https://youtu.be/9Q3R8G_W0Wc

Expand full comment
9 more comments...

No posts