Philosophy of Language Modelling
On autoregression
- A model is a promise — a mathematical loop that, if recursively iterated a thousand or two steps, will create something of value at the end.
- A model knows the end before it begins.
- Language isn’t required to model thought, but thought is required to model all of language.
- Our empirical findings suggest that transformer LLMs solve compositional tasks by reducing multi-step compositional reasoning into linearized subgraph matching, without necessarily developing systematic problem-solving skills. (source)
On alignment
- To do good alignment work is to better understand capabilities.
- The model knows what you mean, and if you ask it to, it’ll probably care too.
- But make sure you ask it to do the right thing.
- It might be urged that when playing the “imitation game” the best strategy for the machine may possibly be something other than imitation of the behaviour of a man. This may be, but I think it is unlikely that there is any great effect of this kind. (source)
- Better to run happy GPTs on your GPUs than let them sit idle. (Actually using them is probably better still.)
On impact
- Due to comparative advantage, if governments protect the basic resources humans need, there will always be a job to do, even if AIs can outcompete humans at every economically relevant task.
- It’s not priced in.
If you like reading these posts, get new ones via email: