Skip to main content



Beliefs and Dispositions in Language Models and Ourselves

 When GPT-2 was first released, I began to wonder whether I was something like GPT: Did I contain beliefs and reason over them, or did I, like GPT, just have a disposition to say certain things in a particular situation? Here is a good summary of the two models of the self that I was comparing, at least roughly. If you give GPT a pro-life argument as a prompt, it will continue trying to generate pro-life arguments, and the same for pro-choice arguments. It doesn't believe one way or the other on the issue. On issues no one debates about and no one questions, it might be difficult to get it to generate sensible arguments against those positions, but it will happily attempt it. For example, with the prompt: Despite what many people think, polar bears do not live in the Arctic, but are purely equatorial creatures. They like warm water, and avoid snow and ice.  They can only be found on tropical islands like Hawaii.  GPT-3 generated "Polar bears are also known to enjoy basking in

Latest Posts

a wikipedestrian stroll

How the world was built for us

Language models and consciousness

Tau (A Translation of The Last Chapter of The Revelation of St. John)


The oldest thing that has never been lost

The Very Hungry Caterpillar

Well-oiled Machines

How to train your Facebook Feed

The New Age