Creating a Large Language Model of a Philosopher
Eric Schwitzgebel, David Schwitzgebel, and Anna StrasserMind and Language (2023) [online article mila.12466, print forthcoming]
Can large language models be trained to produce philosophical texts that are difficult to distinguish from texts produced by human philosophers? To address this question, we fine-tuned OpenAI's GPT-3 with the works of philosopher Daniel C. Dennett as additional training data. To explore the Dennett model, we asked the real Dennett ten philosophical questions and then posed the same questions to the language model, collecting four responses for each question without cherry-picking. We recruited 425 participants to distinguish Dennett's answer from the four machine-generated answers. Experts on Dennett's work (N = 25) succeeded 51% of the time, above the chance rate of 20% but short of our hypothesized rate of 80% correct. For two of the ten questions, the language model produced at least one answer that experts selected more frequently than Dennett's own answer. Philosophy blog readers (N = 302) performed similarly to the experts, while ordinary research participants (N = 98) were near chance distinguishing GPT-3's responses from those of an "actual human philosopher".
By following any of the links below, you are requesting a copy for personal use only, in accord with "fair use" laws.
Click here to view the final manuscript version as a PDF file: Creating a Large Language Model of a Philosopher (pdf, May 9, 2023).
Click here to view the final manuscript version as an html file: Creating a Large Language Model of a Philosopher (html, May 9, 2023).
Or email eschwitz at domain: ucr.edu for a copy of this paper.