A rough attempt at laying out what in philosophy is most relevant for AI.
While I expected studying Philosophy as an undergraduate to be personally useful and indirectly useful to my professional career, I didn’t forsee how relevant it would be to our increasingly AI-infused world.
In this post, Matt Mandel, after “[coming] to realize that using LLMs is pushing all of us to more closely examine our philosophical assumptions” has sketched out “a rough attempt at laying out what in philosophy is most relevant for AI.”
I’ve taken the list — covering things with which I’m familiar and those things I’d like to follow-up on — and added links. The “What is agency?" section is particularly timely, as I’ve been thinking about this a lot recently but need to do more reading.
Part 1: Philosophical Concepts
What is a mind?
- Nagel, What is it like to be a bat
- Chalmers, Facing up to the problem of consciousness
- Putnam, The Nature of Mental States
- Searle, Minds, Brains, and Programs
What is agency?
- Anscombe, Intention
- Dennett, The Intentional Stance
- Bratman, Shared Cooperative Activity
- Applbaum, Legitimacy (group agents)
Who has moral status?
- Kant, The Metaphysics of Morals (indirect duties)
- Singer, Animal Liberation
- Korsgaard, Fellow Creatures
What are reasons and values?
- Hume, Enquiry Concerning the Principles of Morals
- Parfit, On What Matters (Volume II, Part 6)
- Mackie, Ethics: Inventing Right and Wrong
- Korsgaard, The Sources of Normativity
Part 2: How should we build AI
How might AIs come to implement our values?
- Kripke, Wittgenstein on Rules and Private Language
- Bostrom, The Superintelligent Will
- Plato, The Republic (Book I)
- Anthropic, Emergent Misalignment
How should AI be involved in governance?
- Applbaum, Legitimacy (“domination by a stone”)
- Séb Krier, Coasean Bargaining at Scale
- Newsom, Snell v United Specialty Insurance Company (concurrence)
- Cass Sunstein and Vermeule, Law and Leviathan
How might AI impact human flourishing?
- Aristotle, Nicomachean Ethics
- Mill, On Liberty
- Frankfurt, Freedom of the Will and the Concept of a Person
- Brendan McCord, Live by the Claude, Die by the Claude
How do we find meaning in a post-AGI world?
- Camus, The Myth of Sisyphus
- Nozick, Philosophical Explanations (Philosophy and the Meaning of Life)
- Bostrom, Deep Utopia
- Mandel, The Top Three AGI-Proof Careers and What They Reveal About Our Humanity and The Future
In terms of Part 1, to the What is a mind? section it’s definitely worth adding Turing’s foundational paper from 1950 asking “can machines think?” It’s the one that introduces the “Turing test” and directly sets up the questions that Searle and Nagel are responding to.
I’d also add a section entitled What is knowledge? and include:
- Dreyfus, What Computers Can’t Do — which argues that embodied, situated knowledge resists formalisation.
- Wittgenstein, Philosophical Investigations — not the easiest of reads, but it’s where we get notions of the impossibility of a private language and the difficulty of defining terms such as “game”.
I had to ask Perplexity for help with Part 2. I haven’t read any of these but apparently they’re relevant additions:
- How should we build AI?
- Vallor, Technology and the Virtues — A virtue-ethics framework for emerging technology which Perplexity describes as “arguably the most important contemporary work bridging classical ethics and AI practice.” The original list focuses on alignment as a technical problem, whereas Vallor asks what kind of people we need to be to build good technology.
- How might AI impact human flourishing?
- Crawford, Atlas of AI — I’ve had this book on my shelf for a while but haven’t read it yet. It’s a materialist analysis of AI’s supply chains, labour exploitation, and environmental costs, which sounded too depressing for me to read last year.
- What does AI mean for how we know things?
- Floridi, The Ethics of Artificial Intelligence — A framework of five principles for ethical AI (beneficence, non-maleficence, autonomy, justice, and explicability) which apparently is “the standard reference in AI governance circles.”
- Nguyen, Games: Agency as Art — Explores how gamification narrows our values into simplified metrics and therefore relevant to AI alignment. If we have to specify values precisely enough for machines to optimise, do we inevitably impoverish them?
Source: Substack Notes
Image: Hanna Barakat