Upcoming Masterclass on Philosophy of AI: Additional Materials

I’m offering a philosophy of AI masterclass, “Concepts in Machines”, at the University of Zürich later this month (April 23–25, 2026). In this post, I briefly sketch my intentions for this course and collect extra materials not included in the syllabus.

Sketch of Course Topic

The course explores the role of concepts in neural models. What are the candidates for conceptual representations in LLMs? Starting from this question, we turn to debates in philosophy of AI. For example, we will discuss whether neural language models and especially LLMs understand meaning and produce meaningful output. Both philosophical contributions and publications in computer science venues will be considered.

One of the goals of the course is to ensure that philosophy of AI remains connected to the current and rapidly developing state of AI research. Hence, the course includes

  1. A hands-on component: Participants will interact with the internals of neural models, including small neural language models.
  2. Discussions of recent findings: We will debate findings in recent AI research that bear upon philosophy but have received limited attention so far.

Additional Readings

As the masterclass lasts only three days, we cannot cover all contributions to the central topic. I list here additional readings on the core topic of how LLMs capture meaning/concepts.

Vector Semantics

The Cognitive Alignment of LLMs

More on the Grounding Problem for LLMs

Additional Talks

The seminar of the DIC (Doctorat en informatique cognitive) at the Université du Québec à Montréal has hosted some excellent talks relevant for the masterclass. I recommend at least the following:

There are many more excellent talks on the website!


Note: I used an LLM to copy-edit this post.

Previous
Notes on the Symbol Groundi...