Catégories
Publications

Expertise, opacity, and trust in AI systems

Boisseau, É. Expertise, opacity, and trust in AI systems, Synthese 207, 104 (2026). https://doi.org/10.1007/s11229-026-05484-2

Article accessible en ligne en suivant ce lien : https://rdcu.be/e5NrU

Abstract / Résumé

This paper critically examines a family of arguments by analogy which suggest that the trust granted to an AI system should mirror the one usually granted by a layperson to a human expert. I particularly dispute the idea that, on the grounds that both share some degree of opacity, human experts and AI systems can be considered epistemic ‘black boxes’ and both be subject to the same blind trust on the part of non-experts. To uncover the problematic nature of this rather widespread analogy, I proceed by identifying a form of ambivalence in the notion of opacity mobilised, as well as a number of highly charged presuppositions concerning expertise (notably relating to a kind of obsession with what is sometimes dubbed its ‘veritistic’ character). I suggest that such a reductionist, monomaniacal conception of expertise is flawed, negligent or inadequate. The other forgotten facets of expertise are not merely cosmetic, but constitutive of it. I show that artificial systems cannot instantiate them, and that we cannot expect them to ever do so.

Catégories
Publications

Imitation and Large Language Models

Boisseau, É. « Imitation and Large Language Models » in Minds & Machines 34, 42 (2024). https://doi.org/10.1007/s11023-024-09698-6 SMASH
Article accessible en ligne en suivant ce lien : https://rdcu.be/dWvcH

Abstract / Résumé :

The concept of imitation is both ubiquitous and curiously under-analysed in theoretical discussions about the cognitive powers and capacities of machines, and in particular—for what is the focus of this paper—the cognitive capacities of large language models (LLMs). The question whether LLMs understand what they say and what is said to them, for instance, is a disputed one, and it is striking to see this concept of imitation being mobilised here for sometimes contradictory purposes. After illustrating and discussing how this concept is being used in various ways in the context of conversational systems, I draw a sketch of the different associations that the term ‘imitation’ conveys and distinguish two main senses of the notion. The first one is what I call the ‘imitative behaviour’ and the second is what I call the ‘status of imitation’. I then highlight and untangle some conceptual difficulties with these two senses and conclude that neither of these applies to LLMs. Finally, I introduce an appropriate description that I call ‘imitation manufacturing’. All this ultimately helps me to explore a radical negative answer to the question of machine understanding.