Transformers, contextualism, and polysemy

[thumbnail of Transformers, contextualism, and polysemy - Final.pdf]
Text - Accepted Version
ยท Restricted to Repository staff only
Restricted to Repository staff only

Please see our End User Agreement.

It is advisable to refer to the publisher's version if you intend to cite from this work. See Guidance on citing.

Add to AnyAdd to TwitterAdd to FacebookAdd to LinkedinAdd to PinterestAdd to Email

Grindrod, J. orcid id iconORCID: https://orcid.org/0000-0001-8684-974X (2025) Transformers, contextualism, and polysemy. Ergo. ISSN 2330-4014 (In Press)

Abstract/Summary

The transformer architecture, introduced by Vaswani et al. (2017), is at the heart of the remarkable recent progress in the development of language models, including widely-used chatbots such as Chat-GPT and Claude. In this paper, I argue that we can extract from the way the transformer architecture works a theory of the relationship between context and meaning. I call this the transformer theory, and I argue that it is novel with regard to two related philosophical debates: the contextualism debate regarding the extent of context-sensitivity across natural language, and the polysemy debate regarding how polysemy should be captured within an account of word meaning.

Item Type Article
URI https://reading-clone.eprints-hosting.org/id/eprint/120383
Refereed Yes
Divisions Arts, Humanities and Social Science > School of Humanities > Philosophy
Publisher Michigan Publishing
Download/View statistics View download statistics for this item

University Staff: Request a correction | Centaur Editors: Update this record

Search Google Scholar