Abstract
AbstractMost, if not all, philosophers agree that computers cannot learn what words refers to from raw text alone. While many attacked Searle’s Chinese Room thought experiment, no one seemed to question this most basic assumption. For how can computers learn something that is not in the data? Emily Bender and Alexander Koller (2020) recently presented a related thought experiment—the so-called Octopus thought experiment, which replaces the rule-based interlocutor of Searle’s thought experiment with a neural language model. The Octopus thought experiment was awarded a best paper prize and was widely debated in the AI community. Again, however, even its fiercest opponents accepted the premise that what a word refers to cannot be induced in the absence of direct supervision. I will argue that what a word refers to is probably learnable from raw text alone. Here’s why: higher-order concept co-occurrence statistics are stable across languages and across modalities, because language use (universally) reflects the world we live in (which is relatively stable). Such statistics are sufficient to establish what words refer to. My conjecture is supported by a literature survey, a thought experiment, and an actual experiment.
Publisher
Springer Science and Business Media LLC
Subject
Artificial Intelligence,Philosophy
Reference76 articles.
1. Abdou, M., Kulmizev, A., Hershcovich, D., Frank, S., Pavlick, E., & Søgaard, A. (2021). Can language models encode perceptual structure without grounding? a case study in color. In: Proceedings of the 25th Conference on Computational Natural Language Learning, pp. 109–132. Association for Computational Linguistics, Online.
2. Aleksander, I. (2002). Neural depictions of ‘world’ and ‘self’: Bringing computational understanding to the Chinese room. In J. M. Preston & J. M. Bishop (Eds.), Views Into the Chinese room: New essays on Searle and artificial intelligence. Oxford University Press.
3. Arnulf, I., Uguccioni, G., Gay, F., Baldayrou, E., Golmard, J.-L., Gayraud, F., & Devevey, A. (2017). What does the sleeping brain say? Syntax and semantics of sleep talking in healthy subjects and in parasomnia patients. Sleep, 40(11).
4. Artetxe, M., Labaka, G., & Agirre, E. (2017) Learning bilingual word embeddings with (almost) no bilingual data. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), (pp. 451–462). Association for Computational Linguistics, Vancouver, Canada.
5. Bender, E. M., & Koller, A. (2020). Climbing towards NLU: On meaning, form, and understanding in the age of data. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Online.
Cited by
20 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. Figure Credits;Concepts at the Interface;2024-09-05
2. Concluding Thoughts;Concepts at the Interface;2024-09-05
3. Metacognition;Concepts at the Interface;2024-09-05
4. Representational Structure;Concepts at the Interface;2024-09-05
5. Drawing on Meaning;Concepts at the Interface;2024-09-05