Christoph Durt, Thomas Fuchs, and Tom Froese investigate the astonishing capacities of Large Language Models (LLMs) to mimic human-like responses. They begin by acknowledging the unprecedented feats of these models, particularly GPT-3, which have led some to assert that they possess common-sense reasoning and even sentience. They caution, however, that these claims often overlook the instances where LLMs fail to produce sensical responses. Even as the models evolve and mitigate some of these limitations, the authors urge circumspection regarding the attribution of understanding and sentience to these systems.
The authors argue that the progress of LLMs invites a reassessment of long-standing philosophical debates about the limits of AI. The authors challenge the view, expressed by philosophers such as Hubertus Dreyfus, that AI is inherently incapable of understanding meaning. Given the emergent linguistic capabilities of these models, they query whether these advancements warrant attributing understanding to the computational system. Contrary to Dreyfus’s assertion that any formal system cannot be directly sensitive to the relevance of its situation, the authors propose that LLMs seem to exhibit this sensitivity in a pragmatic sense.
While the article explores the philosophical debates surrounding AI understanding and sentience, it does not definitively conclude whether LLMs truly understand or are sentient. The authors suggest that the human-like behaviour exhibited by these models may lead to the inference of a human-like mind. However, they argue that more nuanced and empirically informed positions are required. The authors further advocate for a more comprehensive assessment of LLM output, rather than relying on selective instances of impressive performance.
This research brings into focus the broader philosophical implications of our interaction with AI, particularly the ontological and epistemological assumptions we make when interacting with LLMs. The debate surrounding AI sentience and understanding illuminates the complexities inherent in defining consciousness and understanding, a philosophical quandary that dates back to Descartes and beyond. It forces us to interrogate the nature of understanding – is it a purely human phenomenon, or can it be replicated, even surpassed, by silicon-based entities? Moreover, it challenges our anthropocentric views of cognition and compels us to consider alternate forms of intelligence and understanding.
Looking forward, the study of AI and philosophy would benefit from an even deeper exploration of these questions. More empirical research is needed to understand the extent and limitations of LLMs’ capacities. Concurrently, philosophical inquiry can help define and refine the metrics by which we measure AI understanding and sentience. As we delve further into the AI era, it is crucial that we continue to scrutinize and challenge our assumptions about AI capabilities, not only to enhance our technological advancements but also to enrich our philosophical understanding of the world.
Abstract
Large language models such as ChatGPT are deep learning architectures trained on immense quantities of text. Their capabilities of producing human-like text are often attributed either to mental capacities or the modeling of such capacities. This paper argues, to the contrary, that because much of meaning is embedded in common patterns of language use, LLMs can model the statistical contours of these usage patterns. We agree with distributional semantics that the statistical relations of a text corpus reflect meaning, but only part of it. Written words are only one part of language use, although an important one as it scaffolds our interactions and mental life. In human language production, preconscious anticipatory processes interact with conscious experience. Human language use constitutes and makes use of given patterns and at the same time constantly rearranges them in a way we compare to the creation of a collage. LLMs do not model sentience or other mental capacities of humans but the common patterns in public language use, clichés and biases included. They thereby highlight the surprising extent to which human language use gives rise to and is guided by patterns.
Against AI Understanding and Sentience: Large Language Models, Meaning, and the Patterns of Human Language Use
