| |
An action-oriented perspective changes the role of an individual from a passive observer to an actively engaged agent interacting in a closed loop with the world as well as with others. Cognition exists to serve action within a landscape that contains both. This chapter surveys this landscape and addresses the status of the pragmatic turn. Its potential influence on science and the study of cognition are considered (including perception, social cognition, social interaction, sensorimotor entrainment, and language acquisition) and its impact (...) on how neuroscience is studied is also investigated (with the notion that brains do not passively build models, but instead support the guidance of action). A review of its implications in robotics and engineering includes a discussion of the application of enactive control principles to couple action and perception in robotics as well as the conceptualization of system design in a more holistic, less modular manner. Practical applications that can impact the human condition are reviewed (e.g., educational applications, treatment possibilities for developmental and psychopathological disorders, the development of neural prostheses). All of this foreshadows the potential societal implications of the pragmatic turn. The chapter concludes that an action-oriented approach emphasizes a continuum of interaction between technical aspects of cognitive systems and robotics, biology, psychology, the social sciences, and the humanities, where the individual is part of a grounded cultural system. (shrink) | |
A discussion is going on in cognitive science about the use of representations to explain how intelligent behavior is generated. In the traditional view, an organism is thought to incorporate representations. These provide an internal model that is used by the organism to instruct the motor apparatus so that the adaptive and anticipatory characteristics of behavior come about. So-called interactionists claim that this representational specification of behavior raises more problems than it solves. In their view, the notion of internal representational (...) models is to be dispensed with. Instead, behavior is to be explained as the intricate interaction between an embodied organism and the specific make up of an environment. The problem with a non-representational interactive account is that it has severe difficulties with anticipatory, future oriented behavior. The present paper extends the interactionist conceptual framework by drawing on ideas derived from the study of morphogenesis. This extended interactionist framework is based on an analysis of anticipatory behavior as a process which involves multiple spatio-temporal scales of neural, bodily and environmental dynamics. This extended conceptual framework provides the outlines for an explanation of anticipatory behavior without involving a representational specification of future goal states. (shrink) | |
It has been just over 100 years since the birth of Alan Turing and more than 65 years since he published in Mind his seminal paper, Computing Machinery and Intelligence. In the Mind paper, Turing asked a number of questions, including whether computers could ever be said to have the power of “thinking”. Turing also set up a number of criteria—including his imitation game—under which a human could judge whether a computer could be said to be “intelligent”. Turing’s paper, as (...) well as his important mathematical and computational insights of the 1930s and 1940s led to his popular acclaim as the “Father of Artificial Intelligence”. In the years since his paper was published, however, no computational system has fully satisfied Turing’s challenge. In this paper we focus on a different question, ignored in, but inspired by Turing’s work: How might the Artificial Intelligence practitioner implement “intelligence” on a computational device? Over the past 60 years, although the AI community has not produced a general-purpose computational intelligence, it has constructed a large number of important artifacts, as well as taken several philosophical stances able to shed light on the nature and implementation of intelligence. This paper contends that the construction of any human artifact includes an implicit epistemic stance. In AI this stance is found in commitments to particular knowledge representations and search strategies that lead to a product’s successes as well as its limitations. Finally, we suggest that computational and human intelligence are two different natural kinds, in the philosophical sense, and elaborate on this point in the conclusion. (shrink) | |
An introduction to philosophy of language since Frege, focusing on the 20th century. | |
The development of artificial intelligence necessarily implies the anthropological question of the difference between human and artificial intelligence for two reasons: on the one hand artificial intelligence tends to be conceived on the model of human intelligence, on the other hand, a large part of types of artificial intelligence are designed in order to exhibit at least some features of what is conceived as being human intelligence. In this article I address this anthropological question in two parts. First I bring (...) into review and classify some of the main answers that have been proposed until now to this question. I argue that these variety of answers can be broadly classified in three categories, namely a (1) behaviorist, (2) a representational, and (3) a holistic understanding of human intelligence. In a second moment I propose an alternative way of understanding the difference between human and artificial intelligence, which is not essentialist but contextualist and content-related. Contrary to possible answers that I analyse in the first section, this alternative model does not aim at grasping the essence of human intelligence, which could or could not be reproduced in principle by artificial intelligence. It situates rather the fundamental differences between human and artificial intelligence in the context of human existence and the conceptual content of human intelligence, following the phenomenological description of one of its most fundamental features, namely its life-world. Grounding on this approach, it is possible to argue that human and artificial intelligence could be distinct, even if one could prove that they are eidetically, i.e. by their essence, identical. (shrink) | |
We describe a project to capitalize on newly available levels of computational resources in order to understand human cognition. We are building an integrated physical system including vision, sound input and output, and dextrous manipulation, all controlled by a continuously operating large scale parallel MIMD computer. The resulting system will learn to "think" by building on its bodily experiences to accomplish progressively more abstract tasks. Past experience suggests that in attempting to build such an integrated system we will have to (...) fundamentally change the way artificial intelligence, cognitive science, linguistics, and philosophy think about the organization of intelligence. We expect to be able to better reconcile the theories that will be developed with current work in neuroscience. (shrink) | |
No categories | |
A large body of compelling evidence has been accumulated demonstrating that embodiment – the agent’s physical setup, including its shape, materials, sensors and actuators – is constitutive for any form of cognition and as a consequence, models of cognition need to be embodied. In contrast to methods from empirical sciences to study cognition, robots can be freely manipulated and virtually all key variables of their embodiment and control programs can be systematically varied. As such, they provide an extremely powerful tool (...) of investigation. We present a robotic bottom-up or developmental approach, focusing on three stages: (a) low-level behaviors like walking and reflexes, (b) learning regularities in sensorimotor spaces, and (c) human-like cognition. We also show that robotic based research is not only a productive path to deepening our understanding of cognition, but that robots can strongly benefit from human-like cognition in order to become more autonomous, robust, resilient, and safe. (shrink) | |
Complex systems and complex missions take years of planning and force launches to become incredibly expensive. The longer the planning and the more expensive the mission, the more catastrophic if it fails. The solution has always been to plan better, add redundancy, test thoroughly and use high quality components. Based on our experience in building ground based mobile robots (legged and wheeled) we argue here for cheap, fast missions using large numbers of mass produced simple autonomous robots that are small (...) b y today's standards (1 to 2 Kg). We argue that the time between mission conception and implementation can be radically reduced, that launch mass can be slashed, that totally autonomous robots can be more reliable than ground controlled robots, and that large numbers of robots can change the tradeoff between reliability of individual components and overall mission success. Lastly, we suggest that within a few years it will be possible at modest cost to invade a planet with millions of tiny robots. (shrink) | |
In this paper, the current AI view that emergent functionalities apply only to the study of subcognitive agents is questioned; a hypercognitive view of autonomous agents as proposed in some AI subareas is also rejected. As an alternative view, a unified theory of social interaction is proposed which allows for the consideration of both cognitive and extracognitive social relations. A notion of functional effect is proposed, and the application of a formal model of cooperation is illustrated. Functional cooperation shows the (...) role of extracognitive phenomena in the interaction of intelligent agents, thus representing a typical example of emergent functionality. (shrink) |