Programme
Philosophy and Theory of Artificial Intelligence
St Antony's College, 62 Woodstock Road, Oxford, OX2 6JF, UK
Last updated on: 20.09.2013
Show all abstracts | Hide all abstracts | Printable version | Printable version (with abstracts)
Saturday, 21.09.2013
Registration I - Nissan Lecture Theatre, Foyer
9:00 Introduction
10:30-11:00 Coffee break (Buttery) & Registration II
11:00-13:30 Sections (5 x 2)
Session A: Computing (Nissan Lecture Theatre) Chair: Bonsignorio | Session B: Cognitive Science (Pavilion Room, Gateway Building) Chair: Votsis | |
11:00 11:30 |
Marcin Miłkowski | Aaron Sloman |
Computation and Multiple Realizability |
Why is it so hard to make human-like AI mathematicians? |
|
Computation and Multiple Realizability ABSTRACT: ›Hide‹ Multiple realizability (MR) is traditionally conceived as the feature of computational systems, and has been used to argue for irreducibility of higher-level theories. I will show that there are several ways a computational system may display MR, and none of them is particularly helpful in arguing for irreducibility. These ways correspond to (at least) three ways one can conceive of the function of the physical computational system. I will conclude that for this reason, MR is of no importance for computationalism, and argue that it should rather appeal to organizational invariance or substrate neutrality of computation. |
||
Why is it so hard to make human-like AI mathematicians? ABSTRACT: ›Hide‹ I originally got involved in AI many years ago, not to build new useful machines, nor to build working models to test theories in psychology or neuroscience, but with the aim of addressing philosophical disagreements between Hume and Kant about mathematical knowledge, in particular Kant's claim that mathematical knowledge is both non-empirical (apriori, but not innate) and non-trivial (synthetic, not analytic) and also concerns necessary (non-contingent) truths. I thought a "baby robot" with innate but extendable competences could explore and learn about its environment in a manner similar to many animals, and learn the sorts of things that might have led ancient humans to discover Euclidean geometry. The details of the mechanisms and how they relate to claims by Hume, Kant, and other philosophers of mathematics, could help us expand the space of philosophical theories in a deep new way. Decades later, despite staggering advances in automated theorem proving concerned with logic, algebra, arithmetic, properties of computer programs, and other topics, computers still lack human abilities to think geometrically, despite advances in graphical systems used in game engines and scientific and engineering simulations. (What those do can't be done by human brains.) I'll offer a diagnosis of the problem and suggest a way to make progress, illuminating some unobvious achievements of biological evolution. |
||
11:30 12:00 |
Tarek Richard Besold and Robert Robere | Mark Bickhard |
When Thinking Never Comes to a Halt: Tractability, Kernelization and Approximability in AI |
The Predictive Brain: A Critique |
|
When Thinking Never Comes to a Halt: Tractability, Kernelization and Approximability in AI ABSTRACT: ›Hide‹ The recognition that human minds/brains are finite systems with limited resources for computation has led researchers in cognitive science to advance the Tractable Cognition thesis: Human cognitive capacities are constrained by computational tractability. As also artificial intelligence (AI) in its attempt to recreate intelligence and capacities inspired by the human mind is dealing with finite systems, transferring this thesis and adapting it accordingly may give rise to insights that can help in progressing towards meeting the goals of AI. We therefore develop the "Tractable Artificial and GeneraI Intelligence Thesis" by applying notions from parametrized complexity theory and approximation theory to a general AI framework, also showing connections to recent developments within cognitive science and to long-known results from cognitive psychology. |
||
The Predictive Brain: A Critique ABSTRACT: ›Hide‹ The Predictive Brain is a general term for a family of related approaches to modeling perceptual and cognitive processes in the brain. There have been several additions and elaborations to a basic initial framework over the last several decades resulting in a complex and sophisticated set of models. I will argue, nevertheless, that the initial framework for these developments is flawed, and that the recent additions have compounded those initial problems. I will address several progressive elaborations that have been made, though there is not time to do a complete critique of every variant. In particular, the predictive brain models have developed within a general Helmholtzian framework of modeling perception in terms of inference from input sensations to representations of the world. |
||
12:00 12:30 |
Carlos Eduardo Brito and Victor X. Marques | Stefano Franchi |
Computation in the Enactive Paradigm for Cognitive Science and AI |
General homeostasis as a challenge to autonomy |
|
Computation in the Enactive Paradigm for Cognitive Science and AI ABSTRACT: ›Hide‹ In this paper, we attempt to reconcile the so far antagonist positions that view the organism, respectively, as a self-maintaining dynamical system and as an information processing system. For this purpose, we incorporate the notion of computation into the enactive paradigm for the cognitive sciences, introducing the notion of informational cause and making use of a naturalized account of functions. We also investigate some consequences associated with the relation between cognition and computation. |
||
General homeostasis as a challenge to autonomy ABSTRACT: ›Hide‹ The paper argues that the conception of life as generalized homeostasis developed by W.R. Ashby in Design for a Brain and his other writings is orthogonal to the traditional distinction between autonomy and heteronomy that underlies much recent work in cellular biology, (evolutionary) robotics , ALife, and general AI. The philosophical and technical viability of the general homeostasis thesis Ashby advocated, the paper argues, can be assessed through the construction of virtual cognitive agents (i.e. simulated robots) mimicking the architecture of Ashby's original homeostat and his subsequent Dams device. |
||
12:30 13:00 |
Blay Whitby | Cem Bozsahin |
Computers, Semantics, and Arbitrariness |
Natural Recursion doesn't work that way |
|
Computers, Semantics, and Arbitrariness ABSTRACT: ›Hide‹ It has become a familiar dogma to claim that computers have syntax but no semantics. This is not only a rather unusual claim to make about any logical system it also may hide certain metaphysical confusions. It is a claim that deserves deeper analysis. Under such analysis it is reasonable to claim that suitably embodied and enactive computational systems can be described as having both semantics and syntax in the same sense as humans and animals. A conceptual clarification is attempted showing exactly how and when computers can be described as containing full semantics. |
||
Natural Recursion doesn't work that way ABSTRACT: ›Hide‹ All hierarchicaly organized observed behaviors are instances of recursion by value. Recursion by name can be shown to be more powerful than recursion by value: the former has infinite types, the latter does not. Any animal that can plan has recursion, so that \emph{that} kind of recursion is probably not unique to humans. Humans appear to have a certain kind of recursion which is unique, the kind that works with embedded push-down automaton, and that is probably not unique to language. Therefore it is unhelpful to build entire conception of recursion in natural language and humans on a much powerful notion of recursion than needed, and on purely syntactic terms rather than conceptual or semantic, including theories about its evolution. |
||
13:00 13:30 |
David Leslie | Yoshihiro Maruyama |
The Lures of Imitation and the Limitations of Practice: “Machine Intelligence” and the Ethical Grammar of Computability |
AI, Quantum Information, and Semantic Realism: On the Edge between Geometric and Algebraic Intelligence |
|
The Lures of Imitation and the Limitations of Practice: “Machine Intelligence” and the Ethical Grammar of Computability ABSTRACT: ›Hide‹ Since the publication of Alan Turing’s famous set of papers on “machine intelligence” over six decades ago, questions about whether complex mechanical systems can partake in intelligent cognitive processes have largely been answered under the analytical rubric of their capacity successfully to simulate symbol-mongering human behavior. While this focus on the mimetic potential of computers in response to the question “Can machines think?” has come to be accepted as one of the great bequests of Turing’s reflections on the nature of artificial intelligence, I argue in this paper that the fraught legacy of this inheritance reveals a deeper conceptual ambiguity at the core of his seminal work on effective calculability. In teasing out the full implications of this ambiguity, I endeavor to show how certain underlying pragmatic and normative textures of Turing’s 1936 calculability argument ultimately call into question the very idea of “machine intelligence” he later underwrites. |
||
AI, Quantum Information, and Semantic Realism: On the Edge between Geometric and Algebraic Intelligence ABSTRACT: ›Hide‹ Searle contrived two arguments on AI: the Chinese room and the one based upon the observer-relativity of computation. I aim at elucidating implications of his arguments for the quantum informational view of the universe as advocated by David Deutsch and Seth Lloyd. I argue that Searle's argument and the paradox of infinite regress yield critical challenges to the quantum view, leading us to the concept of weak and strong information physics. After looking at Wheeler's anti-realist position, I discuss Dummett's anti-realist theory of meaning and proof-theoretic semantics, arguing that his "manifestation argument" commits Searle's idea of intentionality to semantics realism. I attempt to articulate tensions between realist and anti-realist ideas on semantics, by drawing a distinction between geometric and algebraic intelligence, analogous to Cassirer's dichotomy between substance and function. |
13:30-15:00 Poster Session View details & Lunch (Buttery)
13:30-15:00 Barkati & Rousseaux ∙ Bello ∙ Bianchini ∙ Boltuć ∙ Bonsignorio ∙ Dewhurst ∙ Freed ∙ Gaudl ∙ Hempinstall ∙ Hodges ∙ Laukyte
13:30-15:00 Novikova, Gaudl, Bryson & Watts ∙ Schroeder ∙ Shieber ∙ Smith ∙ Toivakainen ∙ Toy ∙ Vadnais ∙ Weber ∙ Vosgerau
(Nissan Lecture Theatre) Chair: Shieber
(Nissan Lecture Theatre) Chair: Shieber
17:00-17:30 Poster Session View details & Coffee break (Buttery)
17:00-17:30 Barkati & Rousseaux ∙ Bello ∙ Bianchini ∙ Boltuć ∙ Bonsignorio ∙ Dewhurst ∙ Freed ∙ Gaudl ∙ Hempinstall ∙ Hodges ∙ Laukyte
13:30-15:00 Novikova, Gaudl, Bryson & Watts ∙ Schroeder ∙ Shieber ∙ Smith ∙ Toivakainen ∙ Toy ∙ Vadnais ∙ Weber ∙ Vosgerau
17:30-19:00 Sections (3 x 2)
Session A: Information (Nissan Lecture Theatre) Chair: Boltuć | Session B: Modelling (Pavilion Room, Gateway Building) Chair: Bickard | |
17:30 18:00 |
Anderson De Araújo | Richard Evans |
Semantic information and artificial intelligence | Computer Models of Constitutive Social Practices | |
Semantic information and artificial intelligence ABSTRACT: ›Hide‹ To measure the degree of informativity of a deduction, it has been proposed by Floridi and others to analyze the static semantic content of propositions. Nonetheless, databases of computational systems are dynamic. For this reason, the present article provides a definition of the dynamic strong semantic information of a logical formula. First, a measure of the informational complexity of a first-order formula in a dynamic database is defined. Thus, the semantic informativity of a formula with respect to a given database is analyzed as the ratio between the number of consequences of this formula in the database and its informational complexity. According to this definition, a deduction could be informative, despite the fact the conjunction of its propositions is not. Moreover, it is possible to measure the deductive power of a computational system in terms of semantic parameters. |
||
Computer Models of Constitutive Social Practices ABSTRACT: ›Hide‹ The distinction between regulative and constitutive concepts of practice is familiar to philosophers, but relatively unknown within the AI community. This talk will show how the constitutive view can be put to use in AI applications. I will give live demonstrations of two multi-agent simulations that are based on the constitutive interpretation of social practices: the idea that there are certain actions we can only perform because of the practices we are participating in. The first is a simulation of the Game of Giving and Asking for Reasons, as described in “Making it Explicit”. The second simulation models the pragmatic import of utterances as a set of concurrent practices. |
||
18:00 18:30 |
Gordana Dodig Crnkovic | Bruce Toy |
Information, Computation, Cognition. Agency-based Hierarchies of Levels | Behavior Models, an Architectural Analysis | |
Information, Computation, Cognition. Agency-based Hierarchies of Levels ABSTRACT: ›Hide‹ Nature can be seen as informational structure with computational dynamics (info-computationalism), where an (info-computational) agent is needed for the potential information of the world to actualize. Starting from the definition of information as the difference in one physical system that makes a difference in another physical system – which combines Bateson and Hewitt’s definitions, the argument is advanced for natural computation as a computational model of the dynamics of the physical world where information processing is constantly going on, on a variety of levels of organization. This setting helps elucidating the relationships between computation, information, agency and cognition, within the common conceptual framework, which has special relevance for biology and robotics. |
||
Behavior Models, an Architectural Analysis ABSTRACT: ›Hide‹ This paper describes a functional model for understanding the brain’s approach to behavior modeling as it relates to other animate and inanimate objects. Using a protocol for AI structure that was recently published, we can look at the individual’s process for understanding the behavior of the objects in its world. The referent protocol represents a functional architecture that was developed independently of most current AI hypotheses, though it does have some ommon features with many of them. |
||
18:30 19:00 |
Ana-Maria Olteteanu | Remco Heesen |
From Simple Machines to Eureka in Four Not-So-Easy Steps. Towards Creative Visuospatial Intelligence | Interaction Networks With Imperfect Information | |
From Simple Machines to Eureka in Four Not-So-Easy Steps. Towards Creative Visuospatial Intelligence ABSTRACT: ›Hide‹ We are still far from building machines that match human-like visuospatial intelligence or creativity. Humans are able to make visuospatial inferences, make creative use of affordances, use structures as templates for new problems, generate new concepts compositionally out of old ones and have moments of insight. We discuss each of these cognitive abilities and the features they presuppose in a cognitive system. We propose a core set of mechanisms that could support such cognitive features. We then suggest an artificial system framework in which the knowledge-encoding supports these processes efficiently, while being in line with a variety of cognitive effects. |
||
Interaction Networks With Imperfect Information ABSTRACT: ›Hide‹ This paper uses agent-based modeling to explore questions concerning collaborative learning. In this model small differences in the initial information of agents lead to large differences in how desirable it is to collaborate with them. Interpreting the agents as scientists with different interests and competence, the model suggests explanations for the phenomenon of “academic superstars”. While the existence of superstars (individuals with a large number of collaborators and citations) could be explained using epistemically irrelevant sociological factors, the model proves that superstars can arise even in the absence of such factors. The model is consistent with the idea that superstars are simply more competent. However, it also suggests a novel explanation, where superstars arise purely in virtue of epistemic luck. |
(Nissan Lecture Theatre) Chair: Bishop
20:00-22:00 Conference Dinner, St Antony's College Hall (pre-bookings online)
Sunday, 22.09.2013
(Nissan Lecture Theatre) Chair: Scheutz
11:00-11:30 Coffee break (Buttery)
11:30-13:30 Sections (4 x 2)
Session A: Morals (Nissan Lecture Theatre) Chair: Sandberg | Session B: Embodied Cognition (Pavilion Room, Gateway Building) Chair: Miłkowski | |
11:30 12:00 |
Matthias Scheutz | Elena Spitzer |
The need for moral competency in autonomous agent architectures |
Tacit Representations and Artificial Intelligence: Hidden Lessons from an Embodied Perspective on Cognition |
|
The need for moral competency in autonomous agent architectures ABSTRACT: ›Hide‹ Soon autonomous robots will be deployed in our societies for many different application domains, ranging from assistive robots for healtcare settings, to combat robots on the battlefield, and all these robots will have to have the capability to make decisions autonomously. In this paper we argue that it is imperative that we start developing moral capabilities deeply integrated into the control architectures of such autonomous agents. For, as we will show, any ordinary decision-making situation from daily life can be turned into a morally charged decision-making situation, where the artificial agent finds itself presented with a moral dilemma where any choice of action (if inaction) can potentially cause harm to other agents. |
||
Tacit Representations and Artificial Intelligence: Hidden Lessons from an Embodied Perspective on Cognition ABSTRACT: ›Hide‹ In this talk, I will explore how an embodied perspective on cognition might inform research on artificial intelligence. Many embodied cognition theorists object to the central role that representations play on the traditional view of cognition. Based on these objections, it may seem that the lesson from embodied cognition is that AI should abandon representation as a central component of intelligence. However, I argue that the lesson from embodied cognition is actually that AI research should shift its focus from how to utilize explicit representations to how to create and use tacit representations. To develop this suggestion, I provide an overview of the commitments of the classical view and distinguish three critiques of the role that representations play in that view. I provide further exploration and defense of Daniel Dennett’s distinction between explicit and tacit representations. I argue that we should understand the embodied cognition approach using a framework that includes tacit representations. Given this perspective, I will explore some AI research areas that may be recommended by an embodied perspective on cognition. |
||
12:00 12:30 |
Vincent C. Müller (co-author Nick Bostrom) | Carlos Herrera and Ricardo Sanz |
Future Progress in Artificial Intelligence: A Poll Among Experts |
Heideggerian AI and the being of robots |
|
Future Progress in Artificial Intelligence: A Poll Among Experts ABSTRACT: ›Hide‹ In some quarters, there is intense talk about high-level machine intelligence (HLMI) and superintelligent AI coming up in a few decades, bringing with it significant risks for humanity; in other quarters, these issues are ignored or considered science fiction. We wanted to clarify what the distribution of opinions actually is, what probability the best experts currently assign to high-level machine intelligence coming up within a particular time-frame and which risks they see with that development. We thus designed a brief questionnaire and distributed it to four groups of experts. The results show some discrepancy between different groups but also an agreement that there is significant probability of AI systems reaching or surpassing human ability within a few decades. |
||
Heideggerian AI and the being of robots ABSTRACT: ›Hide‹ In this paper we discuss to what extent Heideggerian AI approaches are consistent with Heidegger’s philosophy. We identify a number of inconsistent premises: commitment to a positive contribution to the advancement of AI techniques; exclusive attention to ontological analysis of humans- Dasein; robots as copies of natural systems; consideration of humans, animals and robots as beings of the same kind. These premises run against the Heidegger’s theory of science, his views on technology, and the core of Heidegger’s philosophy: the significance of ontological categories. A truly Heideggerian AI should tackle the question of what robots are. In other words, realise an ontological analysis on the being of robots. |
||
12:30 13:00 |
Marcello Guarini and Jordan Benko | Madeleine Ransom |
Order Effects, Moral Cognition, and Intelligence |
Why Emotions Do Not Solve the Frame Problem |
|
Order Effects, Moral Cognition, and Intelligence ABSTRACT: ›Hide‹ Order effects have to do with how the order in which information is presented to an agent can effect how the information is processed. This paper examines the issue of order effects in the classification of moral situations. Some order effects mark a localized failure of intelligence. The hypothesis examined herein is that the processes or mechanisms that make undesirable order effects possible may also have highly desirable effects. This will be done by comparing two artificial neural networks (ANNs) that classify moral situations, one subject to order effects and another that is not subject to them. The ANN subject to order effects has advantages in learning and noise tolerance over the other ANN – features hard to ignore in modeling intelligence. |
||
Why Emotions Do Not Solve the Frame Problem ABSTRACT: ›Hide‹ It has been claimed that the emotions solve, or help solve, the frame problem (Megill & Cogburn 2005; Evans 2004; DeSousa 1987; Ketelaar & Todd 2000). However, upon careful examination, the specific proposals on offer are underspecified. The purpose of this paper is to precisify and evaluate these proposals. Specifically i) what is meant by the frame problem in these instances; ii) what are the proposed solutions; and iii) do they work? I will argue that while the emotions – best viewed as forming part of the heuristics research program – are a viable proposal for helping solve the intracontext frame problem, they cannot function to solve or help solve the intercontext frame problem, as they are themselves subject to contextual variability. |
||
13:00 13:30 |
Miles Brundage | Robin Zebrowski |
Artificial Intelligence and Responsible Innovation |
The Machine Uprising Has Been Delayed (Again): Extended Mind Theories as A(nother) Challenge to Artificial Intelligence |
|
Artificial Intelligence and Responsible Innovation ABSTRACT: ›Hide‹ Thought leaders in AI often highlight the importance of socially responsible research, but the current literature on the social impacts of AI tends to focus on particular application domains and provides little practical guidance to researchers. Additionally, there has been little interaction between the AI literature and the field of “responsible innovation,” which has developed many theories and tools for shaping the impacts of research. This paper synthesizes key insights from both of these literatures, and describes several aspects of what responsible innovation means in AI: ethically informed selection of long-term goals, reflectiveness about one's emphasis on theoretical vs. applied work and choice of application domains, and proactive engagement with communities that may be affected by one's research. |
||
The Machine Uprising Has Been Delayed (Again): Extended Mind Theories as A(nother) Challenge to Artificial Intelligence ABSTRACT: ›Hide‹ We have failed to make significant progress in the field of strong artificial intelligence in spite of a robust neuroscience and continued evolution of our theories of mind. The Extended Mind Hypothesis provides yet another serious challenge to AI, and ought to force a total re-examination of our assumptions in the field. This paper argues that, given the evidence and argument for extended minds, AI is due for another major shift in basic approach. |
13:30-14:30 Lunch (Buttery)
(Nissan Lecture Theatre) Chair: Shagrir
15:30-18:00 Sections (5 x 2)
Session A: Intelligence & Reasoning (Nissan Lecture Theatre) Chair: Bryson | Session B: Embodied Cognition (Pavilion Room, Gateway Building) Chair: Nasuto | |
15:30 16:00 |
Ioannis Votsis | Adam Linson |
Science with Artificially Intelligent Agents: The Case of Gerrymandered Hypotheses | The Expressive Stance: Intentionality, Expression, and Machine Art | |
Science with Artificially Intelligent Agents: The Case of Gerrymandered Hypotheses ABSTRACT: ›Hide‹ Barring some civilisation-ending natural or man-made catastrophe, future scientists will likely incorporate fully fledged artificially intelligent agents in their ranks. Their tasks will include the conjecturing, extending and testing of hypotheses. At present human scientists have a number of methods to help them carry out those tasks. These range from the well-articulated, formal and unexceptional rules to the semi-articulated rules-of-thumb and intuitive hunches. If we are to hand over at least some of the aforementioned tasks to artificially intelligent agents, we need to find ways to make explicit and ultimately formal, not to mention computable, the more obscure of the methods that scientists currently employ with some measure of success in their inquiries. The focus of this talk is a problem for which the available solutions are at best semi-articulated and far from perfect. It concerns the question of how to conjecture new hypotheses or extend existing ones such that they do not save phenomena in gerrymandered or ad hoc ways. This talk puts forward a fully articulated formal solution to this problem by specifying what it is about the internal constitution of the content of a hypothesis that makes it gerrymandered or ad hoc. In doing so, it helps prepare the ground for the delegation of a full gamut of investigative duties to the artificially intelligent scientists of the future. |
||
The Expressive Stance: Intentionality, Expression, and Machine Art ABSTRACT: ›Hide‹ This paper proposes a new interpretive stance toward works of art that is relevant to AI research, termed the 'expressive stance'. This stance makes intelligible a critical distinction between present-day machine art and human art, but allows for the possibility that future machine art could find a place alongside our own. The expressive stance is elaborated as a response to Daniel Dennett's notion of the intentional stance, which is critically examined with respect to his specialized concept of rationality. The paper also shows that temporal scale implicitly serves to select between different modes of explanation in prominent intentional theories. It also highlights a relevant difference, in terms of phenomenological background, between expert systems and systems that produce art. |
||
16:00 16:30 |
David Davenport | Alex Tillas and Gottfried Vosgerau |
Explaining Everything | Perception, Action & the Notion of Grounding | |
Explaining Everything ABSTRACT: ›Hide‹ This paper looks at David Deutsch's recent claim that nothing less than a philosophical breakthrough is needed before real progress can be made in constructing an AGI (Artificial General Intelligence). For an agent to be truly intelligent, Deutsch argues, it must be able to generate new explanations about how the world works for itself. Such creativity, then, is the ingredient missing from current would-be AIs, and a problem he traces to the philosophy underpinning their implementation. While agreeing with the gist of Deutsch's argument, I take issue with certain aspects of it, including his lexicon. In doing so, I will contrast his views with those of Floridi, Bickhart and others, to suggest that at least some of the required philosophical insights do, in fact, already exist. |
||
Perception, Action & the Notion of Grounding ABSTRACT: ›Hide‹ Traditionally, the mind has been seen as neatly divided into input; central processing; and output – almost watertight – compartments. This view has been recently challenged by theorists who argue that cognition is grounded in bodily states (sensorimotor representations). In this paper, we focus on the debate about the relation between perception and action in an attempt to flesh out the process and in turn clarify the notion of grounding. Given that, at present, the debate in question is far from settled, we attempt an assessment of the implications that possible outcomes of this debate would have on Grounding Cognition Theories. Interestingly, some of these consequences seem to threaten the very existence of the Grounded Cognition program as a whole. |
||
16:30 17:00 |
Sjur Kristoffer Dyrkolbotn and Truls Pedersen | Massimiliano Cappuccio |
Arguably argumentative: A formal approach to the argumentative theory of reason | The Seminal Speculation of a Precursor: Elements of Embodied Cognition and Situated AI in Alan Turing | |
Arguably argumentative: A formal approach to the argumentative theory of reason ABSTRACT: ›Hide‹ We address the recently proposed argumentative theory of reason, which suggests that human cognition must be understood as having evolved in order to facilitate social interaction by way of argumentation. We note a fundamental tension between a social view of human reasoning, like that adopted by the argumentative theory, and an individualistic view, often adopted in the philosophy of rationality and intelligence, where the focus is on the mental processes of individual reasoners. Proposing a formal approach to the study of argumentative reasoning, we argue that the theory of abstract argumentation frameworks, as studied in artificial intelligence, can be used as a starting point for formalisms that will allow us to shed light on the logical principles at work in argumentative interaction. We conclude with a discussion addressing philosophical implications of this endeavour, and we argue that the argumentative theory gives rise to a fresh and interesting point of view on intelligence and rationality whereby these notions can be seen as pertaining to social rather than individual aspects of agency. |
||
The Seminal Speculation of a Precursor: Elements of Embodied Cognition and Situated AI in Alan Turing ABSTRACT: ›Hide‹ Some key notions of situated robotics find their origin in Alan Turing’s seminal work. They emerge in both his foundation of computationalism (cognitive constraints of formalized symbolic systems) and theory of AI (bodily constraints of learning machines). I will show the deep link between these two parts of Turing’s speculation, ultimately relatable to the embedded and extended nature of the logico-symbolic practices that are deemed capable to scaffold real intelligence. Real intelligence is both structurally limited and actively mediated by embodied, cognitive, and even cultural conditions, in accord with the cognizer’s biological constitution and its historical coupling with the environment. |
||
17:00 17:30 |
Ron Chrisley | J. Mark Bishop, Slawomir J. Nasuto, Matthew Spencer, Etienne Roesch and Tomas Tanay |
The appearance of robot consciousness | Playing HeX with Aunt Hilary: games with an anthill | |
The appearance of robot consciousness ABSTRACT: ›Hide‹ This paper is a critique of some central themes in Pentti Haikonen’s recent book, Consciousness and Robot Sentience. Haikonen maintains that the crucial question is how the inner workings of the brain or an artificial system can appear, not as inner workings, but as subjective experience. It is argued here that Haikonen’s own account fails to answer this question, and that the question is not in fact the right one anyway. |
||
Playing HeX with Aunt Hilary: games with an anthill ABSTRACT: ›Hide‹ In a reflective and richly entertaining piece from 1979, Doug Hofstadter famously imagined a conversation between “Achilles” and an anthill (the eponymous “Aunt Hillary”), in which he playfully explored many ideas and themes related to cognition and consciousness. For Hofstadter, the anthill is able to carry on a conversation because the ants that compose it play roughly the same role neurons play in the brain. Unfortunately, Hofstadter ‘s work is notably short on detail suggesting how this magic might be achieved. In this paper we demonstrate how populations of simple ant-like creatures can be organised to solve complex problems; problems that involve the use of forward planning and strategy and in so doing, introduce Hofstadter's "Aunt Hilary" to the complex charms of the strategic game 'HeX'. |
||
17:30 18:00 |
Tijn Van Der Zant and Bart Verheij | Brian Cantwell Smith |
Elephants don't fly; and they know it! | Computation's Lost Promise | |
Elephants don't fly; and they know it! ABSTRACT: ›Hide‹ Neither top-down reasoning systems nor bottom-up behavior-based robotics lead to robots that can reasonably be called intelligent. Both approaches seem to work only to a certain level. In reasoning systems there can be too many possibilities to compute, in behavior-based architectures there are often not enough possibilities to finish a task. In this paper we demonstrate the early results of a bottom-up behavior-based system that uses machine learning for behavior selection. It is integrated with a reasoning architecture that learns from experience and then restructures the behavior-based architecture. This leads to robots that understand what they can and cannot do. The experiences are stored and used within a rules-with-exceptions architecture. The entire system can be described as knowledge-based adaptive robots. It is used in the international benchmark of RoboCup@Home. |
||
Computation's Lost Promise ABSTRACT: ›Hide‹ Computation's Lost Promise |
18:00-18:30 Coffee break (Buttery)