View Chapter

Chapter 26 — Flying Robots

Stefan Leutenegger, Christoph Hürzeler, Amanda K. Stowers, Kostas Alexis, Markus W. Achtelik, David Lentink, Paul Y. Oh and Roland Siegwart

Unmanned aircraft systems (UASs) have drawn increasing attention recently, owing to advancements in related research, technology, and applications. While having been deployed successfully in military scenarios for decades, civil use cases have lately been tackled by the robotics research community.

This chapter overviews the core elements of this highly interdisciplinary field; the reader is guided through the design process of aerial robots for various applications starting with a qualitative characterization of different types of UAS. Design and modeling are closely related, forming a typically iterative process of drafting and analyzing the related properties. Therefore, we overview aerodynamics and dynamics, as well as their application to fixed-wing, rotary-wing, and flapping-wing UAS, including related analytical tools and practical guidelines. Respecting use-case-specific requirements and core autonomous robot demands, we finally provide guidelines to related system integration challenges.

Flight stability in aerial redundant manipulators

Author  Christopher Korpela, Matko Orsag, Todd Danko, Bryan Kobe, Clayton McNeil, Robert Pisch, Paul Oh

Video ID : 693

Aerial manipulation tests conducted by the Drexel Autonomous Systems Lab.

Chapter 76 — Evolutionary Robotics

Stefano Nolfi, Josh Bongard, Phil Husbands and Dario Floreano

Evolutionary Robotics is a method for automatically generating artificial brains and morphologies of autonomous robots. This approach is useful both for investigating the design space of robotic applications and for testing scientific hypotheses of biological mechanisms and processes. In this chapter we provide an overview of methods and results of Evolutionary Robotics with robots of different shapes, dimensions, and operation features. We consider both simulated and physical robots with special consideration to the transfer between the two worlds.

More complex robots evolve in more complex environments

Author  Josh Bongard

Video ID : 772

This set of videos demonstrates that complex environments influence the evolution of robots with more complex body plans.

Chapter 13 — Behavior-Based Systems

François Michaud and Monica Nicolescu

Nature is filled with examples of autonomous creatures capable of dealing with the diversity, unpredictability, and rapidly changing conditions of the real world. Such creatures must make decisions and take actions based on incomplete perception, time constraints, limited knowledge about the world, cognition, reasoning and physical capabilities, in uncontrolled conditions and with very limited cues about the intent of others. Consequently, one way of evaluating intelligence is based on the creature’s ability to make the most of what it has available to handle the complexities of the real world. The main objective of this chapter is to explain behavior-based systems and their use in autonomous control problems and applications. The chapter is organized as follows. Section 13.1 overviews robot control, introducing behavior-based systems in relation to other established approaches to robot control. Section 13.2 follows by outlining the basic principles of behavior-based systems that make them distinct from other types of robot control architectures. The concept of basis behaviors, the means of modularizing behavior-based systems, is presented in Sect. 13.3. Section 13.4 describes how behaviors are used as building blocks for creating representations for use by behavior-based systems, enabling the robot to reason about the world and about itself in that world. Section 13.5 presents several different classes of learning methods for behavior-based systems, validated on single-robot and multirobot systems. Section 13.6 provides an overview of various robotics problems and application domains that have successfully been addressed or are currently being studied with behavior-based control. Finally, Sect. 13.7 concludes the chapter.

Experience-based learning of high-level task representations: Demonstration (3)

Author  Monica Nicolescu

Video ID : 32

This is a video recorded in early 2000s, showing a Pioneer robot learning to traverse "gates" and move objects from a source place to a destination - the human demonstration stage. The robot execution stage is also shown in a related video in this chapter. Reference: M. Nicolescu, M.J. Mataric: Learning and interacting in human-robot domains, IEEE Trans. Syst. Man Cybernet. A31(5), 419-430 (2001)

Chapter 34 — Visual Servoing

François Chaumette, Seth Hutchinson and Peter Corke

This chapter introduces visual servo control, using computer vision data in the servo loop to control the motion of a robot. We first describe the basic techniques that are by now well established in the field. We give a general overview of the formulation of the visual servo control problem, and describe the two archetypal visual servo control schemes: image-based and pose-based visual servo control. We then discuss performance and stability issues that pertain to these two schemes, motivating advanced techniques. Of the many advanced techniques that have been developed, we discuss 2.5-D, hybrid, partitioned, and switched approaches. Having covered a variety of control schemes, we deal with target tracking and controlling motion directly in the joint space and extensions to under-actuated ground and aerial robots. We conclude by describing applications of visual servoing in robotics.

IBVS on a 6-DOF robot arm (1)

Author  Francois Chaumette, Seth Hutchinson, Peter Corke

Video ID : 59

This video shows an IBVS on a 6-DOF robot arm with Cartesian coordinates of image points as visual features and a desired interaction matrix in the control scheme. It corresponds to the results depicted in Figure 34.2.

Chapter 26 — Flying Robots

Stefan Leutenegger, Christoph Hürzeler, Amanda K. Stowers, Kostas Alexis, Markus W. Achtelik, David Lentink, Paul Y. Oh and Roland Siegwart

Unmanned aircraft systems (UASs) have drawn increasing attention recently, owing to advancements in related research, technology, and applications. While having been deployed successfully in military scenarios for decades, civil use cases have lately been tackled by the robotics research community.

This chapter overviews the core elements of this highly interdisciplinary field; the reader is guided through the design process of aerial robots for various applications starting with a qualitative characterization of different types of UAS. Design and modeling are closely related, forming a typically iterative process of drafting and analyzing the related properties. Therefore, we overview aerodynamics and dynamics, as well as their application to fixed-wing, rotary-wing, and flapping-wing UAS, including related analytical tools and practical guidelines. Respecting use-case-specific requirements and core autonomous robot demands, we finally provide guidelines to related system integration challenges.

Towards valve turning using a dual-arm aerial manipulator

Author  Christopher Korpela, Matko Orsag, Paul Oh, Stjepan Bogdan

Video ID : 719

A framework was proposed for valve turning using an aerial vehicle endowed with dual multi-degree of freedom manipulators. A tightly integrated control scheme between the aircraft and manipulators is mandated for tasks requiring aircraft-to-environment coupling. Feature detection is well-established for both ground and aerial vehicles and facilitates valve detection and arm tracking. Force feedback upon contact with the environment provides compliant motions in the presence of position error and coupling with the valve. The video presents results validating the valve turning framework using the proposed aircraft-arm system during flight tests.

Chapter 40 — Mobility and Manipulation

Oliver Brock, Jaeheung Park and Marc Toussaint

Mobile manipulation requires the integration of methodologies from all aspects of robotics. Instead of tackling each aspect in isolation,mobilemanipulation research exploits their interdependence to solve challenging problems. As a result, novel views of long-standing problems emerge. In this chapter, we present these emerging views in the areas of grasping, control, motion generation, learning, and perception. All of these areas must address the shared challenges of high-dimensionality, uncertainty, and task variability. The section on grasping and manipulation describes a trend towards actively leveraging contact and physical and dynamic interactions between hand, object, and environment. Research in control addresses the challenges of appropriately coupling mobility and manipulation. The field of motion generation increasingly blurs the boundaries between control and planning, leading to task-consistent motion in high-dimensional configuration spaces, even in dynamic and partially unknown environments. A key challenge of learning formobilemanipulation consists of identifying the appropriate priors, and we survey recent learning approaches to perception, grasping, motion, and manipulation. Finally, a discussion of promising methods in perception shows how concepts and methods from navigation and active perception are applied.

Autonomous robot skill acquisition

Author  Scott Kuindersma, George Konidaris

Video ID : 669

This video demonstrates the autonomous-skill acquisition of a robot acting in a constrained environment called the "Red Room". The environment consists of buttons, levers, and switches, all located at points of interest designated by ARTags. The robot can navigate to these locations and perform primitive manipulation actions, some of which affect the physical state of the maze (e.g., by opening or closing a door).

Chapter 18 — Parallel Mechanisms

Jean-Pierre Merlet, Clément Gosselin and Tian Huang

This chapter presents an introduction to the kinematics and dynamics of parallel mechanisms, also referred to as parallel robots. As opposed to classical serial manipulators, the kinematic architecture of parallel robots includes closed-loop kinematic chains. As a consequence, their analysis differs considerably from that of their serial counterparts. This chapter aims at presenting the fundamental formulations and techniques used in their analysis.

3-DOF dynamically balanced parallel robot

Author  Clément Gosselin

Video ID : 49

This video demonstrates a 3-DOF dynamically balanced parallel robot. References: 1. S. Foucault, C. Gosselin: On the development of a planar 3-DOF reactionless parallel mechanism, Proc. ASME Mech. Robot. Conf., Montréal (2002); 2. Y. Wu, C. Gosselin: Synthesis of reactionless spatial 3-DOFf and 6-DOF mechanisms without separate counter-rotations, Int. J. Robot. Res. 23(6), 625-642 (2004)

Chapter 53 — Multiple Mobile Robot Systems

Lynne E. Parker, Daniela Rus and Gaurav S. Sukhatme

Within the context of multiple mobile, and networked robot systems, this chapter explores the current state of the art. After a brief introduction, we first examine architectures for multirobot cooperation, exploring the alternative approaches that have been developed. Next, we explore communications issues and their impact on multirobot teams in Sect. 53.3, followed by a discussion of networked mobile robots in Sect. 53.4. Following this we discuss swarm robot systems in Sect. 53.5 and modular robot systems in Sect. 53.6. While swarm and modular systems typically assume large numbers of homogeneous robots, other types of multirobot systems include heterogeneous robots. We therefore next discuss heterogeneity in cooperative robot teams in Sect. 53.7. Once robot teams allow for individual heterogeneity, issues of task allocation become important; Sect. 53.8 therefore discusses common approaches to task allocation. Section 53.9 discusses the challenges of multirobot learning, and some representative approaches. We outline some of the typical application domains which serve as test beds for multirobot systems research in Sect. 53.10. Finally, we conclude in Sect. 53.11 with some summary remarks and suggestions for further reading.

CKBOTS reconfigurable robots

Author  Mark Yim

Video ID : 196

This video shows reconfigurable robots, which are capable of a variety of configurations and modes of locomotion, including bipeds that can stand up and walk. This system is robust in a variety of situations, as shown in the video. The system has three clusters: when clusters disconnect, they enter a search mode and approach each other to assemble. After successful self-reassembling, the robot system stands up to continue its task.

Chapter 72 — Social Robotics

Cynthia Breazeal, Kerstin Dautenhahn and Takayuki Kanda

This chapter surveys some of the principal research trends in Social Robotics and its application to human–robot interaction (HRI). Social (or Sociable) robots are designed to interact with people in a natural, interpersonal manner – often to achieve positive outcomes in diverse applications such as education, health, quality of life, entertainment, communication, and tasks requiring collaborative teamwork. The long-term goal of creating social robots that are competent and capable partners for people is quite a challenging task. They will need to be able to communicate naturally with people using both verbal and nonverbal signals. They will need to engage us not only on a cognitive level, but on an emotional level as well in order to provide effective social and task-related support to people. They will need a wide range of socialcognitive skills and a theory of other minds to understand human behavior, and to be intuitively understood by people. A deep understanding of human intelligence and behavior across multiple dimensions (i. e., cognitive, affective, physical, social, etc.) is necessary in order to design robots that can successfully play a beneficial role in the daily lives of people. This requires a multidisciplinary approach where the design of social robot technologies and methodologies are informed by robotics, artificial intelligence, psychology, neuroscience, human factors, design, anthropology, and more.

A scene of deictic interaction

Author  Takayuki Kanda

Video ID : 807

This video illustrates the "deictic interaction" in which the robot and a user interact using pointing gestures and verbal-reference terms. The robot has a capability to understand the user's deictic interaction recognizing both the pointing gesture and the reference term. In addition, there is a 'facilitation' mechanism (e.g., the robot engages in real-time joint attention), which makes the interaction smooth and natural.

Chapter 74 — Learning from Humans

Aude G. Billard, Sylvain Calinon and Rüdiger Dillmann

This chapter surveys the main approaches developed to date to endow robots with the ability to learn from human guidance. The field is best known as robot programming by demonstration, robot learning from/by demonstration, apprenticeship learning and imitation learning. We start with a brief historical overview of the field. We then summarize the various approaches taken to solve four main questions: when, what, who and when to imitate. We emphasize the importance of choosing well the interface and the channels used to convey the demonstrations, with an eye on interfaces providing force control and force feedback. We then review algorithmic approaches to model skills individually and as a compound and algorithms that combine learning from human guidance with reinforcement learning. We close with a look on the use of language to guide teaching and a list of open issues.

Learning from failure II

Author  Aude Billard

Video ID : 477

This video illustrates in a second example how learning from demonstration can benefit from failed demonstrations (as opposed to learning from successful demonstrations). Here, the robot Robota must learn how to coordinate its two arms in a timely manner for the left arm to hit the ball with the racket right on time, after the left arm sent the ball flying by hitting the catapult. More details on this work is available in: A. Rai, G. de Chambrier, A. Billard: Learning from failed demonstrations in unreliable systems, Proc. IEEE-RAS Int. Conf. Humanoid Robots (Humanoids), Atlanta (2013), pp. 410 – 416; doi: 10.1109/HUMANOIDS.2013.7030007 .