uman gaze and blinking behaviours have been recently considered, to empower humanlike robots to convey a realistic behaviour in a social human-robot interaction. This paper reports the findings of our investigation on human eye-blinking behaviour in relation to human gaze behaviour, in a human-human interaction. These findings then can be used to design a humanlike eye-blinking model for a social humanlike robot. In an experimental eye-tracking study, we showed to 11 participants, a 7-minute video of social interactions of two people, and collected their eye-blinking and gaze behaviours with an eye-tracker. Analysing the collected data, we measured information such as participants’ blinking rate, maximum and minimum blinking duration, number of frequent (multiple) blinking, as well as the participants’ gaze directions on environment. The results revealed that participants’ blinking rate in a social interaction are qualitatively correlated to the gaze behaviour, as higher number of gaze shift increased the blinking rate. Based on the findings of this study, we can propose a context-dependent blinking model as an important component of the robot’s gaze control system that can empower our robot to mimic human blinking behaviour in a multiparty social interaction.
Sensing and interpreting the interlocutor’s social behaviours is a core challenge in the development of social robots. Social robots require both an innovative sensory apparatus able to perceive the “social and emotional world” in which they act and a cognitive system able to manage this incoming sensory information and plan an organized and pondered response. In order to allow scientists to design cognitive models for this new generation of social machines, it is necessary to develop control architectures that can be easily used also by researchers without technical skills of programming such as psychologists and neuroscientists. In this work an innovative hybrid deliberative/reactive cognitive architecture for controlling a social humanoid robot is presented. Design and implementation of the overall architecture take inspiration from the human nervous system. In particular, the cognitive system is based on the Damasio’s thesis. The architecture has been preliminary tested with the FACE robot. A social behaviour has been modeled to make FACE able to properly follow a human subject during a basic social interaction task and perform facial expressions as a reaction to the social context.
The advent of humanoids has brought new challenges in the real-world application. As a part of ongoing efforts to foster functionality of the robot accommodating a real environment, this paper introduces a recent progress on a dooropeningtask with our compliant humanoid, CoMan. We presents a task-prioritized impedancecontrol framework for an upperbody system that includes a dual-arm, a waist, two soft hands, and 3D camera. Aimed to create desired responses to open the door, a novel stiffness modulation method is proposed, incorporating a realtime optimization. As a preliminary experiment, a full door-opening scenario (approaching to the door and reaching, grasping, rotating and pulling the door handle) is demonstrated under a semi-autonomous operation with a pilot. The experimental result shows the effectiveness and efficacy of the proposed impedancecontrol approach. Despite of uncertainties from sensory data, the dooropeningtask is successfully achieved and safe and robust interaction is established without creating excessive forces.
With the purpose of achieving a desired interaction performance for our complianthumanoid robot (COMAN), in this paper we propose a semi-autonomous control framework and evaluate it experimentally in a valveturning setup. The control structure consists of various modules and interfaces to identify the valve, locate the robot in front of it and perform the manipulation. The manipulation module implements four motion primitives (Reach, Grasp, Rotate and Disengage) and realizes the corresponding desired impedance profile for each phase to accomplish the task. In this direction, to establish a stable and compliant contact between the valve and the robot hands, while being able to generate the sufficient rotational torques depending on the valve's friction, Rotate incorporates a novel dual-arm impedance control technique to plan and realize a task-appropriate impedance profile. Results of the implementation of the proposed control framework are firstly evaluated in simulation studies using Gazebo. Subsequent experimental results highlight the efficiency of the proposed impedance planning and control in generation of the required interaction forces to accomplish the task.
Measuring contact forces applied by a hand to a grasped object is a necessary step to understand the mysteries that still hide in the unparalleled human grasping ability. Nevertheless, simultaneous collection of information about the position of contacts and about the magnitude and direction of forces is still an elusive task. In this paper we introduce a wearable device that addresses this problem, and can be used to measure generalized forces during grasping. By assembling two supports around a commercial 6-axis force/torque sensor we obtain a thimble that can be easily positioned on a fingertip. The device is used in conjunction with an active marker-based motion capture system to simultaneously obtain absolute position and orientation of the thimbles, without requiring any assumptions on the kinematics of the hand. Finally, using the contact centroid algorithm, introduced in [1], position of contact points during grasping are determined. This paper shows the design and implementation of the device, as well as some preliminary experimental validation.