poster - Institute For Systems and Robotics

Report 4 Downloads 90 Views
Generation of Meaningful Robot Expressions with Active Learning Giovanni Saponaro, Alexandre Bernardino VisLab, Institute for Systems and Robotics, Instituto Superior Técnico, Lisbon, Portugal

Computer Vision Laboratory

PÓLO DO I.S.T

Active Learning

Overview



We address the problem of communicating emotions with a humanoid robot merely with its body joint movements, without facial information, querying human users for emotional scores attributed to the movements. Machine learning can help convey the intended emotions more clearly, by selecting the next actions and parameters that need tuning and rewarding successful action–emotion matches.

Motivation • Transmit emotions to human observers with the iCub humanoid robot • Explore expressivity (or lack thereof) of body movements when facial expressions are disabled

Setup and Proposed Approach Each combination is a simple Bayesian Network with an Action and an Emotion node:

Library of Movements select movement with max entropy gain

θ11 θ12 P (E = el |A = an) = [θln] = . . θ1A                 

                

• Action–emotion matches modelled as a multinomial distribution • Rows: actions (with a discretized joint parameter), columns: emotions, Σl θln = 1 ∀ action • Active: learner selects the row yielding highest entropy gain as next query, we update the

parameters of that row P (E=el |A=an) P (E=el |A=an)+s • ← : update step, where #an is the number of trials performed #an #an+1 with A = an up to this point and s is the Likert score resulting from the current trial answer (normalized to a probability) • The framework is scalable to use a matrix with hundres of rows, where clusters of rows encode the same action with different joint parameters, and can be awarded/penalized

Results • Head actions are associated to emotions clearly, arm gestures are often ambiguous • Unclear actions are chosen and displayed more times than clear ones, as reflected in the

higher denominators of match scores

0.8

0.6

match score

human scores E

Active Learning

θ11=0.87/2

uniform θln=0.2/1

0.8

4/5 θ41=0.53/2

0.6

0.4 0.2

0.2

nod

punch

punch

look out disapproval success distraction

look out disapproval success distraction

thumbs up thumbs down

anger agreement

Action

5.02/10

0.2

3.95/18

0.4 0.2

6.45/23 0.4 0.2

0

0

nod

nod

nod

punch

punch

look out

Action

disapproval success distraction

look out disapproval success distraction

thumbs up thumbs down

anger agreement

Action

thumbs down

Emotion

Entropy evolution for the "nod" action

Emotion

Entropy evolution for the "punch" action

Entropy evolution for the "look out" action

1.8

1.8

1.6

1.6

1.6

1.4

1.4

1.4

1.2

1.2

1.2

1

1

1

0.8

0.8

0.8

0.6

0.6

0.6

0.4

0.4

0.4

0.2

0.2

0.2

1

5

10

15

anger agreement

Action

1.8

init

disapproval success distraction

thumbs up

anger agreement

Emotion

Figure: The iCub performing the “thumbs down” gesture.

20

0

init

Number of interviewed subjects

1

5

10

15

20

Number of interviewed subjects

Entropy evolution for the "thumbs up" action

Meaning initial position of first neck joint final position of first neck joint initial velocity of first neck joint time to transition from (0) to (1) final velocity of first neck joint time to transition from (1) to (0)

4.51/18

0

thumbs down

Parameter (0) x0 (1) x0 (0) x˙ 0 t0→1 (1) x˙ 0 t1→0

θ14=6.46/23

0.6

Score

Score

Score

2.97/12

thumbs up

Table: Parameters of the “nod” action

12.08/20 12.07/19 0.8

0.6

look out

Description Expected perceived emotion el head tilts up and down agreement rapidly extend fist in front of robot anger abruptly deviate robot head and gaze to a side distraction show fist and move thumb up success show fist and move thumb down disapproval

Matches after interviewing 20 people

0.8

punch

Table: Library of robot action movements

Emotion

θ42=4.77/18

3/12

anger agreement

Action

Matches after interviewing 15 people

0.6 0.4

thumbs down

anger agreement

6.28/10 7.8/11

disapproval success distraction

thumbs up

Emotion

Matches after interviewing 10 people

0.8

Example Library of Movements and Parameters

0.2

nod

Emotion

show (the one requiring more correction, i.e., with most entropy) • Active learner can also choose an action movement parameter and a discretized value to fine-tune

0.4

nod

look out

0

1.34/7

0

Action

• Input human opinions into an active machine learning program that selects the next action to

0.6

0

thumbs down

Figure: Human users being interviewed about what emotion they perceive from robot movements.

3.34/6

θ33=θ44=1.2/2

0

thumbs up

• Survey human users about what they feel when the robot performs said movements

2.9/6

0.8

0.4

punch

• Design a library of pre-defined movements

Matches after interviewing 5 people

Score

A

Matches after interviewing 1 person

Score

tuned movement parameters

Initial matches

Score

Questionnaire

Action an nod punch look out thumbs up thumbs down

· · · θL1 · · · θL2 . . . .. · · · θLA



0

init

1

5

10

15

20

Number of interviewed subjects

Entropy evolution for the "thumbs down" action

1.8

1.8

1.6

1.6

1.4

1.4

1.2

1.2

1

1

0.8

0.8

0.6

0.6

0.4

0.4

0.2

0.2

0

0

Entropy Posterior entropy Entropy gain init

1

5

10

15

Number of interviewed subjects

20

init

1

5

10

15

20

Number of interviewed subjects

Future Work • Improve corpus of robot movements, explore using other limbs (legs), systematically analyze

the human affective experience with or without robot facial expressions • Study the degree of robot emotional expressivity due to its appearance vs the mechanical range of permitted movements • Try different parameter initialization: uniform vs expert knowledge • Explore optimization strategies other than maximum-entropy criterion: artificial neural networks, reinforcement learning

Summary We model a mapping between robot movements (with no facial expressions) and emotions perceived by human users. Match parameters are updated with a questionnaire, and as correspondences take shape we can observe (i) which body joint actions are expressive, (ii) which actions should be performed by a robot in order to transmit a given emotion. As opposed to random action selection or querying for a fixed sequence of actions, the Active Learning framework gives the system the ability to inquire about movements that are ambiguous, showing them more often than easily-perceived ones.

User Survey • Human subjects see robot actions and respond to a Likert questionnaire • After each robot movement, a user is asked to evaluate statements such as “This action

expresses anger” with a score among [strongly disagree] [disagree] [neither agree nor disagree] [agree] [strongly agree] • Responses are sent as a probability vector to the Active Learning module, which updates action–emotion parameters • We interviewed 20 people (non-roboticists): 50% males, 50% females, mean age 29.7, stddev 4.49 • Type of study: within-subject with self-assessment evaluation

References [1] J. Li and M. Chignell. Communication of Emotion in Social Robots through Simple Head and Arm Movements. International Journal of Social Robotics, 2010. [2] S. Tong and D. Koller. Active Learning for Parameter Estimation in Bayesian Networks. In T. K. Leen, T. G. Dietterich, and V. Tresp, editors, Neural Information Processing Systems (NIPS), pages 647–653. MIT Press, 2000. This work was supported by Fundação para a Ciência e a Tecnologia (ISR/IST pluriannual funding and grant SFRH/BD/61910/2009).