23
Introduction Techniques and Approach The Domestic Scenario The Manufacturing Scenario Conclusions Verification and Validation of Robotic Assistants Clare Dixon Department of Computer Science University of Liverpool 1 University of Liverpool (UoL) 2 University of Hertfordshire (UoH) 3 Bristol Robotics Lab (BRL) www.robosafe.org Farshid Amirabdollahian 2 Kerstin Dautenhahn 2 Anthony Pipe 3 Kerstin Eder 3 Maha Salem 2 Michael Fisher 1 Joe Saunders 2 Dejanira Araiza Illan 3 Matt Webster 1 Kheng Lee Koay 2 David Western 3 Clare Dixon Verification and Validation Robotic Assistants 1 / 23

Verification and Validation of Robotic Assistants

  • Upload
    adacore

  • View
    208

  • Download
    1

Embed Size (px)

Citation preview

Introduction Techniques and Approach The Domestic Scenario The Manufacturing Scenario Conclusions

Verification and Validation of Robotic Assistants

Clare Dixon

Department of Computer ScienceUniversity of Liverpool

1 University of Liverpool (UoL)2 University of Hertfordshire (UoH)3 Bristol Robotics Lab (BRL)

www.robosafe.org

Farshid Amirabdollahian2

Kerstin Dautenhahn2 Anthony Pipe3

Kerstin Eder3 Maha Salem2

Michael Fisher1 Joe Saunders2

Dejanira Araiza Illan3 Matt Webster1

Kheng Lee Koay2 David Western3

Clare Dixon Verification and Validation Robotic Assistants 1 / 23

Introduction Techniques and Approach The Domestic Scenario The Manufacturing Scenario Conclusions

Robotic Assistants

Robotic assistants are being developed tohelp, or work closely with humans inindustrial, domestic and health careenvironments (e.g. RI-MAN, Pearl,Wakamaru, . . . )The robots will need to be able to actautonomously and make decisions tochoose between a range of activities.In addition they will need to operate closeto, or in collaboration with humans.How do we make sure they are trustworthy,safe, reliable and do what they aresupposed to?

Wakamaru image by Nesnad (Own work) [CC BY-SA 3.0 (http://creativecommons.org/licenses/by-sa/3.0) or GFDL(http://www.gnu.org/copyleft/fdl.html)], via Wikimedia Commons

Clare Dixon Verification and Validation Robotic Assistants 2 / 23

Introduction Techniques and Approach The Domestic Scenario The Manufacturing Scenario Conclusions

What is Trustworthiness and Safety?

Safety involves showing that the robot does nothing that(unnecessarily) endangers the person.There are ISO safety requirements and guidelines forindustrial robots (ISO 10218, 2011), personal care robots(ISO 13482, 2014), and for collaborative robots (ISO15066, 2016).Trustworthiness involves social issues beyond pure safety.It is not just a question of whether the robots are safe butwhether they are perceived to be safe, useful and reliable.There are also legal (and ethical) issues such as whathappens when

the robot spills a hot drink on someone;the robot doesn’t remind the person to take their medicine;the robot doesn’t go to the kitchen when told?

Clare Dixon Verification and Validation Robotic Assistants 3 / 23

Introduction Techniques and Approach The Domestic Scenario The Manufacturing Scenario Conclusions

Robots in the Workplace and at Home

Currently many robots used in industry or domestic use operatein limited physical space or have limited functionality. This helpsassure their safety.

Robots’ industrial environments are limited so they canonly move in a fixed area and have limited interactions withhumans e.g. welding or paint spraying robots.Small or limited capability domestic robots, e.g., vacuumcleaning robots, robot lawn mowers, pool cleaning robotsetc

Clare Dixon Verification and Validation Robotic Assistants 4 / 23

Introduction Techniques and Approach The Domestic Scenario The Manufacturing Scenario Conclusions

Trustworthy Robotic Assistants Project

The EPSRC funded Trustworthy Robotic Assistants Projectdevelops three different approaches to verification andvalidation of robotic assistants.Each approach is aimed at increasing trust in robotic assistants.

Formal Verification (Liverpool)Simulation-based Testing (Bristol Robotics Laboratory)End-user Validation (Hertfordshire)

Clare Dixon Verification and Validation Robotic Assistants 5 / 23

Introduction Techniques and Approach The Domestic Scenario The Manufacturing Scenario Conclusions

Robotic Assistants

We consider two use cases, domesticand manufacturing.

A personal robot assistant (the Care-O-bot R©,) located in a domestic housein the University of Hertfordshire.

A co-operative manufacturing taskwith BERT a robot at Bristol RoboticsLab.

8 Journal Title XX(X)

• System model inaccuracies. All the verificationtechniques use models of the real-world. The modelsmight have been constructed erroneously, or may beinconsistent with the real world, or relative to oneanother.

• Requirement model inaccuracies. In our approach,the real-world requirements of the system are con-verted into textual requirements, assertions and prop-erties for verification. These requirements modelsmay not have been correctly formulated.

• Tool inaccuracies. It is possible that numericalapproximations affect the verification results. Inaddition, third party tools can contain bugs that areunknown to us.

We could now proceed to perform “Experiments.” Asbefore, we may find a problem with the textual require-ments or the physical system during experimentation. Atthe same time, the assurances from formal verificationand/or simulation-based testing can be compared againstthe experiment results. We may also discover that one of theassurances holds during simulation-based testing or formalverification, but not during the experiments. In this case wemay need to refine any of the other assets, as explainedbefore.

Careful comparisons must be made between the dif-ferent representations in order to discover the cause ofthe assurance conflicts. Such comparisons are indicatedby the bi-directional arrows between “Formal Verification”and “Simulation-based Testing”, “Simulation-based Test-ing” and “Experiments”, and “Formal Verification” and“Experiments”, respectively, in Figure 1.

4 The BERT Handover Task: A Case StudyIn this section, we present a case study to demonstratethe application of assurance-based verification to an HRIscenario considering the following research question: canassurance-based verification provide a higher degree ofconfidence in the resulting assurances than when usingverification techniques in isolation?

BERT 2 is an upper-body humanoid robot designed tofacilitate research into complex human-robot interactions,including verbal and non-verbal communication, such asgaze and physical gestures (Lenz et al. 2010) (see Figure 2).BERT 2’s software architecture was originally developedusing YARP¶. More recently, this system has been wrappedwith a ROS interface.

We verify an object handover to exemplify our approach,in the context of a broader collaborative manufacturescenario where BERT 2 and a person work together toassemble a table (Lenz et al. 2012). In the handover, thefirst step is an activation signal from the human to the

Figure 2. BERT 2 engaged in the handover task.

robot. BERT 2 then picks up a nearby object, and holdsit out to the human. The robot announces that it is readyto handover. The human responds verbally to indicate thatthey are ready to receive. (For practical reasons, human-to-robot verbal signals were relayed to the robot by a humanoperator pressing a key.) Then, the human is expected topull gently on the object while looking at it. BERT 2 thencalculates three binary sensor conditions:

• Gaze: The human’s head position and orientationrelative to the object are tracked using the Vicon R�

motion-tracking system for an approximate measureof whether he/she is looking at the object.

• Pressure: Changes in the robot’s finger positionsare sensed to detect whether the human is applyingpressure to take the weight of the object.

• Location: The Vicon R� motion-tracking system isused to determine whether the human’s hand islocated on the object.

The sensor conditions must be calculated within a timethreshold for BERT 2 to determine if the human “is ready”.The robot should release its grip on the object if allthree conditions are satisfied. Otherwise, the robot shouldterminate the handover and not release the object. Thehuman may disengage and the robot can timeout, whichwould cancel the remainder of the handover task. Thesensors are not completely accurate and may sometimesgive incorrect readings.

A safety requirement ensures that “nothing badhappens”, whereas a liveness requirement ensures that“something good happens eventually” or inside a threshold

¶http://www.yarp.it

Prepared using sagej.cls

Clare Dixon Verification and Validation Robotic Assistants 6 / 23

Introduction Techniques and Approach The Domestic Scenario The Manufacturing Scenario Conclusions

Formal Verification

A mathematical analysis of all behaviours using logics, andtools such as theorem provers or model checkers.We focus on temporal verification using automatic toolsand techniques that do not require user interaction.Model checking is a fully automatic, algorithmic techniquefor verifying the temporal properties of systems.Input to the model checker is a model of the system and aproperty to be checked on that model.Output is that the property is satisfied or a counterexample is given.

Model Checker

Property holds

or

counter example

Property eg

"always p"

Clare Dixon Verification and Validation Robotic Assistants 7 / 23

Introduction Techniques and Approach The Domestic Scenario The Manufacturing Scenario Conclusions

Robot Architectures and Verification

We assume an architecture where there is a separationbetween the high level decision making layer and the low levelcontrol layer.

etc

Control System

Sense and act

High level choices

Rational Agent

Low level control

Decision making

Avoidance

Reactive

Goal selection

Plan selection

Predictionetc

We aim to represent and verify the decision making layer andwe don’t deal with low level control such as movement etc.

Clare Dixon Verification and Validation Robotic Assistants 8 / 23

Introduction Techniques and Approach The Domestic Scenario The Manufacturing Scenario Conclusions

Simulation Based Testing

This is an exhaustive testing methodology widely used inthe design of micro-electronic and avionics systems.These appeal to Monte-Carlo techniques and dynamic testrefinement in order to cover a wide range or practicalsituations.Tools are used to automate the testing and analyse thecoverage of the tests.HRI Handover Scenario

27

Requirements: !  Functional and safety (ISO 13482:2014, ISO 10218-1) Clare Dixon Verification and Validation Robotic Assistants 9 / 23

Introduction Techniques and Approach The Domestic Scenario The Manufacturing Scenario Conclusions

End User Validation

This approach involves experiments and user evaluationsin practical robotic scenarios.Scenarios relating to robot human interaction aredeveloped to test some hypothesis and experiments withusers carried out.This helps establish whether the human participantsindeed view the robotic assistants as safe and trustworthy.

Clare Dixon Verification and Validation Robotic Assistants 10 / 23

Introduction Techniques and Approach The Domestic Scenario The Manufacturing Scenario Conclusions

Overall Approach

Clare Dixon Verification and Validation Robotic Assistants 11 / 23

Introduction Techniques and Approach The Domestic Scenario The Manufacturing Scenario Conclusions

A Domestic Robot AssistantHere we apply model checking to thehigh level behaviours controlling the(commercially available) Care-O-bot R©,manufactured by Fraunhofer IPA.It is based on the concept of a “robotbutler” which has been developed as amobile robotic assistant to supportpeople in domestic environments.It has a manipulator arm, an articulatedtorso, stereo sensors serving as “eyes”,LED lights, a graphical user interface,and a moveable tray.The robot’s sensors monitor its current location, the stateof the arm, torso, eyes and tray.Its software is based on the Robot Operating System.

Clare Dixon Verification and Validation Robotic Assistants 12 / 23

Introduction Techniques and Approach The Domestic Scenario The Manufacturing Scenario Conclusions

Care-O-bot and Robot House

This is deployed in a domestic-type house (the robothouse) at the University of Hertfordshire.The robot house is equipped with sensors which provideinformation on the state of the house and its occupants,such as whether the fridge door is open and whethersomeone is seated on the sofa.Low-level robot actions such as movement, speech, lightdisplay, etc., are controlled by groups of high-level rulesthat together define particular behaviours.

3

Fig. 2. A plan view of the ground floor of the University of Hertfordshire Robot House. Numbered boxes show the locations of sensors.

models, and their formal verification, are described inSection IV.

• Figs. 2 and 3 have been added to provide additionalinformation on the Robot House and the user activitywithin it.

• Section V on related work has been updated, and Sec-tion VI on conclusions and future work has been ex-tended.

II. MODELLING THE CARE-O-BOT USING BRAHMS

The autonomous decision making within the Robot Houseand Care-O-bot R� at the University of Hertfordshire is carriedout by a high-level planning/scheduling system described inthe previous section. The code base includes a database of 31default rules for the Robot House and Care-O-bot to follow.Careful examination of these rules revealed that they aresimilar in structure to the various constructs within the Brahmsmulti-agent workflow programming language.

The first step in modelling was to convert the full set ofCare-O-bot rules into a more convenient if-then rule repre-sentation. For example, the rule in the previous section wasrewritten as:

IF tray_is_raised AND tray_is_emptyTHEN set_light(yellow)

move_tray_and_wait(lowered_position)set_light(white)wait()set(tray_is_raised,false)set(tray_is_lowered,true)

Once translated into this format, these rules could then bestraightforwardly translated into Brahms. A key concept inBrahms is the ‘workframe’, which specifies a sequence ofthings to be done when a given condition holds. The RobotHouse rules were translated into Brahms workframes withinthe Care-O-bot agent, with the IF a THEN b rules trans-lated into the when a do { b } construct in Brahms.For example, the rule above was translated into a Brahmsworkframe called wf_lowerTray:

workframe wf_lowerTray {repeat: true;priority: 10;

when(knownval(current.trayIsRaised = true)andknownval(current.trayIsEmpty = true))

do{conclude((current.lightColour =

current.colourYellow));lowerTrayAndWait();conclude((current.lightColour =

current.colourWhite));waitForLightColourChange();

Clare Dixon Verification and Validation Robotic Assistants 13 / 23

Introduction Techniques and Approach The Domestic Scenario The Manufacturing Scenario Conclusions

Care-O-bot Decision Making: Behaviours

The Care-O-bot’s high-level decision making is determinedby a set of behaviours of the form precondition → action(each a sequence of rules).Examples of high-level rules can take the form “lower tray”,“move to sofa area of the living room”, “say ‘The fridgedoor is open’ ”, set a flag, check a sensor etc.Only one behaviour executes at once.Each behaviour has a priority (integer between 0 and 90).Higher priority behaviours are executed in preference tolower priority behaviours.Each behaviour is flagged as interruptible or not.Once it has started executing, a behaviour will execute tocompletion, if it is not interruptible.Users can add new behaviours.

Clare Dixon Verification and Validation Robotic Assistants 14 / 23

Introduction Techniques and Approach The Domestic Scenario The Manufacturing Scenario Conclusions

The S1-alertFridgeDoor Behaviour

Behaviours (a set of high level rules) take the form:

Precondition-Rules -> Action-Rules

27 Fridge Freezer Is *ON* AND has been ON for more than 30 secs31 ::514:: GOAL-fridgeUserAlerted is false

32 Turn light on ::0::Care-o-Bot 3.2 to yellow34 move ::0::Care-o-Bot 3.2 to ::2:: Living Room and wait for

completion35 Turn light on ::0::Care-o-Bot 3.2 to white and wait for

completion36 ::0::Care-o-Bot 3.2 says ‘The fridge door is open!’ and

wait for completion37 SET ::506::GOAL-gotoCharger TO false38 SET ::507::GOAL-gotoTable TO false39 SET ::508::GOAL-gotoSofa TO false40 ::0::Care-o-Bot 3.2 GUI, S1-Set-GoToKitchen, S1-Set-WaitHere41 SET ::514::GOAL-fridgeUserAlerted TO true

Its priority is 60 and it is not interruptible.Clare Dixon Verification and Validation Robotic Assistants 15 / 23

Introduction Techniques and Approach The Domestic Scenario The Manufacturing Scenario Conclusions

Models and Properties

We need to abstract away from some of the timing detailsincluded in the database to obtain a model that is discrete,finite and not too large.We developed a (by hand) model in the input language forthe model checker NuSMV and later developed a tool(CRutoN) to automatically translate from behaviours toNuSMV input.We also need a set of properties of the system to checkover the model.Ideally these would come from a specification or standardsdocuments about what is expected of the robot withrespect to functionality, safety etc.Here we focus on issues relating to the scheduling ofbehaviours, priorities and interruptions (which at leastprovide a sanity check).

Clare Dixon Verification and Validation Robotic Assistants 16 / 23

Introduction Techniques and Approach The Domestic Scenario The Manufacturing Scenario Conclusions

Sample Properties and Model Checking Results

1 ((fridge_freezer_on ∧ ¬goal_fridge_user_alerted)⇒♦(location = livingroom ∧♦say = fridge_door_open))

2 ((fridge_freezer_on ∧ ¬goal_fridge_user_alerted ∧schedule = schedule_alert_fridge_door)⇒♦(location = livingroom ∧♦say = fridge_door_open))

Property Output Time (sec)1 FALSE 11.12 TRUE 12.3

The model had 130,593 reachable states.We did find a small bug in the behaviours (a flag waswrongly set) but this was by inspection of the behaviours.It would be better to try properties relating to therequirements of the robot.

Clare Dixon Verification and Validation Robotic Assistants 17 / 23

Introduction Techniques and Approach The Domestic Scenario The Manufacturing Scenario Conclusions

Discussion

CRutoN allowed us to translate from different databases ofbehaviours into input for a model checker, settingparameters to control particular aspects of the translation.CRutoN uses an intermediate representation so that inputto different model checkers can potentially be generated.Understanding the semantics of the robot execution cycletook a lot of close work and interaction with UoH.The state explosion problem means we have to find abalance between the level of detail/abstraction andverification times (timing details were not well represented).We could deal better with uncertainty or timing constraintsby applying a different model checker.The model of a person in the robot house was notrepresented but this could be incorporated showing theirlocation for example.

Clare Dixon Verification and Validation Robotic Assistants 18 / 23

Introduction Techniques and Approach The Domestic Scenario The Manufacturing Scenario Conclusions

Experiments with Trust and Reliability

In the robot house UoH experimented using two scenarioswhere the robot appeared faulty or not.

In both scenarios the person was asked to carry out a task withthe robot.

Results suggested that although errors in a robot’s behavior arelikely to affect participant’s perception of its reliability andtrustworthiness, this doesn’t seem to influence their decisionsto comply with instructions (or not).

Their willingness to comply with therobot’s instructions seem to dependon the nature if the task, in particular,whether its effects are irrevocable.

Clare Dixon Verification and Validation Robotic Assistants 19 / 23

Introduction Techniques and Approach The Domestic Scenario The Manufacturing Scenario Conclusions

The Manufacturing Scenario: Verification

The focus was on a table leg handover task. The gaze, handlocation and hand pressure of the human should be correctbefore the handover takes place.

8 Journal Title XX(X)

• System model inaccuracies. All the verificationtechniques use models of the real-world. The modelsmight have been constructed erroneously, or may beinconsistent with the real world, or relative to oneanother.

• Requirement model inaccuracies. In our approach,the real-world requirements of the system are con-verted into textual requirements, assertions and prop-erties for verification. These requirements modelsmay not have been correctly formulated.

• Tool inaccuracies. It is possible that numericalapproximations affect the verification results. Inaddition, third party tools can contain bugs that areunknown to us.

We could now proceed to perform “Experiments.” Asbefore, we may find a problem with the textual require-ments or the physical system during experimentation. Atthe same time, the assurances from formal verificationand/or simulation-based testing can be compared againstthe experiment results. We may also discover that one of theassurances holds during simulation-based testing or formalverification, but not during the experiments. In this case wemay need to refine any of the other assets, as explainedbefore.

Careful comparisons must be made between the dif-ferent representations in order to discover the cause ofthe assurance conflicts. Such comparisons are indicatedby the bi-directional arrows between “Formal Verification”and “Simulation-based Testing”, “Simulation-based Test-ing” and “Experiments”, and “Formal Verification” and“Experiments”, respectively, in Figure 1.

4 The BERT Handover Task: A Case StudyIn this section, we present a case study to demonstratethe application of assurance-based verification to an HRIscenario considering the following research question: canassurance-based verification provide a higher degree ofconfidence in the resulting assurances than when usingverification techniques in isolation?

BERT 2 is an upper-body humanoid robot designed tofacilitate research into complex human-robot interactions,including verbal and non-verbal communication, such asgaze and physical gestures (Lenz et al. 2010) (see Figure 2).BERT 2’s software architecture was originally developedusing YARP¶. More recently, this system has been wrappedwith a ROS interface.

We verify an object handover to exemplify our approach,in the context of a broader collaborative manufacturescenario where BERT 2 and a person work together toassemble a table (Lenz et al. 2012). In the handover, thefirst step is an activation signal from the human to the

Figure 2. BERT 2 engaged in the handover task.

robot. BERT 2 then picks up a nearby object, and holdsit out to the human. The robot announces that it is readyto handover. The human responds verbally to indicate thatthey are ready to receive. (For practical reasons, human-to-robot verbal signals were relayed to the robot by a humanoperator pressing a key.) Then, the human is expected topull gently on the object while looking at it. BERT 2 thencalculates three binary sensor conditions:

• Gaze: The human’s head position and orientationrelative to the object are tracked using the Vicon R�

motion-tracking system for an approximate measureof whether he/she is looking at the object.

• Pressure: Changes in the robot’s finger positionsare sensed to detect whether the human is applyingpressure to take the weight of the object.

• Location: The Vicon R� motion-tracking system isused to determine whether the human’s hand islocated on the object.

The sensor conditions must be calculated within a timethreshold for BERT 2 to determine if the human “is ready”.The robot should release its grip on the object if allthree conditions are satisfied. Otherwise, the robot shouldterminate the handover and not release the object. Thehuman may disengage and the robot can timeout, whichwould cancel the remainder of the handover task. Thesensors are not completely accurate and may sometimesgive incorrect readings.

A safety requirement ensures that “nothing badhappens”, whereas a liveness requirement ensures that“something good happens eventually” or inside a threshold

¶http://www.yarp.it

Prepared using sagej.cls

11

RobotController_3S

Modelling was carried out using Probabilistic Timed Automataand verification via the PRISM probabilistic model checker.

Clare Dixon Verification and Validation Robotic Assistants 20 / 23

Introduction Techniques and Approach The Domestic Scenario The Manufacturing Scenario Conclusions

Manufacturing Scenario: Simulation and Experiments

Simulation based testing and real robot experiments (BRL)12 Journal Title XX(X)

code to be used in simulation and in the actual robot,providing consistency between simulations, experiments,and deployed use. A screenshot of the ROS/Gazebosimulation can be seen in Figure 5.

For the simulator, additional ROS nodes were con-structed in Python, to simulate BERT 2’s sensor sys-tems and embedded actuation controllers. The pre-existingURDF file describing BERT 2 was extended as describedpreviously for use in Gazebo. The simulated humanbehaviour was controlled by a ROS node written in Python,driving a simplified physical model of the head and hand.

Figure 5. Screenshot of the simulated handover task. Thehuman head and hand are represented in orange. The objectto be handed over is shown in blue.

A testbench was incorporated into the simulator. Thetestbench comprised a test generator, a driver, a checkerand a coverage collector. Achieving the exploration ofmeaningful and interesting sequences of behaviours fromthe robot and its environment in an HRI task is achallenging task. For this reason, we stimulate the robot’scode in the simulation indirectly through stimulating itsenvironment (e.g., the person’s behaviour) instead, and weuse a combination of model-based and pseudorandom testgeneration. Also, to alleviate the complexity of generatingand timing different types of system inputs, the testgenerator is based on a two-tiered approach (Araiza-Illanet al. 2016) where an abstract test is generated first andthen concretized by instantiating low-level parameters. Thehigh-level actions of the human in the simulator includesending signals to the robot, or setting abstract parametersfor gaze, location and pressure. Low-level parametersinclude the robot’s initial pose and the poses and forcevectors applied by the human during the interaction. Forexample, we computed an abstract test of high-level actionsfor the human, by exploring the model in UPPAAL⇤⇤, so

that the robot was activated (sending a signal to activate therobot and wait for the robot to present the object), the gaze,pressure and location sensor readings were correct (setgaze, pressure and location to mean “ready”), and the robotreleased the object. This allowed testing the requirement “ifthe human is ready, BERT 2 should hand over the object”.

The driver distributed the test components in thesimulator. A self-checker — i.e., automated assertionmonitors — was added according to the requirements,described in more detail in the following subsection.Finally, a coverage collector gathered statistics on thetriggering of the assertion monitors.

The simulator code is available online††.

5.2.1 Assertion Monitors For requirements checking,assertion monitors were implemented as state machinesin Python, allowing sequences of events to be captured.If the precondition of an assertion is satisfied, themachine transitions to check the relevant postconditions,to determine if the assertion holds, or not. Otherwise, thepostconditions are never checked.

For example, requirements Reqs. 1a-b and Req. 3 wereboth initially monitored as the following sequence:if (sensors_OK)

wait_for(robot_decision)assert(robot_released_object)

Note that, as with the logical properties, there maybe different ways to implement an assertion for thesame textual requirement, and there is scope formisinterpretation.

The results of the assertion checks, if triggered, arecollected and a conclusion about the satisfaction of theverified requirements can be drawn at the end of simulation.The number of times each assertion monitor has beentriggered in a set of tests can be used as a measure of thecoverage achieved by that test set.

5.3 Experiments5.3.1 Experiment Design BERT 2 can be verified exper-imentally with respect to the textual requirements using acustom facility at the Bristol Robotics Laboratory, as shownin Figure 2. When seeking to verify probabilistic propertiesof a system, the experiments should ideally provide anunbiased sampling, representative of the system’s deployedenvironment. However, some phenomena may be difficultto reproduce naturally in experiments, due to their rarity,safety considerations, or other practical limitations. Conse-quently, experiment-based estimates of their likelihood maybe inaccurate, as may estimates of dependent propertiessuch as the overall success rate of the task.

⇤⇤http://www.uppaal.org††https://github.com/robosafe/testbench_ABV

Prepared using sagej.cls

We carried out a small user valida-tion study with 10 participants eachcarrying out 10 handover tasks.

Clare Dixon Verification and Validation Robotic Assistants 21 / 23

Introduction Techniques and Approach The Domestic Scenario The Manufacturing Scenario Conclusions

The Manufacturing Scenario: Discussion

A number of properties checked inspired by the ISOrequirements, e.g. “At least 95% (60%) of handoverattempts should be completed successfully”.Disagreement between outcomes from some of thetechniques meant further investigation and refinement ofthe models was needed:

simulation based testing revealed that the robot sometimesdropped the table leg accidentally (gripper failure) whichwas not modelled in the formal verification;real experiments revealed false negatives for the pressuresensor and location sensor (they were wrongly reported astoo low/incorrect hand position) not represented elsewhere.

Some of the techniques were not suitable for verifyingsome of the requirements, for example for aspects such asspeed or closeness formal verification may not be the besttechnique to use.

Clare Dixon Verification and Validation Robotic Assistants 22 / 23

Introduction Techniques and Approach The Domestic Scenario The Manufacturing Scenario Conclusions

Concluding Remarks

We gave an overview to the research carried out on the projectTrustworthy Robotic Assistants and discussed approaches totrust and safety for robotic assistants.

We advocate the use of a suite of verification and validationtechniques at different levels of abstraction and coverability tohelp gain assurance of the robot’s safety, reliability andfunctional correctness.

We considered the combination of formal verification (modelchecking), simulation-based testing, and user validation inexperiments with real robots in a domestic and collaborativemanufacturing scenario.

Papers available at www.robosafe.org

Clare Dixon Verification and Validation Robotic Assistants 23 / 23