110
Risk-bounded Programming on Continuous State Prof. Brian Williams March 30 th , 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy MIT News 1

Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

  • Upload
    others

  • View
    6

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Risk-bounded Programming on Continuous State

Prof. Brian Williams

March 30th, 2016

Cognitive Robotics (16.412J / 6.834J)

photo courtesy MIT News

1

Page 2: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Today: Risk-bounded Motion Planning

• M. Ono and B. C. Williams, "Iterative Risk Allocation: A New Approach to Robust Model Predictive Control with a Joint Chance Constraint,” IEEE Conference on Decision and Control, Cancun, Mexico, December 2008.

• M. Ono, B. Williams and L. Blackmore, “Probabilistic Planning for Continuous Dynamic Systems under Bounded Risk,” Journal of Artificial Intelligence Research, v. 46, 2013.

After Advanced Lectures: Risk-bounded Scheduling

• C. Fang, P. Yu, and B. C. Williams, “Chance-constrained Probabilistic Simple Temporal Problems,” AAAI, Montreal, CN, 2014.

• A. Wang and B. C. Williams, “Chance-constrained Scheduling via Conflict-directed Risk Allocation,“ AAAI, Austin, TX, January, 2015.

After Advanced Lectures: Risk-bounded Probabilistic Activity Planning

• Santana, P., Thiébaux, S., Williams, B.C., "RAO*: an Algorithm for Chance-Constrained POMDP's,” AAAI, Phoenix, AZ, February 2016.

Homework:

• Changing Pset order; different from syllabus.

• IRA pset soon (next 1-2 weeks)

• Let Steve know what times work for advanced lecture dry runs

Assignments

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 2

Page 3: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

• Maximizing utility under bounded risk makes sense.

• Risk allocation can help us solve.

Key takeaways

October 29th, 2015

Risk Bounded Goal-directed Motion Planning 3

Page 4: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

• Review

• Risk-aware Trajectory Planning

• Iterative Risk Allocation (IRA)

• Generalizing to Risk-aware Systems

• Convex Risk Allocation (CRA)

Outline

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 4

Page 5: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Depth Navigation for Bathymetric Mapping – Jan. 23rd, 2008

© MBARI. All rights reserved. This content is excluded fromour Creative Commons license. For more information,see https://ocw.mit.edu/help/faq-fair-use/.

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 5

Page 6: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Dynamic Execution of State Plans

Command script

00:00 Go to x1,y100:20 Go to x2,y200:40 Go to x3,y3…

04:10 Go to xn,yn

Plant

Commands

Leaute & Williams, AAAI 05© MBARI. All rights reserved. This content is excluded fromour Creative Commons license. For more information,see https://ocw.mit.edu/help/faq-fair-use/.

16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 63/30/16

Page 7: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

of State Plans

Sulu

Model-based Executive

Observations Commands

“Explore mapping region for at least 100s, then explore bloom region for at least 50s, then return to pickup region.

Avoid obstacles at all times”State Plan

Plant

Leaute & Williams, AAAI 05

Optimal

Robust

Dynamics

+Constraints

16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 73/30/16

© MBARI. All rights reserved. This content is excluded fromour Creative Commons license. For more information,see https://ocw.mit.edu/help/faq-fair-use/.

Execution Dynamic

Page 8: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Sulu: Dynamic Execution of State Plans

Remain in [safe region]

Remain in

[bloom region]

e1e5

Remain in

[mapping region]e2 e3 e4End in

[pickup region]

[50,70] [40,50]

[0,300]

Obstacle 1

Obstacle 2

Mapping

Region

Bloom

RegionPickup

Region

“Explore bloom region for between 50 and 70seconds. Afterwards, explore mapping regionfor between 40s and 50s. End in the pickup region. Avoid obstacles at all times. Complete the mission within 300s”

Issue: Activities couple through time and state constraints.

Approach: Frame as Model-Predictive Controlusing Mixed Logic or Integer / Linear Programming.

[Leaute & Williams, AAAI 05]

A state plan is a model-based program that is unconditional, timed, and hybrid and provides flexibility in state and time.

16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 83/30/16

Page 9: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Frame Planning as a Mathematical Program

)1,1,0(

),1,0( )1,1,0(

..

),(min

maxmax

goal

start0

1

, ::

Nk

NkNk

ts

J

k

N

kkk

NNNN

uuuxxxx

gHxBuAxx

uuxx

k

11ux 11Cost function

Dynamics

Spatial constraints

Initial position and velocity

Goal position and velocity

Thrust limits

TkykxkT

kkkkk FFyxyx ,, , ux

)( Nf x

Cost-to-go function

16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 93/30/16

Page 10: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Encode “Remain In” Constraints, . . .

Remain in [ R ] eEeS

ReTteT kEkS

Nk

k

x)()(0

•Thomas Léauté, "Coordinating Agile Systems through the Model-based Execution of Temporal Plans, " S. M. Thesis,

Massachusetts Institute of Technology, August 2005.

•Thomas Léauté, Brian Williams, “Coordinating Agile Systems Through the Model-based Execution of Temporal Plans,"

Proceedings of the Twentieth National Conference on Artificial Intelligence (AAAI-05), Pittsburgh, PA, July 2005, pp. 114-120.

16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 103/30/16

Page 11: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

• Review

• Risk-aware Trajectory Planning

• Iterative Risk Allocation (IRA)

• Generalizing to Risk-aware Systems

• Convex Risk Allocation (CRA)

Outline

16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 113/30/16

Page 12: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 123/30/16

Page 13: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Issue: Frequent Mission Aborts

Minimum Altitude

Planned trajectory

Actual trajectory

Attitude is less than the minimum altitude

Mission abort

16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 133/30/16

Page 14: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Chanced Constrained, Robust Path Planning

– “Plan optimal path to goal such that p(failure) ≤ Δ.”

p(failure) ≤ Δ

Expected path

October 29th, 2015

Risk Bounded Goal-directed Motion Planning 14

Page 15: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Risk – Performance Tradeoff• Maximum probability of failure is used to

trade performance against risk-aversion.

October 29th, 2015

Risk Bounded Goal-directed Motion Planning 15

2 4 6 8 10 12 14 160

2

4

6

8

10

12

14

x(meters)

y(m

eter

s)

No uncertaintyΔ = 0.1Δ = 0.001Δ = 0.0001

10−5

10−4

10−3

10−2

10−1

60

65

70

75

80

85

90

95

100

105

110

Fue

l Use

Maximum Probability of Collision (Δ)

Method: Uniform Risk Allocation[Blackmore, PhD]

Page 16: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Goal-directed, Risk-bounded Planning

Remain in [safe region]

Explore

[bloom region]

e1e5

Explore

[mapping region]e2 e3 e4End in

[pickup region]

[50,70] [40,50]

[0,300]

2. p( End in [pickup region] fails OR Remain in [safe region] fails ) < .1%.

1. p( Remain in [bloom region] fails OR Remain in [mapping region] fails ) < 5%.

Constraints on risk of failure (Chance Constraints):

1. Science Activities

2. Safety Activities

Instance of Chance-constrained Programming.

“Stay over science region with 95% success, avoid collision and achieve pickup with 99.9% success.”

Operator: Specifies acceptable risk.Executive: Decides how to use risk effectively.

16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 163/30/16

Page 17: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

• p-Sulu: Probabilistic Sulu (plan executive)

Control Sequence

Output

Input and Output

p

p-Sulu

Input

Goal

(cc-QSP)

© source unknown. All rights reserved. This content isexcluded from our Creative Commons license. For moreinformation, see https://ocw.mit.edu/help/faq-fair-use/.

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 17

Page 18: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Example Execution

Remain in [safe region]

Remain in

[bloom region]

e1e5

Remain in

[mapping region]e2 e3 e4End in

[pickup region]

[50,70] [40,50]

[0,300]

T(e1)=0

T(e2)=70

T(e3)=110 T(e4)=150 T(e5)=230

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 18

Page 19: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Problems

Waypoint

Goal

Start

t = 1

t = 5

Convex chance-constrained traj opt

Fixed schedule

Non-convex, chance-constrained traj opt

C

Waypoint

Goal

Start

Obstacle

t = 1

t = 5

Fixed schedule

Flexible schedule (QSP)

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 19

Page 20: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Deterministic Finite-Horizon Optimal Control

ttt

T

tBuAxx

1

1

0

s.t.

itt

iTt

N

i

T

tgxh

11State constraints(Convex)

Discrete-time linear dynamics

Cost function (e.g. fuel consumption)

)(min :1:1

Tu

uJT

T U

Ni

N

i

Ni

N

i

CCCC

CCCC

211

211Notations:

Logicalconjunctions

Logicaldisjunctions

Convex function

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 20

Page 21: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

),0(~ tt Nw

Chance-Constrained FH Optimal Control

),(~ 0,00 xxNx

itt

iTt

N

i

T

tgxh

11State constraints

Stochastic dynamics

Gaussian distribution

s.t.

tttt

T

twBuAxx

1

1

0

State estimation error

Exogenous disturbance

)(min :1:1

Tu

uJT

T U

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 21

Page 22: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

),0(~ tt Nw

itt

iTt

N

i

T

tgxh

11

Chance-Constrained FH Optimal Control

),(~ 0,00 xxNx

Stochastic dynamics

State constraints

s.t.

tttt

T

twBuAxx

1

1

0

)(min :1:1

Tu

uJT

T U

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 22

Page 23: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

),0(~ tt Nw

1Pr

11

itt

iTt

N

i

T

tgxh

Chance-Constrained FH Optimal Control

),(~ 0,00 xxNx

Chance constraint

Stochastic dynamics

Risk bound(Upper bound of the probability of failure)

Assumption: Δ < 0.5

s.t.

tttt

T

twBuAxx

1

1

0

)(min :1:1

Tu

uJT

T U

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 23

Page 24: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Example: Connected Sustainable Homejoint w F. Casalegno & B. Mitchell, MIT Mobile Experience Lab

• Goal: Optimally control HVAC, window opacity, washer and dryer, e-car.

• Objective: Minimize energy cost.

© MIT Mobile Experience Lab. All rights reserved. This content is excluded from our CreativeCommons license. For more information, see https://ocw.mit.edu/help/faq-fair-use/.

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 24

Page 25: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Qualitative State Plan (QSP)

Maintain room

temperature

Wake

up

Wake

up

Maintain room

temperature

Go to

work

Home

from

work

Go to

sleep

Maintain comfortable

sleeping temperature

[24 hours]

“Maintain room temperature after waking up until I go to work. No temperature constraints while I’m at work, but when I get home, maintain room temperatureuntil I go to sleep. Maintain a comfortable sleeping temperature while I sleep.”

……

[1-3 hour] [6-8 hours] [7-8 hours][7-9 hours]

Sulu [Leaute & Williams, AAAI05]

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 25

Page 26: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Maintain room

temperature

Wake

up

Wake

up

Maintain room

temperature

Go to

work

Home

from

work

Go to

sleep

Maintain comfortable

sleeping temperature

[24 hours]

(p)Sulu Results

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 26

Page 27: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

• Review

• Risk-aware Trajectory Planning

• Iterative Risk Allocation (IRA)

• Generalizing to Risk-aware Systems

• Convex Risk Allocation (CRA)

Outline

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 27

Page 28: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Example: Race Car Path Planning

Start

Goal

Walls

Planned Path

Actual PathSafety Margin

Safety Margin

Idea:

• Create safety margin that satisfies the risk bound from start to the goal.

• Reduce to simpler, deterministic optimization problem.

Problem

Find the fastest path to the goal, while limiting the probability of crashthroughout the race to 0.1%

Risk bound

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 28

Page 29: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Executive creates safety margins that satisfy risk bounds while maximizing expected utility

Start

Goal

Safety margin

Walls

(a) Uniform width safety margin

(b) results in better path → takes risk when most beneficial

Start

Goal

Walls

Safety margin

(b) Uneven width safety margin

[Ono & Williams, AAAI 08]Approach: Algorithmic Risk Allocation

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 29

Page 30: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

q

Key Idea - Risk Allocation

CornerNarrow safety margin= higher risk

StraightawayWide safety margin= lower risk

• Taking risk at the cornerresults in a shorter path, than taking the same risk at the straightaway.

• Sensitivity of path length to risk is higher at the corner.

• Risk Allocation

– Optimize allocation of risk to time steps and constraints.

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 30

Page 31: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Iterative Risk Allocation (IRA) Algorithm

Iteration

•Descent algorithm

•Starts from a suboptimal risk allocation•Improves it by iteration

)()()( 2*

1*

0* δδδ JJJ

(Refer to paper for proof)

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 31

Page 32: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Iterative Risk Allocation AlgorithmAlgorithm IRA

1 Initialize with arbitrary risk allocation

2 Loop3 Compute the best

available path given the current risk allocation

4 Decrease the risk where the constraint is inactive

5 Increase the risk where the constraint is active

6 End loopStart

Goal

Safety margin

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 32

Page 33: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Iterative Risk Allocation AlgorithmAlgorithm IRA

1 Initialize with arbitrary risk allocation

2 Loop3 Compute the best

available path given the current risk allocation

4 Decrease the risk where the constraint is inactive

5 Increase the risk where the constraint is active

6 End loopStart

Goal

Safety margin

No gap = Constraint is active

Gap = constraint is inactive

Best available path given the safety margin

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 33

Page 34: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Iterative Risk Allocation AlgorithmAlgorithm IRA

1 Initialize with arbitrary risk allocation

2 Loop3 Compute the best

available path given the current risk allocation

4 Decrease the risk where the constraint is inactive

5 Increase the risk where the constraint is active

6 End loopStart

Goal

Safety margin

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 34

Page 35: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Iterative Risk Allocation AlgorithmAlgorithm IRA

1 Initialize with arbitrary risk allocation

2 Loop3 Compute the best

available path given the current risk allocation

4 Decrease the risk where the constraint is inactive

5 Increase the risk where the constraint is active

6 End loopStart

Goal

Safety margin

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 35

Page 36: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Iterative Risk Allocation AlgorithmAlgorithm IRA

1 Initialize with arbitrary risk allocation

2 Loop3 Compute the best

available path given the current risk allocation

4 Decrease the risk where the constraint is inactive

5 Increase the risk where the constraint is active

6 End loopStart

Goal

Safety margin

ActiveInactive

Inactive

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 36

Page 37: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Iterative Risk Allocation AlgorithmAlgorithm IRA

1 Initialize with arbitrary risk allocation

2 Loop3 Compute the best

available path given the current risk allocation

4 Decrease the risk where the constraint is inactive

5 Increase the risk where the constraint is active

6 End loopStart

Goal

Safety margin

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 37

Page 38: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Iterative Risk Allocation AlgorithmAlgorithm IRA

1 Initialize with arbitrary risk allocation

2 Loop3 Compute the best

available path given the current risk allocation

4 Decrease the risk where the constraint is inactive

5 Increase the risk where the constraint is active

6 End loopStart

Goal

Safety margin

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 38

Page 39: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Iterative Risk Allocation AlgorithmAlgorithm IRA

1 Initialize with arbitrary risk allocation

2 Loop3 Compute the best

available path given the current risk allocation

4 Decrease the risk where the constraint is inactive

5 Increase the risk where the constraint is active

6 End loopStart

Goal

Safety margin

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 39

Page 40: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Monterey Bay Mapping Example

Sea floor level

Risk allocation ( )%5Ono & Williams, AAAI 08

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 40

Page 41: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

• Review

• Risk-aware Trajectory Planning (IRA)

• Iterative Risk Allocation

• Generalizing to Risk-aware Systems

• Convex Risk Allocation (CRA)

Outline

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 41

Page 42: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Risk-Sensitive Architectures for Exploration• In collaboration with JPL, WHOI and Caltech.

• Initial year study, funded by Keck Institute for Spaces Sciences.

• 2 year follow on for demonstration.

NereId Under IceVenus Sage © Woods Hole Oceanographic Institution. All rights reserved. Thiscontent is excluded from our Creative Commons license. For moreinformation, see https://ocw.mit.edu/help/faq-fair-use/.

This image is in the public domain.

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 42

Page 43: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Risk-aware Planning & Execution

UserKirk

Pike

Sulu

Goals &

models

in RMPL

Control

Commands

Enterprise

Coordinates and monitors task risk

Plans paths within risk bounds

Sketches mission and assigns tasks within risk-bounds

Burton

Plans actions within risk bounds.

Bones

Diagnoses incipient failures

Uhura

Collaboratively adapts goal to reduce risk

Start

GoalWalls

Safety margin

11:00 a.m. 4:00 p.m. 10:00 p.m.

© source unknown. All rights reserved. This content isexcluded from our Creative Commons license. For moreinformation, see https://ocw.mit.edu/help/faq-fair-use/.

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 43

Page 44: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Falkor Cruise – March-April, 2015

Falkor - Schmidt Ocean Institute

Slocum Glider

© Teledyne Webb Research. All rights reserved. This contentis excluded from our Creative Commons license. For moreinformation, see https://ocw.mit.edu/help/faq-fair-use/.

© Schmidt Ocean Institute. All rights reserved. This contentis excluded from our Creative Commons license. For moreinformation, see https://ocw.mit.edu/help/faq-fair-use/.

© Schmidt Ocean Institute. All rights reserved. This contentis excluded from our Creative Commons license. For moreinformation, see https://ocw.mit.edu/help/faq-fair-use/.

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 44

Page 45: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 45

Page 46: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Glider Primer

© Woods Hole Oceanographic Institution. All rights reserved. This content is excluded from ourCreative Commons license. For more information, see https://ocw.mit.edu/help/faq-fair-use/.

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 46

Page 47: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

User Kirk

Pike

Sulu-lite

Goals &

models

in RMPL

Control

Commands

Enterprise

Coordinates and monitors tasks

Plans paths

Sketches mission and assigns tasks

© source unknown. All rights reserved. This content isexcluded from our Creative Commons license. For moreinformation, see https://ocw.mit.edu/help/faq-fair-use/.

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 47

Page 48: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Activity Planning in Scott Reef

region1

no-fly zone

region2region3

region4

no-fly zone

ship1

Iver1

glider

Iver2

10:00-11:30am

10:30-13:00

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 48

Page 49: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

START POINT

END GOAL

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 49

Page 50: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

April/May 2016 Deployment:Slocum Glider at Santa Barbara• Mission goal:

– Use miniaturized mass spectrometer to find and characterize oil seeps off the coast of Santa Barbara.

• Primary research goal:– Adaptive science using planning and rule-based algorithms.

• Secondary research goal:– Risk-aware path planning.

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 50

Page 51: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Advisory System for WHOI Cruise AT26-06: San Francisco, California to Los Angeles

• Find and sample methane seeps near the coast.

• Locate and sample a 60 year-old DDT dumping site.

• Recover and replace incubators on the seafloor.

Courtesy WHOI

Yu, Fang, Williams, Camilli, Fall 13

© Woods Hole Oceanographic Institution. All rights reserved. This content is excluded from ourCreative Commons license. For more information, see https://ocw.mit.edu/help/faq-fair-use/.

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 51

Page 52: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

• [Day 1] Jason failed after 30 min into its first dive, entered an uncontrollable spin and broke its optic fiber tether.

• [Day 1] The new camera installed on Sentry did not work well in low light situations. It had been replaced during its second dive.

Everything Can Go Wrong

© Woods Hole Oceanographic Institution. All rights reserved. Thiscontent is excluded from our Creative Commons license.For moreinformation, see https://ocw.mit.edu/help/faq-fair-use/.

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 52

Page 53: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

User

PTS

pUhura

pKirk

pSulu

Goals in

Natural

Language

Control

Commands

pEnterprise

Collaboratively diagnoses and resolves goal failures

Plans mission and contingencies.

Navigates within risk bounds

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 53

Page 54: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

• Review

• Risk-aware Trajectory Planning

• Iterative Risk Allocation (IRA)

• Generalizing to Risk-aware Systems

• Convex Risk Allocation (CRA)– Intuitions

– Math (optional)

Outline

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 54

Page 55: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Problems

Waypoint

Goal

Start

t = 1

t = 5

Convex chance-constrained traj opt

Fixed schedule

Non-convex, chance-constrained traj opt

C

Waypoint

Goal

Start

Obstacle

t = 1

t = 5

Fixed schedule

Flexible schedule (QSP)

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 55

Page 56: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Risk-Allocation Overview: One Variable

Idea 1: We easily solve a chance constrained problem with one linear constraint C and one normally distributed random variable x, by reformulating C to a deterministic constraint C’ on .

< 1%

C(x)Chance constraint:

Risk < 1%x m

x

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 58

C '(x) C(x)m

Page 57: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Risk-Allocation Overview: Many Variables

Idea 2: Generalize to a single constraint over an

N-dimensional random variable,

by projecting its distribution onto the axis perpendicular to the constraint boundary.

99.9%

99%

90%

80%

Chance constraint:

Risk < 1%

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 59

Page 58: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Risk-Allocation Overview: Many Constraints

99.9%

99%

90%

80%

Chance constraint:

Risk < 1%

Idea 3: Generalize to a joint chance-constraint over multiple constraints C1, C2,

by distributing risk.

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 58

Page 59: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Risk-Allocation Overview: Many Constraints

99.9%

99%

90%

80%

Chance constraint:

Risk < 1%

Find a solution such that:

1. Each constraint Ci takes less than δi risk, and

2. Σi δi ≤ 1%

Note: this bound is derived from Boole’s inequality.3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 59

Page 60: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Risk-Allocation Overview: Conservatism

99.9%

99%

90%

80%

Chance constraint:

Risk < 1%

Conservatism

Significantly less conservative than the elliptic approximation, especially in a high-dimensional problem.

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 60

Page 61: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

• Review

• Risk-aware Trajectory Planning

• Iterative Risk Allocation (IRA)

• Generalizing to Risk-aware Systems

• Convex Risk Allocation (CRA)– Intuitions

– Math (optional)

Outline

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 61

Page 62: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

),0(~ tt Nw

1Pr

11

itt

iTt

N

i

T

tgxh

Reformulation: Now Lets Do the Math!

),(~ 0,00 xxNx

Chance constraint

Stochastic dynamics

Risk bound(Upper bound of the probability of failure)

Assumption: Δ < 0.5

s.t.

tttt

T

twBuAxx

1

1

0

)(min :1:1

Tu

uJT

T U

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 62

Page 63: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Conversion of Joint Chance Constraint

1Pr

00

itt

iTt

N

i

T

tgxhJoint chance

constraint

Intractable- Requires computation of an integral over a multivariate Gaussian.

1

A set of individual chance constraints.- involves univariate Gaussian distribution.

2

A set of deterministic state constraint.

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 63

Page 64: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Decomposition of Joint Chance Constraint

1Pr

00

itt

iTt

N

i

T

tgxhJoint chance

constraint

Using Boole’s inequality (union bound)

BABA PrPrPr

1

Where A and B denote constraint failures

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 64

Page 65: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

1Pr

00

itt

iTt

N

i

T

tgxhJoint chance

constraint

is implied by:

0

1Pr

it11

,

11

N

i

T

t

it

it

it

itt

iTt

N

i

T

tgxh

Individual chanceconstraints

Upper bound of the probability of violating ithconstraint at time t

Upper bound of the probability of violating anyconstraints over the planning horizon

Risk allocation:

NT 2

111 ,δ

1Decomposition of Joint Chance Constraint

Variable

Constant

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 65

Page 66: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

)(min:1

UJT

Tu U

),0(~ tt Nw

),(~ 0,00 xxNx

it

it

it

itt

iTtit

gxh

,

1Pr

Individual chanceconstraints

1Pr

00

itt

iTt

N

i

T

tgxhJoint chance

constraint

δmin

Risk allocationoptimization

1Decomposition of Joint Chance Constraint

s.t.

tttt

T

twBuAxx

1

1

0

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 66

Page 67: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

itit

itt

iTt mgxh Deterministic constraint

it

itt

iTt gxh 1PrChance constraint

Conversion to Deterministic Constraint2

tiTt xh

itg

Mean state (deterministic variable)

Prob. Distribution of t

iTt xh

it

it

itm

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 67

Page 68: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

itit

itt

iTt mgxh Deterministic constraint

Safety Margin

it

itt

iTt gxh 1PrChance constraint

1

00,, )(

t

nx

Tnw

ntx AA

),(~ ,txtt xNx

12erf2 1, i

tittx

iTt

it

it hhm

where

(Inverse of cdf of Gaussian)

[Charnes et. al. 1959]

t = 1

t = 2

t = 3t = 4t = 5

Safety Margin

Mean stateNominal path

11 m

22 m

Conversion to Deterministic Constraint2

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 68

Page 69: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

)(min :1:1

Tu

uJT

T U

),0(~ tt Nw

),(~ 0,00 xxNx

it

it

it

itt

iTt

I

i

T

tgxh

,

111Pr

Individual chanceconstraints

δmin

s.t.

tttt

T

twBuAxx

1

1

0

Conversion to Deterministic Constraint2

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 69

Page 70: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

)(min :1:1

Tu

uJT

T U

it

it

it

it

itt

iTt

I

i

T

tmgxh

,

11

δmin

s.t.

ttt

T

tBuxAx

1

1

0

Conversion to Deterministic Constraint2

Convex if δ < 0.5

1.00.50

it

m ConvexConvex optimization

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 70

Page 71: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

• Maximizing utility under bounded risk makes sense.

• Risk allocation can help us solve.

Key takeaways

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 71

Page 72: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Summary: Risk Allocation

Iteration

1. IRA: reallocates risk manually.2. CRA,NRA: standard solver reallocates risk.

)()()( 2*

1*

0* δδδ JJJ

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 72

Page 73: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Approach: Programming Cognitive Systems

1. An embedded programming language elevated to the goal-level through partial specification and operations on state (RMPL).

2. A language executive that achieves robustness by reasoning over constraint-based models and by bounding risk (Enterprise).

3. Interfaces that support natural human interaction fluidly and at the cognitive level (Uhura).

This image is in the public domain.

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 73

Page 74: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Risk-aware Planning & Execution

User

Kirk

Pike

Sulu

Goals &

models

in RMPL

Control

Commands

Enterprise

Coordinates and monitors task risk

Plans paths within risk bounds

Sketches mission and assigns tasks within risk-bounds

Burton

Plans actions within risk bounds.

Bones

Diagnoses incipient failures

Uhura

Collaboratively adapts goal to reduce risk

Start

Goal

Walls

Safety margin

11:00 a.m. 4:00 p.m. 10:00 p.m.

© source unknown. All rights reserved. This content isexcluded from our Creative Commons license. For moreinformation, see https://ocw.mit.edu/help/faq-fair-use/.

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 74

Page 75: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

APPENDIX: RISK-BOUNDED PLANNING FOR NON-CONVEX PROBLEMS

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 75

Page 76: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Next …

Waypoint

Goal

Start

t = 1

t = 5

Convex chance-constrained trajectory optimization

Fixed schedule

Non-convex, chance-constrained traj opt

C

Waypoint

Goal

Start

Obstacle

t = 1

t = 5

Fixed schedule

October 29th, 2015

76Risk Bounded Goal-directed Motion Planning

Page 77: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Non-Convex Problem Formulation

Non-convex state constraint

C

Waypoint

Goal

Start

Obstacle

t = 1

t = 5

Fixed schedule

October 29th, 2015

77Risk Bounded Goal-directed Motion Planning

Page 78: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Problem Formulation: Non-Convex Chance Constraint

1Pr ..

)(min

11

:1

ntt

nTt

N

n

T

t

ti

u

gxhts

uJT

T U

CC over convex state constraints

1

A set of individualchance constraints

2

A set of deterministicstate constraints

A joint chance constraints

NT

nt

nt

nt

nt

ntt

nTt

N

n

T

t

ti

u

mgxhts

uJT

T

,

1,1

11

..

)(min:1

U

Convex Deterministic Program

Risk allocation

78Risk Bounded Goal-directed Motion Planning

Page 79: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Problem Formulation: Non-Convex Chance Constraint

1Pr ..

)(min

11

:1

ntt

nTt

N

n

T

t

ti

u

gxhts

uJT

T U

CC over convex state constraints

NT

nt

nt

nt

knt

kntt

kTnt

K

k

N

n

T

t

ti

u

mgxhts

uJT

T

,

1,1

,,,

111

..

)(min:1

U

Convex Disjunctive Deterministic Program

NT

nt

nt

nt

nt

ntt

nTt

N

n

T

t

ti

u

mgxhts

uJT

T

,

1,1

11

..

)(min:1

U

Convex Deterministic Program

1Pr ..

)(min

,,

111

:1

kntt

kTnt

K

k

N

n

T

t

ti

u

gxhts

uJT

T U

CC over non-convex state constraints

Risk allocation Risk allocationRisk selection

79

Page 80: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Solving Disjunctive Programusing Branch and Bound

NT

nt

nt

nt

knt

kntt

kTnt

K

k

N

n

T

t

ti

u

mgxhts

uJT

T

,

1,1

,,,

111

..

)(min:1

U

Convex Disjunctive Programming

nt

kt

ktt

kTttk mgxhCKNT ,1,1,1 ;2,1,2

nt

knt

kntt

kTnt

K

k

N

n

T

tmgxh ,,,

111

)()( 22211211 CCCC

Example:

October 29th, 2015 80Risk Bounded Goal-directed Motion Planning

Page 81: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Stochastic DLP Branch and Bound

)()( 22211211 CCCC

)( 222111 CCC )( 222112 CCC

2111 CC 2211 CC 2112 CC 2212 CC

Convex Optimization Problems

Repeat until no clauses left:1. Select clause.2. Split on disjuncts.

October 29th, 2015

81Risk Bounded Goal-directed Motion Planning

Page 82: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

)()( 22211211 CCCC

Stochastic DLP Branch and Bound

)( 222111 CCC )( 222112 CCC

2111 CC 2211 CC 2112 CC 2212 CC

Convex Optimization Problems

Waypoint

Goal

Start

t = 1

t = 5

Convex, chance-constrained

Fixed scheduleRepeat until no clauses left:

1. Select clause.2. Split on disjuncts.

October 29th, 2015

82Risk Bounded Goal-directed Motion Planning

Page 83: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Bound Sub-ProblemsThrough Convex Relaxation

)()( 22211211 CCCC

)( 222111 CCC )( 222112 CCC

2111 CC 2211 CC 2112 CC 2212 CC

• Bound: Remove all disjunctive clauses [Li & Williams 2005].

• Issue: Computing bound is slow!!

• Cause: Sub-problems include non-linear constraints.

October 29th, 2015

83Risk Bounded Goal-directed Motion Planning

Page 84: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

B&B Subproblem (non-linear)

),0(~ tt Nw

s.t. ttttTtwBuAxx

110

),(~ 0,00 xxNx

)(min :1:1

TuuJ

T

0 ,1

)()()(

1

t

T

tt

tti

tti

ttTti

t

T

tmgxh

Nonlinear

1.00.50

it

m

October 29th, 2015

84Risk Bounded Goal-directed Motion Planning

Page 85: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

B&B Subproblem Relaxation (Linear)

),0(~ tt Nw

s.t. ttttTtwBuAxx

110

),(~ 0,00 xxNx

)(min :1:1

TuuJ

T

0 ,1

)()()(

1

t

T

tt

tti

tti

ttTti

t

T

tmgxh

Nonlinear

t

Fixed RiskRelaxation

October 29th, 2015

85Risk Bounded Goal-directed Motion Planning

Page 86: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

B&B Subproblem Relaxation (Linear)

),0(~ tt Nw

s.t. ttttTtwBuAxx

110

),(~ 0,00 xxNx

)(min :1:1

TuuJ

T

)()()(

1

tit

titt

Ttit

T

tmgxh

Constant

t

Fixed RiskRelaxation

•All constraints are linear (FRR is typically LP or QP).

October 29th, 2015

86Risk Bounded Goal-directed Motion Planning

Page 87: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

FRR Intuition

• Results in an infeasible solution to the original problem.

• Gives lower bound on the cost of the original problem.

Start

Goal

Safety margin

Start

Goal

FRR Safety margin

FRROriginal problemSets safety margin for all constraints to max risk Δ.

October 29th, 2015

87Risk Bounded Goal-directed Motion Planning

Page 88: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Problems

Waypoint

Goal

Start

t = 1

t = 5

Convex traj opt

Fixed schedule

Non-convex, traj opt

C

Waypoint

Goal

Start

Obstacle

t = 1

t = 5

Fixed schedule

C

Waypoint

Goal

Start

Obstacle

[1 3]

[2 4]

[0 5]

Simple temporalconstraints

Goal-directed qualitative state plan traj opt 88

Page 89: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

APPENDIX: OVERVIEW AND ALTERNATIVE APPROACHES

3/30/16 16.412J / 6.834J – L15 Risk-bounded Programs on Continuous States 89

Page 90: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

What are the uncertainties and risks?

Airport

UAV

NFZ

Scenic region

October 29th, 2015

Risk Bounded Goal-directed Motion Planning 90

Page 91: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Robust Model Predictive Control

• Receding horizon (MPC) planners react to uncertainty after something goes wrong.

• Can we take precautionary actions before something goes wrong?

•Ali A. Jalali and Vahid Nadimi, “A Survey on Robust Model Predictive Control from 1999-2006.”

October 29th, 2015

Risk Bounded Goal-directed Motion Planning 91

Page 92: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Robust versus Chance ConstrainedR

obust

Pre

dic

tive C

ontr

ol

x

y

Assume bounds on

uncertainty.

Ch

an

ce-c

onstr

ain

ed

Pre

dic

tive

Co

ntr

ol Wind speed

Wind speed

99%

90%80% x

y

Assume probability

distribution that

characterizes uncertainty.

No Fly Zone

No Fly Zone

• Predicted position has bounded uncertainty.

• Problem: Find control sequence that satisfies

constraints for all realizations of uncertainty.

t =1 t =2

99.9%99%

90%80%

99.9%

99%

90%

80%

t =1 t =2

• Predicted position has probabilistic uncertainty.

• Problem: Find control sequence that satisfies

constraints within a probability bound

(Chance Constraint). 92October 29th, 2015

- 99

Page 93: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Incorporating Uncertainty

• Deterministic discrete-time LTI model.

• Additive uncertainty

• Multiplicative uncertainty

ttt BuAxx 1

ttt BuxAAx )(1

tttt wBuAxx 1

x

y

W

wt W

xy

p(x,y)

p(wt )= N(wt,P0 )

October 29th, 2015

Risk Bounded Goal-directed Motion Planning 93

Page 94: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

What to Minimize? (Bounded Uncertainty)

• Minimize the worst case cost

• Minimize nominal cost

yuncertaint Bounded :

..

Cost when :)(min

Ww

gxhts

wJitt

iTtWw

0U,XU

yuncertaint Bounded :

..

)(maxmin

Ww

gxhts

Jitt

iTtWw

Ww

UX,

U

October 29th, 2015

Risk Bounded Goal-directed Motion Planning 94

Page 95: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

What to Minimize? (Stochastic Uncertainty)

• Utilitarian approach

• Chance constrained optimization

)()(min UU,XU

pfJ

Probability of failurePenalty (constant)

)( ..

)(min

U

U,XU

fts

J

Probability of failure Risk boundOctober 29th, 2015

Risk Bounded Goal-directed Motion Planning 95

Page 96: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Solution Methods forChance-Constrained Problems• Sampling based methods

– Scenario-based• Bernardini and Bemporad, 2009

– Particle control• Blackmore et al., 2010

• Non-sampling-based methods

– Elliptic approximation (direct extension of robust predictive control)

• van Hessem, 2004

– Risk allocation• Ono and Williams, 2008

October 29th, 2015

Risk Bounded Goal-directed Motion Planning 96

Page 97: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Particle Control1. Use particles to sample random variables.

Obstacle 1

Obstacle 2

Goal Region

Initial state distribution.

Particles approximating initial state distribution.

)(~)(t

it p )(~ 0,

)(0, c

ic p xx Ni 1 Ft 0

October 29th, 2015

Risk Bounded Goal-directed Motion Planning 97

Page 98: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Particle Control2. Calculate future state trajectory for each particle, leaving explicit,

dependence on control inputs u0:T-1.

Obstacle 1

Obstacle 2

Goal Region

)(,

)(0,

)(:0,

iFc

ic

iTc

x

xx

Particle 1 for u = uB

Particle 1 for u = uA

t=0

t=1

t=2

t=3

t=4

t=0

t=1

t=2

t=3

t=4

),,( )(1:0

)(0,1:0

)( it

ictt

it f xux

October 29th, 2015

Risk Bounded Goal-directed Motion Planning 98

Page 99: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Particle Control2. Calculate future state trajectory for each particle, leaving explicit,

dependence on control inputs u0:T-1.

Obstacle 1

Obstacle 2

Goal RegionParticles 1…Nfor u = uB

Particles 1…Nfor u = uA

October 29th, 2015

Risk Bounded Goal-directed Motion Planning 99

Page 100: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

October 29th, 2015

Risk Bounded Goal-directed Motion Planning 100

Particle Control3. Express chance-constraints of optimization problem approximately in

terms of particles.

Probability of failureTrue expectation approximated by the approximated byfraction of failing sample mean of costparticles.function:.

Sample meanapproximates state mean.

Obstacle 1

Obstacle 2

Goal Region

t=0t=1

t=2

t=3

t=4

E h(u0:F1,x1:F )

1N

h(u0:F1,x1:F(i) )

i1

N

LB4

Page 101: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Particle Control4. Solve approximate deterministic optimization problem for u0:F-1.

Obstacle 1

Obstacle 2

Goal Region

t=0

t=1

t=2

t=3

t=4

10% of particles fail in optimal solution.

δ = 0.1

October 29th, 2015

Risk Bounded Goal-directed Motion Planning 101

Page 102: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Convergence

– As N∞, approximation becomes exact.

Obstacle 1

Obstacle 2

Goal Region

October 29th, 2015

Risk Bounded Goal-directed Motion Planning 102

Page 103: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Convergence

– As N∞, approximation becomes exact.

Obstacle 1

Obstacle 2

Goal Region

10% probability of failure.

October 29th, 2015

Risk Bounded Goal-directed Motion Planning 103

Page 104: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Solution Methods forChance-Constrained Problems• Sampling based methods

– Scenario-based• Bernardini and Bemporad, 2009

– Particle control• Blackmore et al., 2010

• Non-sampling-based methods

– Elliptic approximation (direct extension of robust predictive control)

• van Hessem, 2004

– Risk allocation• Ono and Williams, 2008

October 29th, 2015

Risk Bounded Goal-directed Motion Planning 104

Page 105: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Elliptic ApproximationNo Fly Zone

1. Derive probability distribution over future states as a function of control inputs.

Chance constraint:

Risk < 1%

Note: When planning in an N-dimensional state space over time steps, a joint distribution over an N-dimensional space must be considered.

99.9%

99%

90%

80%

October 29th, 2015

Risk Bounded Goal-directed Motion Planning 105

Page 106: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Elliptic ApproximationNo Fly Zone

1. Derive probability distribution over future states as a function of control inputs.

2. Find a 99% probability ellipse.

Chance constraint:

Risk < 1%

99.9%

99%

90%

80%

October 29th, 2015

Risk Bounded Goal-directed Motion Planning 106

Page 107: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Elliptic ApproximationNo Fly ZoneChance constraint:

Risk < 1%

1. Derive probability distribution over future states as a function of control inputs.

2. Find a 99% probability ellipse.

3. Find control sequence that makes sure the probability ellipse is within the constraint boundaries.

99.9%

99%

90%

80%

October 29th, 2015

Risk Bounded Goal-directed Motion Planning 107

Page 108: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Conservatism of Elliptic ApproximationNo Fly Zone 99.9%

99%

90%

80%

Issue: often very conservative

Real probability of failure

Probability density function

October 29th, 2015

Risk Bounded Goal-directed Motion Planning 108

Page 109: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

Conservatism of Elliptic ApproximationNo Fly Zone 99.9%

99%

90%

80%

Issue: often very conservative.

Conservatism

October 29th, 2015

Risk Bounded Goal-directed Motion Planning 109

Page 110: Risk-bounded Programming on Continuous State...Risk-bounded Programming on Continuous State Prof. Brian Williams March 30th, 2016 Cognitive Robotics (16.412J / 6.834J) photo courtesy

MIT OpenCourseWarehttps://ocw.mit.edu

16.412J / 6.834J Cognitive RoboticsSpring 2016

For information about citing these materials or our Terms of Use, visit: https://ocw.mit.edu/terms.