41
ICML 2020 On Learning Sets of Symmetric Elements Haggai Maron, Or Litany, Gal Chechik, Ethan Fetaya [1] Nvidia Research [2] Stanford University [3] Bar-Ilan University [1] [2] [1,3] [3]

On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

  • Upload
    others

  • View
    1

  • Download
    0

Embed Size (px)

Citation preview

Page 1: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

ICML 2020

On Learning Sets of Symmetric Elements

Haggai Maron, Or Litany, Gal Chechik, Ethan Fetaya [1] Nvidia Research [2] Stanford University [3] Bar-Ilan University

[1] [2] [1,3] [3]

Page 2: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

Motivation and Overview

Page 3: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

Set Symmetry

{ }Input set

Previous work (DeepSets, PointNet) targeted training a deep network over sets

DeepNet

x1x2⋮xm ,

x1x2⋮xm ,

x1x2⋮xm , …

Page 4: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

Set+Elements symmetry

{ }Input set

Main challenge: What architecture is optimal when elements of the set have their own symmetries?

DeepNet, , …

Both the set and its elements have symmetries.

Page 5: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

Deep Symmetric sets

{ }Input image set Output

Page 6: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

Set symmetry: Order invariance/equivariance

{ }=

Page 7: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

Set symmetry: Order invariance/equivariance

{ }={ }

Page 8: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

Set symmetry: Order invariance/equivariance

{ }={ }

Page 9: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

Element symmetry: Translation invariance/equivariance

{ }

Page 10: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

{ }{ }

Element symmetry: Translation invariance/equivariance

Page 11: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

{ }{ }

Element symmetry: Translation invariance/equivariance

Page 12: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

Applications

Modalities1D signals 2D images 3D pointclouds Graph

Page 13: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

This paperA principled approach for learning sets of complex elements (graphs, point clouds, images)

Characterize maximally expressive linear layers that respect the symmetries (DSS layers)

Prove universality results

Experimentally demonstrate that DSS networks outperform baselines

Page 14: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

Previous work

Page 15: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

Deep sets [Zaheer et al. 2017]

Page 16: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

CNN

CNN

CNN

Siamese

Deep sets [Zaheer et al. 2017]

Page 17: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

CNN

CNN

CNN

FeaturesSiamese

Deep sets [Zaheer et al. 2017]

Page 18: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

CNN

CNN

CNN

DeepSets

FeaturesSiamese Deeps sets block

Deep sets [Zaheer et al.]

Page 19: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

Previous work: information sharing

Aittala and Durand, ECCV 2018

Sridhar et al., NeuriPS 2019

Liu et al., ICCV 2019

CNN

CNN

CNN

CNN

CNN

CNN

Information sharing layer

Page 20: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

Our approach

Page 21: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

InvarianceMany Learning tasks are invariant to natural transformations (symmetries)

More formally. Let be a subgroup:

is invariant if , for all

e.g. image classification

H ≤ Sn

f : ℝn → ℝ f(τ ⋅ x) = f(x) τ ∈ Hτ

f f

“Cat”

Page 22: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

EquivarianceLet be a subgroup:

Equivariant if ,

e.g. edge detection

H ≤ Sn

f(τ ⋅ x) = τ ⋅ f(x)

τ

τf f

Page 23: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

Invariant neural networks

Equivariant FC Invariant

• Invariant by construction

Page 24: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

Deep Symmetric Sets

with symmetry group

Want to be invariant/equivariant to both and the ordering

Formally the symmetry group is

x1, …, xn ∈ ℝd G ≤ Sd

G

H = Sn × G ≤ Snd

G

Page 25: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

Main challenges

• What is the space of linear equivariant layers for specific ?H = SN × G

Page 26: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

• What is the space of linear equivariant layers for a given ?

• Can we compute these operators efficiently?

H = SN × G

Main challenges

Page 27: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

• What is the space of linear equivariant layers for a given ?

• Can we compute these operators efficiently?

• Do we lose expressive power?

H = SN × G

-invariant networksH

-invariant continuous functionsH

Continuous functions

Main challenges

Page 28: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

• What is the space of linear equivariant layers for a given ?

• Can we compute these operators efficiently?

• Do we lose expressive power?

H = SN × G

-invariant networksH

-invariant continuous functionsH

Continuous functions

Gap?

Main challenges

Page 29: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

Theorem: Any linear −equivariant layer is of the form

where are linear -equivariant functions

We call these layers Deep Sets for Symmetric elements layers (DSS)

SN × G L : ℝn×d → ℝn×d

L(X)i = LG1 (xi) + ∑

j≠i

LG2 (xj)

LG1 , LG

2 G

-equivariant layersH

Page 30: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

DSS for images

are images

is the group of circular translations

-equivariant layers are convolutions

x1, …, xn

G 2D

G

Single DSS layer

Page 31: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

DSS for images

are images

is the group of circular translations

-equivariant layers are convolutions

x1, …, xn

G 2D

G

CONV

CONV

CONV

Single DSS layer

Page 32: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

DSS for images

are images

is the group of circular translations

-equivariant layers are convolutions

x1, …, xn

G 2D

G

CONV

CONV

CONV

+

Single DSS layer

Page 33: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

DSS for images

are images

is the group of circular translations

-equivariant layers are convolutions

x1, …, xn

G 2D

G

CONV

CONV

CONV

CONV

+

Single DSS layer

Page 34: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

DSS for images

Siamese part

Information sharing part

CONV

CONV

CONV

CONV

+ +

Single DSS layer

Page 35: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

Expressive powerTheorem

If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS networks for -equivariant functions.

SN × G

Page 36: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

Expressive powerTheorem

If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS networks for -equivariant functions.

• Main tool:

• Noether’s Theorem (Invariant theory)

• For any finite group , the ring of invariant polynomials is finitely generated.

• Generators can be used to create continuous unique encodings for elements in

SN × G

H ℝ[x1, . . . , xn]H

ℝn×d /H

Page 37: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

Results

Page 38: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

Signal classification

Page 39: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

Image selection

Page 40: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

Shape selection

Page 41: On Learning Sets of Symmetric Elements · 2021. 1. 13. · Expressive power Theorem If G-equivariant networks are universal appoximators for G-equivariant functions, then so are DSS

Conclusions

A general framework for learning sets of complex elements

Generalizes many previous works

Expressivity results

Works well in many tasks and data types