1
Non-linear Distribution of Descriptors Descriptors are non- linearly distributed. Observed descriptor pair share common information. Mixture model of Probabilistic CCA Tackle non-linearity with mixture model Encode shared information by specifying latent variable Mathematical Formulation Let v = (x, y), x, y satisfy p(v | k) is Gaussian with Comparison of performance on HMDB51. Comparison of performance on UCF101. Influence of Parameters Number of Components Latent Dimension Multi-View Super Vector for Action Recognition Zhuowei Cai 1 , Limin Wang 1,2 , Xiaojiang Peng 1 , Yu Qiao 1,2 1 Shenzhen Institutes of Advanced Technology 2 The Chinese University of Hong Kong Motivation Multi-View Super Vector Experimental Results Performance of Different Components Challenges in Action Recognition Different feature descriptors capture different aspects of object feature. Current fusion pipelines presume different properties between feature descriptors. Direct concatenation presumes strong correlation between fusion descriptors. Kernel average and score fusion prefer mutual independence between fusion descriptors. Typical fusion pipelines: We utilize the Mixture of Probabilistic Canonical Correlation Analyzers to model a pair of feature descriptors. Our Approach Mixture model of Probabilistic CCA In each local Probabilistic CCA, shared information Z and private information λ jointly generate the observed descriptors Shared information Z is extracted via concatenating latent information z k for each local PCCA Private information g x is extracted by taking the gradients of the expected loglikelihood with respect to λ x (and λ y ) Z, g x and g y form the final MVSV. X Fusion Y X Y Descriptor- level (linear) Kernel- level GMM GMM Fusion GMM SVM Score SVM Score X Y Score- level GMM GMM Fusion Score SVM Score SVM Score X Y MVSV M-PCCA Fusion SVM Score

Non-linear Distribution of Descriptors Descriptors are non-linearly distributed

  • Upload
    shawn

  • View
    31

  • Download
    5

Embed Size (px)

DESCRIPTION

MVSV. Descriptor-level. X. X. M-PCCA. SVM. Fusion. GMM. Fusion. SVM. Score. Score. Y. Y. GMM. (linear) Kernel-level. X. Fusion. SVM. GMM. Score. Y. GMM. SVM. Score-level. X. Score. Fusion. Score. SVM. GMM. Score. Y. Multi-View Super Vector for Action Recognition - PowerPoint PPT Presentation

Citation preview

Page 1: Non-linear Distribution of Descriptors   Descriptors are non-linearly distributed

Non-linear Distribution of Descriptors Descriptors are non-linearly distributed. Observed descriptor pair share common information.

Mixture model of Probabilistic CCA Tackle non-linearity with mixture model Encode shared information by specifying latent variable

Mathematical Formulation Let v = (x, y),

x, y satisfy

p(v | k) is Gaussian with

Comparison of performance on HMDB51.

Comparison of performance on UCF101.

Influence of Parameters Number of Components

Latent Dimension

Multi-View Super Vector for Action RecognitionZhuowei Cai1, Limin Wang1,2, Xiaojiang Peng1, Yu Qiao1,2

1Shenzhen Institutes of Advanced Technology2The Chinese University of Hong Kong

Motivation Multi-View Super Vector Experimental Results

Performance of Different Components

Challenges in Action Recognition Different feature descriptors capture different aspects of object feature.

Current fusion pipelines presume different properties between feature descriptors. Direct concatenation presumes strong

correlation between fusion descriptors. Kernel average and score fusion prefer

mutual independence between fusion descriptors.

Typical fusion pipelines:

We utilize the Mixture of Probabilistic Canonical Correlation Analyzers to model a pair of feature descriptors.

M-PCCA factorize descriptors from two sources into a share part between them and two parts specific to each of them.

These parts construct the super vector representation for subsequent classification.

Our Approach

Mixture model of Probabilistic CCA In each local Probabilistic CCA, shared information Z and private information λ jointly generate the observed descriptors

Shared information Z is extracted via concatenating latent information zk for each local PCCA

Private information gx is extracted by taking the gradients of the expected loglikelihood with respect to λx (and λy)

Z, gx and gy form the final MVSV.

X

Fusion

Y

X

Y

Descriptor-level

(linear) Kernel-level

GMM

GMM

Fusion

GMM

SVM Score

SVM Score

X

Y

Score-levelGMM

GMM

Fusion Score

SVM Score

SVM Score

X

Y

MVSV

M-PCCA Fusion SVM Score