Transcript slides
Deception Detection with
micro-expressions using
Kernelized Perceptron
Abhijit Roy (ar3579)
Why micro-expressions?
•
There has been a lot of interest in the field of studying micro-expressions
•
It is considered as one of the most effective clues to detect lies and
dangerous behaviors
•
Haggard and Isaacs first discovered micro-expression and considered it as
repressed emotions
•
Micro-expression has gained popularity recently because of its potential
applications in the process of diagnosis and national security
•
The Transportation Security Administration in the USA has already
employed Screening Passengers by Observation Techniques (SPOT) which
uses micro-expressions analysis
Why micro-expressions with machine
learning?
•
Higher accuracy in classification of data with higher number of features.
•
Each data point or example is a set of features representing the human face.
•
Less number of features makes it harder to train the model, but higher
number of features can help in better classification.
•
Micro-expressions can be used to create feature rich examples, which can be
used to train classifiers with an expectation of better accuracy.
Perceptron
Positive Example - Truth
Negative Example - Lie
Perceptron
•
Very efficient classification algorithm for linearly separable data.
•
Positive and negative examples for lie detection experiment are highly likely to
be linearly separable.
•
Works well with low dimensional data - that is data with few features.
•
Computationally expensive when run with examples having lot of features, that
is high dimensional data causes performance issues.
•
Micro-expression data is a high dimensional data having multiple features for
each example point.
Kernelized Perceptron
•
Perceptron has a dual form which can be run on high dimensional data, even
with infinite feature space.
•
Made possible due to use of Kernel function which is a simple
computationally inexpensive calculation.
•
Can be used with micro-expressions data, where each example is a high
dimensional vector with each dimension corresponding to a certain feature.
•
Expected to run in reasonable amount of time and still result in a good
classifier with high accuracy in predicting lies.
Data
•
The Chinese Academy of Sciences Micro-expression (CASME) database
•
Contains 195 micro-expressions filmed under 60fps
•
They were selected from more than 1500 elicited facial movements
•
These data points will be preprocessed and presented as labeled example
vectors which will then be fed to the Perceptron algorithm as training data
Results and Other Methods
•
The experiment will do a 70/30 split of the data into training set and holdout
validation set.
•
The training error will be calculated on the training data set, and the test
error will be calculated on the holdout validation set.
•
We can also do Principal Component Analysis to find the most significant
features and then run some other algorithms as well.
•
For data that is not clearly linearly separable, we can use soft margin SVMs
with kernel functions to get a good accuracy.