Dictionary-based face recognition from video

Yi Chen Chen, Vishal M. Patel, P. Jonathon Phillips, Rama Chellappa

Research output: Chapter in Book/Report/Conference proceedingConference contribution

87 Scopus citations

Abstract

The main challenge in recognizing faces in video is effectively exploiting the multiple frames of a face and the accompanying dynamic signature. One prominent method is based on extracting joint appearance and behavioral features. A second method models a person by temporal correlations of features in a video. Our approach introduces the concept of video-dictionaries for face recognition, which generalizes the work in sparse representation and dictionaries for faces in still images. Video-dictionaries are designed to implicitly encode temporal, pose, and illumination information. We demonstrate our method on the Face and Ocular Challenge Series (FOCS) Video Challenge, which consists of unconstrained video sequences. We show that our method is efficient and performs significantly better than many competitive video-based face recognition algorithms.

Original languageEnglish (US)
Title of host publicationComputer Vision, ECCV 2012 - 12th European Conference on Computer Vision, Proceedings
Pages766-779
Number of pages14
EditionPART 6
DOIs
StatePublished - 2012
Externally publishedYes
Event12th European Conference on Computer Vision, ECCV 2012 - Florence, Italy
Duration: Oct 7 2012Oct 13 2012

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
NumberPART 6
Volume7577 LNCS
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Other

Other12th European Conference on Computer Vision, ECCV 2012
Country/TerritoryItaly
CityFlorence
Period10/7/1210/13/12

All Science Journal Classification (ASJC) codes

  • Theoretical Computer Science
  • General Computer Science

Fingerprint

Dive into the research topics of 'Dictionary-based face recognition from video'. Together they form a unique fingerprint.

Cite this