Skip to yearly menu bar Skip to main content


Poster

Deep Learning of invariant features via tracked video sequences

Will Y Zou · Andrew Y Ng · Shenghuo Zhu · Kai Yu

Harrah’s Special Events Center 2nd Floor

Abstract:

We use video sequences produced by tracking as training data to learn invariant features. These features are spatial instead of temporal, and well suited to extract from still images. With a temporal coherence objective, a multi-layer neural network encodes invariance that grow increasingly complex with layer hierarchy. Without fine-tuning with labels, we achieve competitive performance on five non-temporal image datasets and state-of-the-art classification accuracy 61% on STL-10 object recognition dataset.

Live content is unavailable. Log in and register to view live content