Retinomorphic Event-Driven Representations for Video Tasks

Date
2017-11-30
Journal Title
Journal ISSN
Volume Title
Publisher
Abstract

Deep neural networks have revolutionized static image understanding through their amazing performance in tasks such as classification, segmentation and style transfer. However, current architectures have yet to find as much success in video tasks mainly due to increased data dimension, higher information throughput and a sub-optimal frame-driven representation. Inspired by the early vision systems of mammals and insects, we propose an event-driven input representation (EDR) that models several major properties of early retinal circuits: 1. Output scales logrithmically with input intensity, 2. ON/OFF pathways for differentiating event types, 3. Relative change detection and event detection via thresholding. With UCF-101 video action recognition experiments, we show that neural network utilizing EDR as input performs near state-of-the-art in accuracy while achieving a 1,500x speedup in input representation processing (9K frames/sec), as compared to optical flow. EDR's fast processing speed enables real-time inference/learning in time sensitive video applications such as reinforcement learning. In this vein, we use EDR to demonstrate performance improvements over state-of-the-art reinforcement learning algorithms for Atari games, something that has not been possible with optical flow.

Description
Degree
Master of Science
Type
Thesis
Keywords
Retinomorphic, Event-Driven, Motion Representation, Optical-Flow, Action Recognition, Reinforcement Learning
Citation

Liu, Wanjia. "Retinomorphic Event-Driven Representations for Video Tasks." (2017) Master’s Thesis, Rice University. https://hdl.handle.net/1911/105601.

Has part(s)
Forms part of
Published Version
Rights
Copyright is held by the author, unless otherwise indicated. Permission to reuse, publish, or reproduce the work beyond the bounds of fair use or other exemptions to copyright law must be obtained from the copyright holder.
Link to license
Citable link to this page