Shrivastava, Anshumali2019-05-172019-05-172017-122017-11-30December 2Liu, Wanjia. "Retinomorphic Event-Driven Representations for Video Tasks." (2017) Master’s Thesis, Rice University. <a href="https://hdl.handle.net/1911/105601">https://hdl.handle.net/1911/105601</a>.https://hdl.handle.net/1911/105601Deep neural networks have revolutionized static image understanding through their amazing performance in tasks such as classification, segmentation and style transfer. However, current architectures have yet to find as much success in video tasks mainly due to increased data dimension, higher information throughput and a sub-optimal frame-driven representation. Inspired by the early vision systems of mammals and insects, we propose an event-driven input representation (EDR) that models several major properties of early retinal circuits: 1. Output scales logrithmically with input intensity, 2. ON/OFF pathways for differentiating event types, 3. Relative change detection and event detection via thresholding. With UCF-101 video action recognition experiments, we show that neural network utilizing EDR as input performs near state-of-the-art in accuracy while achieving a 1,500x speedup in input representation processing (9K frames/sec), as compared to optical flow. EDR's fast processing speed enables real-time inference/learning in time sensitive video applications such as reinforcement learning. In this vein, we use EDR to demonstrate performance improvements over state-of-the-art reinforcement learning algorithms for Atari games, something that has not been possible with optical flow.application/pdfengCopyright is held by the author, unless otherwise indicated. Permission to reuse, publish, or reproduce the work beyond the bounds of fair use or other exemptions to copyright law must be obtained from the copyright holder.RetinomorphicEvent-DrivenMotion RepresentationOptical-FlowAction RecognitionReinforcement LearningRetinomorphic Event-Driven Representations for Video TasksThesis2019-05-17