The initial development of two First-Person Perspective Video Activity Recognition Systems is discussed. The first system, the First Person Fall Detection or UFall, can be used to recognize when a person wearing or holding the mobile vision system has fallen. The problem of fall detection is tackled from the unique first-person perspective. The second system, the directed CrossWalk System (UCross), involves detection of the user movement across a crosswalk and is intended for use in helping a low vision person navigate. In both cases, the user is wearing or holding the camera device for purposes of monitoring or inspection of the environment. This first-person perspective yields unusual fall data and this is captured and used for the creation of a fall detection system. For both systems Machine Learning is employed using video input to trained Long-Term Short-Term (LSTM) Networks. These first-perspective video activity recognition systems use the Tensorflow framework [1] and is deployed using mobile phones for proof of concept. These applications could be useful for low vision people and in the case of fall detection for senior citizens, police, construction and other inspection-oriented jobs to help users who have fallen. The success and challenges faced with this unique first-person perspective data are presented along with future avenues of work.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.