ATTENTION: The works hosted here are being migrated to a new repository that will consolidate resources, improve discoverability, and better show UTA's research impact on the global community. We will update authors as the migration progresses. Please see MavMatrix for more information.
Show simple item record
dc.contributor.advisor | Huber, Manfred | |
dc.creator | Gupta, Alankrit | |
dc.date.accessioned | 2020-09-10T14:21:17Z | |
dc.date.available | 2020-09-10T14:21:17Z | |
dc.date.created | 2020-08 | |
dc.date.issued | 2020-09-03 | |
dc.date.submitted | August 2020 | |
dc.identifier.uri | http://hdl.handle.net/10106/29386 | |
dc.description.abstract | The recognition of activities from video is a capability that is important for a wide range of applications, ranging from basic scene understanding to the successful prediction of behavior in autonomous vehicle applications. At this time, human capabilities in this task by far outperform computer applications and thus the idea to mimic human perception should be promising. In this thesis we are proposing an architecture that processes videos to extract important action instances that describe the essential behaviors contained in any video and help us map the information from the video to a machine-understandable form. This is an important research area, as it could help us interpret the surrounding environment for the visually impaired, detect and characterize human behavior for autonomous vehicles, as well as enhance security at some of the most vulnerable places by identifying suspicious behavior. All of this illustrates the vast range of possibilities to this technology. The architecture proposed here is divided into three major sub-modules, namely: i) Localization; ii) Action Detection; iii) Description mapping. In this thesis, all the submodules are introduced and their interaction and operation is described before the action detection module is implemented and its performance is demonstrated. In addition, the thesis will describe how we could use transfer learning to combine all the proposed specialized components to mimic human perception. | |
dc.format.mimetype | application/pdf | |
dc.language.iso | en_US | |
dc.subject | Human perception | |
dc.subject | Activity recognition | |
dc.title | ACTIVITY RECOGNITION TO MIMIC HUMAN PERCEPTION | |
dc.type | Thesis | |
dc.degree.department | Computer Science and Engineering | |
dc.degree.name | Master of Science in Computer Science | |
dc.date.updated | 2020-09-10T14:21:17Z | |
thesis.degree.department | Computer Science and Engineering | |
thesis.degree.grantor | The University of Texas at Arlington | |
thesis.degree.level | Masters | |
thesis.degree.name | Master of Science in Computer Science | |
dc.type.material | text | |
Files in this item
- Name:
- GUPTA-THESIS-2020.pdf
- Size:
- 544.6Kb
- Format:
- PDF
This item appears in the following Collection(s)
Show simple item record