https://scholars.lib.ntu.edu.tw/handle/123456789/413023
標題: | Egocentric activity recognition by leveraging multiple mid-level representations | 作者: | Hsieh P.-J. Lin Y.-L. Chen Y.-H. WINSTON HSU |
關鍵字: | Egocentric Activity Recognition; Egocentric Video; Multiple Feature Fusion | 公開日期: | 2016 | 卷: | 2016-August | 來源出版物: | IEEE International Conference on Multimedia and Expo | 摘要: | Existing approaches for egocentric activity recognition mainly rely on a single modality (e.g., detecting interacting objects) to infer the activity category. However, due to the inconsistency between camera angle and subject's visual field, important objects may be partially occluded or missing in the video frames. Moreover, where the objects are and how we interact with the objects are usually ignored in prior works. To resolve these difficulties, we propose multiple mid-level representations (e.g., objects manipulated by a user, background context, and motion patterns of hands) to compensate the insufficiency of a single modality, and jointly consider what, where, and how a subject is interacting with. To evaluate the method, we introduce a new and challenging egocentric activity dataset (ADL+) that contains video and wrist-worn accelerometer data of people performing daily-life activities. Our approach significantly outperforms the state-of-the-art method on the public ADL dataset (i.e., 36.8% to 46.7%) and our ADL+ dataset (i.e., 32.1 % to 60.0%) in terms of classification accuracy. In addition, we also conduct a series of analyses to explore relative merits of each modality to egocentric activity recognition. ? 2016 IEEE. |
URI: | https://scholars.lib.ntu.edu.tw/handle/123456789/413023 | ISBN: | 9781467372589 | ISSN: | 19457871 | DOI: | 10.1109/ICME.2016.7552937 |
顯示於: | 資訊工程學系 |
在 IR 系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。