Lin, Yan ChingYan ChingLinHu, Min ChunMin ChunHuWEN-HUANG CHENGHsieh, Yung HuanYung HuanHsiehChen, Hong MingHong MingChen2023-03-132023-03-132012-12-269781450310895https://scholars.lib.ntu.edu.tw/handle/123456789/629232Observing the widespread use of Kinect-like depth cameras, in this work, we investigate into the problem of using sole depth data for human action recognition and retrieval in videos. We proposed the use of simple depth descriptors without learning optimization to achieve promising performances as compatible to those of the leading methods based on color images and videos, and can be effectively applied for real-time applications. Because of the infrared nature of depth cameras, the proposed approach will be especially useful under poor lighting conditions, e.g. the surveillance environments without sufficient lighting. Meanwhile, we proposed a large Depth-included Human Action video dataset, namely DHA, which contains 357 videos of performed human actions belonging to 17 categories. To the best of our knowledge, the DHA is one of the largest depth-included video datasets of human actions. © 2012 ACM.depth information | human action recognition | human action video retrieval[SDGs]SDG1Human action recognition and retrieval using sole depth informationconference paper10.1145/2393347.23963812-s2.0-84871359520https://api.elsevier.com/content/abstract/scopus_id/84871359520