remarkable efficiency, surpassing CPU capabilities
by a significant margin of 202 times, and exhibiting
latency levels comparable to GPU processing, with
differences ranging from just 3 to 12 milliseconds.
Notably, while achieving impressive performance
metrics, FPGAs exhibit minimal power consumption,
consuming less than 36 watts, in contrast to the con-
siderable power requirements of GPUs (around 220
watts). Our findings not only contribute to the ex-
panding field of FPGA-accelerated deep learning but
also provide a valuable perspective on the feasibility
of deploying custom models for real-time skeleton-
based human action recognition.
References:
[1] Xilinx, “Dpuczdx8g for zynq ultra-
scale+ mpsocs product guide (pg338),”
https://docs.xilinx.com/r/en-US/pg338-dpu,
2023.
[2] A. Shahroudy, J. Liu, T.-T. Ng, and G. Wang,
“Ntu rgb+ d: A large scale dataset for 3d hu-
man activity analysis,” in Proceedings of the
IEEE conference on computer vision and pat-
tern recognition, pp. 1010–1019, 2016.
[3] P. Zhang, C. Lan, J. Xing, W. Zeng, J. Xue,
and N. Zheng, “View adaptive neural networks
for high performance skeleton-based human ac-
tion recognition,” IEEE transactions on pattern
analysis and machine intelligence, vol. 41, no. 8,
pp. 1963–1978, 2019.
[4] A. Banerjee, P. K. Singh, and R. Sarkar,
“Fuzzy integral-based cnn classifier fusion for
3d skeleton action recognition,” IEEE transac-
tions on circuits and systems for video technol-
ogy, vol. 31, no. 6, pp. 2206–2216, 2020.
[5] S. Li, W. Li, C. Cook, and Y. Gao, “Deep in-
dependently recurrent neural network (indrnn),”
arXiv preprint arXiv:1910.06251, 2019.
[6] C. Li, C. Xie, B. Zhang, J. Han, X. Zhen,
and J. Chen, “Memory attention networks
for skeleton-based action recognition,” IEEE
Transactions on Neural Networks and Learning
Systems, vol. 33, no. 9, pp. 4800–4814, 2021.
[7] S. Yan, Y. Xiong, and D. Lin, “Spatial tempo-
ral graph convolutional networks for skeleton-
based action recognition,” in Thirty-second
AAAI conference on artificial intelligence, 2018.
[8] Y.-F. Song, Z. Zhang, C. Shan, and L. Wang,
“Constructing stronger and faster baselines
for skeleton-based action recognition,” IEEE
Transactions on Pattern Analysis and Machine
Intelligence, 2022.
[9] A. Mansouri, T. Bakir, and S. Femmam, “Hu-
man action recognition with skeleton and in-
frared fusion model,” Journal of Image and
Graphics, vol. 11, no. 4, pp. 309–320, 2023.
[10] M. T. Ajili and Y. Hara-Azumi, “Multimodal
neural network acceleration on a hybrid cpu-
fpga architecture: A case study,” IEEE Access,
vol. 10, pp. 9603–9617, 2022.
[11] N. Attaran, A. Puranik, J. Brooks, and T. Mohs-
enin, “Embedded low-power processor for per-
sonalized stress detection,” IEEE Transactions
on Circuits and Systems II: Express Briefs,
vol. 65, no. 12, pp. 2032–2036, 2018.
[12] A. N. Mazumder, H. Ren, H.-A. Rashid,
M. Hosseini, V. Chandrareddy, H. Homayoun,
and T. Mohsenin, “Automatic detection of respi-
ratory symptoms using a low-power multi-input
cnn processor,” IEEE Design & Test, vol. 39,
no. 3, pp. 82–90, 2021.
[13] A. Jafari, A. Ganesan, C. S. K. Thalisetty,
V. Sivasubramanian, T. Oates, and T. Mohsenin,
“Sensornet: A scalable and low-power deep
convolutional neural network for multimodal
data classification,” IEEE Transactions on Cir-
cuits and Systems I: Regular Papers, vol. 66,
no. 1, pp. 274–287, 2018.
[14] A. N. Mazumder, H.-A. Rashid, and T. Mohs-
enin, “An energy-efficient low power lstm pro-
cessor for human activity monitoring,” in 2020
IEEE 33rd International System-on-Chip Con-
ference (SOCC), pp. 54–59, IEEE, 2020.
[15] A. Reiss and D. Stricker, “Introducing a new
benchmarked dataset for activity monitoring,”
in 2012 16th international symposium on wear-
able computers, pp. 108–109, IEEE, 2012.
[16] S. Yao, S. Hu, Y. Zhao, A. Zhang, and T. Ab-
delzaher, “Deepsense: A unified deep learning
framework for time-series mobile sensing data
processing,” in Proceedings of the 26th interna-
tional conference on world wide web, pp. 351–
360, 2017.
[17] B. Zhang, J. Han, Z. Huang, J. Yang, and
X. Zeng, “A real-time and hardware-efficient
processor for skeleton-based action recognition
with lightweight convolutional neural network,”
IEEE Transactions on Circuits and Systems II:
Express Briefs, vol. 66, no. 12, pp. 2052–2056,
2019.
[18] K. He, X. Zhang, S. Ren, and J. Sun, “Deep
residual learning for image recognition,” in Pro-
ceedings of the IEEE conference on computer
WSEAS TRANSACTIONS on COMPUTER RESEARCH
DOI: 10.37394/232018.2024.12.31
Amine Mansouri, Abdellah Elzaar,
Mahdi Madani, Toufik Bakir