A novel recurrent neural networks architecture for behavior analysis

Neziha Jaouedi, Noureddine Boujnah, Mohamed Bouhlel

Research output: Contribution to journalArticlepeer-review

3 Citations (Scopus)


Behavior analysis is an important yet challenging task on computer vision area. However, human behavior is still a necessity in differents sectors. In fact, in the increase of crimes, everyone needs video surveillance to keep their belongings safe and to automatically detect events by collecting important information for the assistance of security guards. Moreover, the surveillance of human behavior is recently used in medicine fields to quickly detect physical and mental health problems of patients. The complex and the variety presentation of human features in video sequence encourage researches to find the effective presentation. An effective presentation is the most challenging part. It must be invariant to changes of point of view, robust to noise and efficient with a low computation time. In this paper, we propose new model for human behavior analysis which combine transfer learning model and Recurrent Neural Network (RNN). Our model can extract human features from frames using the pre-trained model of Convolutional Neural Network (CNN) the Inception V3. The human features obtained are trained using RNN with Gated Recurrent Unit (GRU). The performance of our proposed architecture is evaluated by three different dataset for human action, UCF Sport, UCF101 and KTH, and achieved good classification accuracy.

Original languageEnglish
Pages (from-to)133-139
Number of pages7
JournalInternational Arab Journal of Information Technology
Issue number2
Publication statusPublished - 2021
Externally publishedYes


  • Activity recognition
  • Behavior modelling
  • Convolutional neural network
  • Deep learning
  • Gated recurrent unit
  • Recurrent neural networks
  • Video classification


Dive into the research topics of 'A novel recurrent neural networks architecture for behavior analysis'. Together they form a unique fingerprint.

Cite this