Gait-based Biometric Identification System using Triangulated Skeletal Models (TSM)
Human gait identification is a behavioral biometric technology which can be used to monitor human beings without user interaction. Recent researches are more focused on investigating gait as one of the biometric traits. Further, gait recognition aims to analyze and identify human behavioral activities and may be implemented in different scenarios including access control and criminal analysis. However, using various techniques in relation to image processing and obtaining better accuracy are remaining challenges. In last decade, Microsoft has introduced the Kinect sensor as an innovative sensor to provide image characteristics, precisely. Therefore, this article uses a Kinect sensor to extract gait characteristics to be used in individual recognition. A set of Triangulated shape are generated as new feature vector and called Triangulated Skeletal Model (TSM). Nearest Neighbor technique is utilized to do the recognition issue based on leave-one-out strategy. The experimental outcomes indicated that the recommended technique provides significant results and outperforms other comparative similar techniques with accuracy of 93.46%.
Choras, Michal. "Ear biometrics based on geometrical feature extraction." ELCVIA Electronic Letters on Computer Vision and Image Analysis 5.3 , 84-95,2005.
Sabir, Azhin Tahir, et al. "Human gait identification using Kinect sensor." Kurdistan Journal of Applied Research 2.3 : 142-146, 2017.
Mather, G., & Murdoch, L. "Gender discrimination in biological motion displays based on dynamic cues" Proceedings of the Royal Society of London. Series B: Biological Sciences 258, no. 1353 : 273-279, 1994
Shirke, Suvarna, S. S. Pawar, and Kamal Shah. "Literature review: Model free human gait recognition." 2014 Fourth International Conference on Communication Systems and Network Technologies. IEEE, 2014.
J. Wang, M. She, S. Nahavandi, and A. Kouzani, "A review of vision-based gait recognition methods for human identification," in Digital Image Computing: Techniques and Applications (DICTA), 2010 International Conference on, pp. 320-327, 2010.
Hong, S., Lee, H., Fareed, I. And An, N., “A New Gait Representation for Human Identification: Mass Vector,” Second IEEE Conference on Industrial Electronics and Applications 673, pp. 669-673, 2007.
Lee, C.-C.et al., 2011. Frame difference history image for gait recognition. s.l., Machine Learning and Cybernetics (ICMLC), International Conference on, pp. 1785-1788, 2011.
Bashir, K., Xiang, T., & Gong, S. "Feature selection on gait energy image for human identification." IEEE international conference on acoustics, speech and signal processing. IEEE, 2008.
Felez, M., Ramon, M., and J. Salvador S. "Human Recognition Based on Gait Poses." Pattern Recognition and Image Analysis, 347—354, 2011.
C. BenAbdelkader, R. Cutler, and L. Davis, "Stride and cadence as a biometric in automatic person identification and verification," in Automatic Face and Gesture Recognition, Proceedings. Fifth IEEE, 2002.
Singh, J. P. and Jain, S., “Person Identification Based on Gait using Dynamic Body Parameters,” IEEE Xplore., pp.248-252, 2010.
Zhang, Z. "Microsoft kinect sensor and its effect" IEEE multimedia, 19(2), 4-10, 2012.
J. Preis, M. Kessel, M. Werner, and C. Linnhoff-Popien, "Gait recognition with kinect," in 1st international workshop on kinect in pervasive computing, pp. P1-P4, 2012.
W. Zeng, X. Zheng, F. Liu, Y. Wang and Q. Wang, "A new Kinect-based frontal view gait recognition method via deterministic learning," in Control Conference (CCC), 2016 35th Chinese, Chengdu, China , pp. 3919-3923, 2016
Gianaria, E., Grangetto, M., Lucenteforte, M., & Balossino, N., "Human classification using gait features," in International Workshop on Biometric Authentication, Sofia, Bulgaria, pp.16-27, 2014.
M. Ahmed, Al-Jawad and A. T. Sabir, "Gait recognition based on Kinect sensor," in Real-Time Image and Video Processing, Brussels, Belgium, pp. 91390B, 2014
Min, Pa Pa, Shohel Sayeed, and Thian Song Ong. "Gait Recognition Using Deep Convolutional Features." 2019 7th International Conference on Information and Communication Technology (ICoICT). IEEE, 2019.
Wang, X., Zhang, J. and Yan, W.Q. . Gait recognition using multichannel convolution neural networks. Neural Computing and Applications, pp.1-11 2019.
S. M. H. Bari and M. L. Gavrilova, "Artificial Neural Network Based Gait Recognition Using Kinect Sensor," in IEEE Access, vol. 7, pp. 162708-162722, 2019.
Huan, Z., Chen, X., Lv, S., & Geng, H. (2019). Gait Recognition of Acceleration Sensor for Smart Phone Based on Multiple Classifier Fusion. Mathematical Problems in Engineering, 2019.
V. O. Andersson and RM. Araujo, "Person Identification Using Anthropometric and Gait Data from Kinect Sensor.," in AAAI, Austin,Texas , pp 425-431, 2015.
How to Cite
Copyright (c) 2021 Azhin Tahir Sabir
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
Authors who publish with this journal agree to the following terms:
- Authors retain copyright and grant the journal right of first publication with the work simultaneously licensed under a Creative Commons Attribution License [CC BY-NC-ND 4.0] that allows others to share the work with an acknowledgment of the work's authorship and initial publication in this journal.
- Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the journal's published version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgment of its initial publication in this journal.
- Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work (See The Effect of Open Access).
AJNU is committed to protecting the privacy of the users of this journal website. The names, personal particulars and e-mail addresses entered in this website will be used only for the stated purposes of this journal and will not be made available to third parties without the user's permission or due process. Users consent to receive communication from the AJNU for the stated purposes of the journal. Queries with regard to privacy may be directed to [email protected]