International Journal of Innovative Research in Computer and Communication Engineering

ISSN Approved Journal | Impact factor: 8.771 | ESTD: 2013 | Follows UGC CARE Journal Norms and Guidelines

| Monthly, Peer-Reviewed, Refereed, Scholarly, Multidisciplinary and Open Access Journal | High Impact Factor 8.771 (Calculated by Google Scholar and Semantic Scholar | AI-Powered Research Tool | Indexing in all Major Database & Metadata, Citation Generator | Digital Object Identifier (DOI) |


TITLE AeroText: Air-Writing Recognition Based on Deep Convolutional Neural Network
ABSTRACT This paper presents AEROText, a novel human computer interaction system designed to interpret in-air finger gestures as written characters, captured in real-time via a standard webcam. The system facilitates a touchless and intuitive method of text input, thereby eliminating the need for physical writing instruments. The architecture incorporates a secure user management module with One-Time Password (OTP) email verification, a robust finger tracking algorithm employing OpenCV for hand segmentation and contour analysis, and a deep learning model for character recognition. The core of the recognition engine is a Convolutional Neural Network (CNN), constructed with PyTorch and trained on the EMNIST dataset, which accurately translates mid-air finger trajectories into digital text. The system demonstrates successful recognition of English alphabets and exhibits a unique capability by mapping certain predictions to the Devanagari script, indicating its potential for multilingual applications. This work constitutes a comprehensive proof-of-concept for the development of accessible, next-generation HCI interfaces, with significant implications for assistive technology and public interactive systems. Index Terms—human-computer interaction, computer vision, deep learning, air handwriting, gesture recognition, OpenCV, PyTorch, touchless interface, EMNIST
AUTHOR SAMYAK KAMBLE, SAHIL SHARMA, RITESH ZANWAR, PROF. KIRAN PATIL Department of Information Technology, G.H. Raisoni College of Engineering and Management, Pune, India
VOLUME 182
DOI DOI: 10.15680/IJIRCCE. 2026.1403098
PDF pdf/98_AeroText Air-Writing Recognition Based on Deep Convolutional Neural Network.pdf
KEYWORDS
References [1] G. Bradski, “The OpenCV Library,” Dr. Dobb’s Journal of Software Tools, vol. 25, no. 1, pp. 120–123, 2000.
[2] A. Paszke et al., “PyTorch: An Imperative Style, High-Performance Deep Learning Library,” in Advances in Neural Information Processing Systems, vol. 32, pp. 8024–8035, 2019.
[3] G. Cohen, S. Afshar, J. Tapson, and A. van Schaik, “EMNIST: An Extension of MNIST to Handwritten Letters,” arXiv preprint arXiv:1702.05373, 2017.
[4] Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-Based Learning Applied to Document Recogni-tion,” Proceedings of the IEEE, vol. 86, no. 11, pp. 2278–2324, 1998.
[5] J. C. Russ, The Image Processing Handbook, 7th ed. Boca Raton, FL: CRC Press, 2016.
[6] R. C. Gonzalez and R. E. Woods, Digital Image Processing, 4th ed. Pearson, 2018.
[7] S. Mitra and T. Acharya, “Gesture Recognition: A Survey,” IEEE Transactions on Systems, Man, and Cy-bernetics, Part C, vol. 37, no. 3, pp. 311–324, May 2007.
[8] P. Viola and M. Jones, “Rapid Object Detection Using a Boosted Cascade of Simple Features,” in Proc. IEEE Conf. Computer Vision and Pattern Recognition (CVPR), vol. 1, pp. I-511–I-518, 2001.
[9] K. He, X. Zhang, S. Ren, and J. Sun, “Deep Residual Learning for Image Recognition,” in Proc. IEEE Conf. Computer Vision and Pattern Recognition (CVPR), pp. 770–778, 2016.
[10] D. G. Lowe, “Distinctive Image Features from Scale-Invariant Keypoints,” International Journal of Com-puter Vision, vol. 60, no. 2, pp. 91–110, 2004.
[11] M. Z. Alom et al., “The History Began from AlexNet: A Comprehensive Survey on Deep Learning Ap-proaches,” arXiv preprint arXiv:1803.01164, 2018.
[12] S. Ren, K. He, R. Girshick, and J. Sun, “Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 39, no. 6, pp. 1137–1149, 2017.
[13] T.-Y. Lin et al., “Microsoft COCO: Common Objects in Context,” in Proc. European Conf. Computer Vision (ECCV), pp. 740–755, 2014.
[14] D. P. Kingma and J. Ba, “Adam: A Method for Stochastic Optimization,” arXiv preprint arXiv:1412.6980, 2014.
[15] J. R. Quinlan, “Induction of Decision Trees,” Machine Learning, vol. 1, no. 1, pp. 81–106, 1986.
[16] C. Cortes and V. Vapnik, “Support-Vector Networks,” Machine Learning, vol. 20, no. 3, pp. 273–297, 1995.
[17] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “ImageNet Classification with Deep Convolutional Neural Networks,” in Advances in Neural Information Processing Systems, vol. 25, pp. 1097–1105, 2012.
[18] S. Ioffe and C. Szegedy, “Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift,” in Proc. Int. Conf. Machine Learning (ICML), pp. 448–456, 2015.
[19] J. Shotton et al., “Real-Time Human Pose Recognition in Parts from Single Depth Images,” in Proc. IEEE Conf. Computer Vision and Pattern Recognition (CVPR), pp. 1297–1304, 2011.
[20] M. Young, The Technical Writer’s Handbook. Mill Valley, CA: University Science, 1989.
image
Copyright © IJIRCCE 2020.All right reserved