Machine Learning Based Customized Solution for Deafness and Mute People

Abstract

Individuals with hearing or speech impairments often possess a higher degree of familiarity with sign language due to its widespread usage. Consequently, a noticeable communication divide exists between those with such impairments and the general populace. A pivotal solution in bridging this gap involves employing human sign language interpreters. Regrettably, the scarcity of sign language interpreters worldwide in comparison to the number of individuals with hearing or speech impairments has resulted in certain individuals being unable to bear the cost of a human interpreter each time they engage in conversation. To address this issue, it is imperative to automate communication in a manner that liberates the deaf community from dependence on human translators. This article centers on the development of an application capable of real-time conversion between American Sign Language and text, along with supplementary functionalities aimed at dismantling the communication barriers faced by individuals with hearing or speech challenges when interacting with the broader public. the primary objective of this application is to accurately perceive and interpret the user's sign language. The initial stage involves training the system to recognize and interpret signs utilizing algorithms for object recognition and motion tracking. For this purpose, a convolutional neural network model was trained employing a meticulously curated American Sign Language dataset. Subsequently, the recognized signs are translated into English, forming grammatically sound words. Incorporated within the system are a language translator and a virtual sign keyboard, augmenting its capabilities. Moreover, a text-to-text transformer, structured on an encoder-decoder framework, is harnessed to identify grammatical inaccuracies and generate coherent phrases. For a comprehensive understanding of the process, detailed elaboration will follow in subsequent sections.

Country : Sri Lanka

1 Suranjini Silva2 Thilini Jayalath3 Madhushani E.A.Y.C.4 Amarasooriya H.D.5 Elpitiya S.N.6 Frank Perera

  1. Department of Information Technology, Sri Lanka Institute of Information Technology, New Kandy Road, Malabe, Sri Lanka
  2. Department of Information Technology, Sri Lanka Institute of Information Technology, New Kandy Road, Malabe, Sri Lanka
  3. Department of Information Technology, Sri Lanka Institute of Information Technology, New Kandy Road, Malabe, Sri Lanka
  4. Department of Information Technology, Sri Lanka Institute of Information Technology, New Kandy Road, Malabe, Sri Lanka
  5. Department of Information Technology, Sri Lanka Institute of Information Technology, New Kandy Road, Malabe, Sri Lanka
  6. Department of Information Technology, Sri Lanka Institute of Information Technology, New Kandy Road, Malabe, Sri Lanka

IRJIET, Volume 8, Issue 1, January 2024 pp. 62-68

doi.org/10.47001/IRJIET/2024.801008

References

  1. G. Eason, B. Noble, and I. N. Sneddon, “On certain integrals of Lipschitz-Hankel type involving products of Bessel functions,” Phil. Trans. Roy. Soc. London, vol. A247, pp. 529–551, April 1955. (references)
  2. J. Clerk Maxwell, A Treatise on Electricity and Magnetism, 3rd ed., vol. 2. Oxford: Clarendon, 1892, pp.68–73.
  3. I.S. Jacobs and C. P. Bean, “Fine particles, thin films and exchange anisotropy,” in Magnetism, vol. III, G. T. Rado and H. Suhl, Eds. New York: Academic, 1963, pp. 271–350.
  4. K. Elissa, “Title of paper if known,” unpublished.
  5. R. Nicole, “Title of paper with only first word capitalized,” J. Name Stand. Abbrev., in press.
  6. Y. Yorozu, M. Hirano, K. Oka, and Y. Tagawa, “Electron spectroscopy studies on magneto-optical media and plastic substrate interface,” IEEE Transl. J. Magn. Japan, vol. 2, pp. 740–741, August 1987 [Digests 9th Annual Conf. Magnetics Japan, p. 301, 1982].
  7. M. Young, The Technical Writer’s Handbook. Mill Valley, CA: University Science, 1989.
  8. Davis, K., Biddulph, R., and Balashek, S., “Automatic Recognition of Spoken Digit,” J. Acoust. Soc. Am. 24: Nov 1952, p. 637.
  9. Hemdal, J.F. and Hughes, G.W., A feature-based computer recognition program for the modeling of vowel perception, in Models for the Perception of Speech and Visual Form., MIT Press, Cambridge, MA. The Journal of the Acoustical Society of America 57 (2), 476-482, 1975.
  10. Kain and M. W. Macon, “Spectral voice conversion for text-to-speech synthesis,” in ICASSP, vol. 1, 1998, pp. 285–288.
  11. Watcher, M. D., Matton, M., Demuynck, K., Wambacq, P., Cools, R., “Template Based Continuous Speech Recognition”, IEEE Transaction.
  12. J. S. Sonkusare, N. B. Chopade, R. Sor, and S. L. Tade, “A review on hand gesture recognition system,” 2015, doi: 10.1109/ICCUBEA.2015.158.
  13. U. Patel and A. G. Ambekar, “Moment Based Sign Language Recognition for Indian Languages,” 2017 Int. Conf. Comput. Commun. Control Autom. ICCUBEA 2017, pp. 1–6, 2018, doi: 10.1109/ICCUBEA.2017.8463901.
  14. takelessons, 2019. ASL sentence structure: Word order in American sign language LASL lessons. YouTube. Available at: https://www.youtube.com/watch?v=nzrbvMeoBnE [Accessed. February 1, 2022].
  15. Z. A. Memon, M. U. Ahmed, S. T. Hussain, Z. A. Baig and U. Aziz, "Real Time Translator for Sign Languages," 2017 International Conference on Frontiers of Information Technology (FIT), 2017, pp. 144-148, doi: 10.1109/FIT.2017.00033.
  16. T. Vichyaloetsiri and P. Wuttidittachotti, "Web Service framework to translate text into sign language," 2017 International Conference on Computer, Information and Telecommunication Systems (CITS), 2017, pp. 180-184, doi: 10.1109/CITS.2017.8035336.
  17. International Advanced Research Journal in Science, Engineering and Technology, Vol. 8, Issue 4, April 2021, DOI: 10.17148/IARJSET.2021.8412.