kzruen

Project: №BR24992875 «Development of sign language resources using artificial intelligence for people with hearing and speech impairments»

Purpose of the project:

Creating a system facilitating communication between deaf, hard of hearing and hearing people via translation of Kazakh-Russian sign language into written text and vice versa, and through spoken language conversion into gestures and text, will improve the information and communication access for people with multilevel hearing functions.

People with hearing and speech impairments use sign language to communicate, but their hearing interlocutors do not always understand it. Mastery of sign language, just like verbal language, plays an important role in the formation of the personality of a person with hearing and speech impairments, as well as in their psychological development and socialization. The development of a set of surdo-resources using artificial intelligence for people with hearing and speech impairments is one of the most important tasks facing modern society. Due to the lack of effective tools for translating and interpreting sign language, the opportunities of people with hearing and speech impairments in education, employment, and social integration are limited.

Kazakh-Russian Sign Language (KRSL) is a natural language used for communication by people with hearing impairments living in Kazakhstan. One of the key achievements was the development of a lexico-morphological analysis (LMA) module for processing texts in the Kazakh language. Within this framework, multimedia thesauri were created and expanded, a database of a sign language thesaurus dictionary was formed, and software for generating dactyl animation models was developed.

In addition, work has been carried out to create a sign language corpus, which made it possible to systematize and standardize the signs used in everyday communication. Software prototypes for automatic translation of words into signs were developed, and a database of multimedia materials was created, including video recordings of signs, dactyl models, and textual descriptions. These results formed the basis for the development of intelligent systems capable of providing inclusive communication between people with hearing impairments and hearing users.

Figure 1 – Machine learning and deep learning models

The process begins with data sources, which include Remote Motion Capture and the Kazakh-KSL Gloss Parallel Corpus. These data go through the stages of annotation and splitting into training, testing, and validation sets.

At the data processing stage, machine learning models are selected and evaluated. Models such as Transformer, Seq2Seq, and Seq2Seq with Attention (Seq2Seq+Attn) are considered. To assess the effectiveness of these models, evaluation metrics such as Accuracy, Recall, Precision, F1-score, Confusion Matrix, Word Error Rate (WER), Edit Distance, and BLEU are used.

At the next stage of data post-processing, various libraries and tools are used, including scikit-learn, Python, TensorFlow, PyTorch, pandas, NumPy, matplotlib, LIVE FACE App, iClone, and Perception Neuron 3. These tools help analyze and visualize data, as well as create animations for the avatar.

The final result is the visualization of an avatar performing gestures according to the input text, enabling real-time translation of text into sign language. This process demonstrates the integration of modern technologies and machine learning methods to enhance communication between humans and machines, particularly in the context of supporting people with hearing and speech impairments.

Publications:

Foreign publications:

  1. Yerimbetova A., Sakenov B., Sambetbayeva M., Daiyrbayeva E., Berzhanova U., Othman M. Creating a Parallel Corpus for the Kazakh Sign Languageand Learning. Sci. 2025, 15, 2808. https://doi.org/ 10.3390/app15052808
  2. Amangeldy, N., Milosz, M., Yerimbetova, A., Tursynova, N., Kurmetbek, B., & Gazizova, N. (2025, July). Current Trends in Developing Parallel Corpora for Text-to-Sign Language Translation. In International Conference on Computational Intelligence in Engineering Science(pp. 135-144). Cham: Springer Nature Switzerland. https://doi.org/10.1007/978-3-031-98164-7_11
  3. Buribayev, Z., Aouani, M., Zhangabay, Z., Yerkos, A., Abdirazak, Z., & Zhassuzak, M. (2025). Enhancing Kazakh Sign Language Recognition with BiLSTM Using YOLO Keypoints and Optical Flow. Applied Sciences15(10), 5685. https://doi.org/10.3390/app15105685

Domestic publications:

  1. Yerimbetova A., Sambetbayeva M., Daiyrbayeva E., Sakenov B Berzhanova U. (2025). Creation of a model for recognition of the Kazakh sign language using the deep learning method. News of the National Academy of Sciences of the Republic of Kazakhstan. Physics and Mathematics series, (1), 108–123. https://doi.org/10.32014/2025.2518-1726.328
  2. Bekarystankyzy A., Mamyrbayev O., Oralbekova, D., 1.    Yerimbetova А., & Turdalyuly M. М. (2025). TESTING THE AUDIO-TEXT DATASET FOR KAZAKH LANGUAGE USING THE CONFORMER ENCODER. Academic Scientific Journal of Computer Science, (2), 50–60. https://doi.org/10.32014/2025.2518-1726.343

Author's certificates:

  1. Abdrazak Z., Buribaev Zh., Zhasuzak M. “Kazakh-Sign-Language-Detector”. Copyright certificates № 50506, 16.10.2024.
  2. Yerimbetova A., Sakenov B., Daiyrbayeva E.«The program of automatic recognition of dactylic gestures of the Kazakh sign language». Copyright certificates № 50629, 21.10.2024.
  3. Yerimbetova A., Daiyrbayeva E., Sakenov B Berzhanova U. «Automated system for extracting gesture coordinates and synchronized audio from video for sign language analysis». Copyright certificates № 62079, 12.09.2025.

Video interview: