Enhancing Cross Language for English-Telugu pairs through the Modified Transformer Model based Neural Machine Translation
DOI:
https://doi.org/10.22399/ijcesen.1740Keywords:
Cross Language Translation, Transformer Networks, Neural machine translation, Feed Forward networks, Multi-scale attention mapsAbstract
Cross-Language Translation (CLT) refers to conventional automated systems that generate translations between natural languages without human involvement. As the most of the resources are mostly available in English, multi-lingual translation is badly required for the penetration of essence of the education to the deep roots of society. Neural machine translation (NMT) is one such intelligent technique which usually deployed for an efficient translation process from one source of language to another language. But these NMT techniques substantially requires the large corpus of data to achieve the improved translation process. This bottleneck makes the NMT to apply for the mid-resource language compared to its dominant English counterparts. Although some languages benefit from established NMT systems, creating one for low-resource languages is a challenge due to their intricate morphology and lack of non-parallel data. To overcome this aforementioned problem, this research article proposes the modified transformer architecture for NMT to improve the translation efficiency of the NMT. The proposed NMT framework, consist of Encoder-Decoder architecture which consist of enhanced version of transformer architecture with the multiple fast feed forward networks and multi-headed soft attention networks. The designed architecture extracts word patterns from a parallel corpus during training, forming an English–Telugu vocabulary via Kaggle, and its effectiveness is evaluated using measures like Bilingual Evaluation Understudy (BLEU), character-level F-score (chrF) and Word Error Rate (WER). To prove the excellence of the proposed model, extensive comparison between the proposed and existing architectures is compared and its performance metrics are analysed. Outcomes depict that the proposed architecture has shown the improvised NMT by achieving the BLEU as 0.89 and low WER when compared to the existing models. These experimental results promise the strong hold for further experimentation with the multi-lingual based NMT process.
References
Chen, Z., Jiang, C., & Tu, K. (2023). Using interpretation methods for model enhancement. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language (pp. XX–XX). Singapore, December 6–10.
Yin, K., & Neubig, G. (2022). Interpreting language models with contrastive explanations. arXiv Preprint, arXiv:2202.10419.
Belinkov, Y., Màrquez, L., Sajjad, H., Durrani, N., Dalvi, F., & Glass, J. (2018). Evaluating layers of representation in neural machine translation on part-of-speech and semantic tagging tasks. arXiv Preprint, arXiv:1801.07772.
Ekin, A., Dale, S., Jacob, A., Tengyu, M., & Denny, Z. (2023). What learning algorithm is in-context learning? Investigations with linear models. In Proceedings of the International Conference on Learning Representations (ICLR) (pp. XX–XX). Kigali, Rwanda, May 1–5.
He, S., Tu, Z., & Wang, X. (2019). Towards understanding neural machine translation with word importance. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP) (pp. XX–XX). Hong Kong, China, November 3–7.
Qiang, J., Liu, K., Li, Y., Zhu, Y., Yuan, Y. H., Hu, X., & Ouyang, X. (2023). Chinese lexical substitution: Dataset and method. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language (pp. XX–XX). Singapore, December 6–10.
Kalpana, P., Almusawi, M., Chanti, Y., Sunil Kumar, V., & Varaprasad Rao, M. (2024). A deep reinforcement learning-based task offloading framework for edge-cloud computing. In Proceedings of the 2024 International Conference on Integrated Circuits and Communication Systems (ICICACS) (pp. 1–5). Raichur, India. https://doi.org/10.1109/ICICACS60521.2024.10498232
Tan, S., Shen, Y., Chen, Z., Courville, A., & Gan, C. (2023). Sparse universal transformer. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language (pp. XX–XX). Singapore, December 6–10.
Müller, M., Jiang, Z., Moryossef, A., Rios, A., & Ebling, S. (2023). Considerations for meaningful sign language machine translation based on glosses. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (pp. XX–XX). Toronto, Canada, July 9–14.
Kalpana, P., Narayana, P. L., Madhavi, S., Dasari, K., Smerat, A., & Akram, M. (2025). Health-Fots—A latency-aware fog-based IoT environment and efficient monitoring of body’s vital parameters in smart healthcare environment. Journal of Intelligent Systems and Internet of Things, 15(1), 144–156. https://doi.org/10.54216/JISIoT.150112
Kai, V., & Frank, K. (2024). Cluster-centered visualization techniques for fuzzy clustering results to judge single clusters. Applied Sciences, 14(1102).
Woosik, L., & Juhwan, L. (2024). Tree-based modeling for large-scale management in agriculture: Explaining organic matter content in soil. Applied Sciences, 14(1811).
Kalpana, P., Malleboina, K., Nikhitha, M., Saikiran, P., & Kumar, S. N. (2024). Predicting cyberbullying on social media in the big data era using machine learning algorithm. In Proceedings of the 2024 International Conference on Data Science and Network Security (ICDSNS) (pp. 1–7). Tiptur, India. https://doi.org/10.1109/ICDSNS62112.2024.10691297
Rikters, M., Pinnis, M., & Krišlauks, R. (2018). Training and adapting multilingual NMT for less-resourced and morphologically rich languages.
Aly, R., Caines, A., & Buttery, P. (2021). Efficient unsupervised NMT for related languages with cross-lingual language models and fidelity objectives. In Workshop on NLP for Similar Languages, Varieties, and Dialects
Kalpana, P., Kodati, S. Smitha, L., Sreekanth, D., Smerat, N., & Akram, M. (2025). Explainable AI-driven gait analysis using wearable internet of things (WIoT) and human activity recognition. Journal of Intelligent Systems and Internet of Things, 15(2), 55–75. https://doi.org/10.54216/JISIoT.150205
Prasanna, R. K., Sudharson, S., Reddy, A. A., Reddy, B. S. J. N., & Anvitha, V. (2023). A comparative analysis of transformers for multilingual neural machine translation. In Proceedings of the 2023 IEEE 7th Conference on Information and Communication Technology (CICT) (pp. 1–6). Jabalpur, India. https://doi.org/10.1109/CICT59886.2023.10455324
Ji, B., Zhang, Z., Duan, X., Zhang, M., Chen, B., & Luo, W. (2020). Cross-lingual pre-training based transfer for zero-shot neural machine translation. Proceedings of the AAAI Conference on Artificial Intelligence, 34(01), 115–122. https://doi.org/10.1609/aaai.v34i01.5341
Dabre, R., Chu, C., & Kunchukuttan, A. (2020). A survey of multilingual neural machine translation. ACM Computing Surveys, 53(5), Article 99. https://doi.org/10.1145/3406095
Kudugunta, S. R., Bapna, A., Caswell, I., Arivazhagan, N., & Firat, O. (2019). Investigating multilingual NMT representations at scale. arXiv Preprint, arXiv:1909.02197.
Ha, T. L., Niehues, J., & Waibel, A. (2016). Toward multilingual neural machine translation with universal encoder and decoder. arXiv Preprint, arXiv:1611.04798.
Kumari, D., Ekbal, A., Haque, R., Bhattacharyya, P., & Way, A. (2021). Reinforced NMT for sentiment and content preservation in low-resource scenarios. ACM Transactions on Asian and Low-Resource Language Information Processing, 20(4), Article 70. https://doi.org/10.1145/3450970
Chan, K. H., Ke, W., & Im, S. K. (2020). CARU: A content-adaptive recurrent unit for the transition of hidden state in NLP. In H. Yang, K. Pasupa, A. C. S. Leung, J. T. Kwok, J. H. Chan, & I. King (Eds.), Neural information processing (Lecture Notes in Computer Science, Vol. 12532, pp. 123–135). Springer.
Shinde, V. R., et al. (2024). Multilingual neural machine translation system for Indic languages. International Journal of Engineering Research & Technology (IJERT), 10(1).
Das, S. B., Panda, D., Mishra, T. K., Patra, B. K., & Ekbal, A. (2024). Multilingual neural machine translation for Indic to Indic languages. ACM Transactions on Asian and Low-Resource Language Information Processing, 23(5), Article 65. https://doi.org/10.1145/3652026
Lu, Y. (2024). Research on English-Chinese neural machine translation based on improved deep Q-network approach. Second International Conference on Data Science and Information System (ICDSIS), Hassan, India. https://doi.org/10.1109/ICDSIS61070.2024.10594422
Prasanna ASD, Latha CBC. (2023) Bi-Lingual Machine Translation Approach using Long Short–Term Memory Model for Asian Languages. Indian Journal of Science and Technology. 16(18):1357-1364. https://doi.org/10.17485/IJST/v16i18.176.
Das, S. B., Biradar, A., Mishra, T. K., & Patra, B. K. (2023). Improving multilingual neural machine translation system for Indic languages. ACM Transactions on Asian and Low-Resource Language Information Processing, 22(6), Article 169. https://doi.org/10.1145/3587932
Shailashree, K. S., Gupta, D., & Costa-Jussà, M. R. (2023). A voyage on neural machine translation for Indic languages. Procedia Computer Science, 218, 2694–2712. https://doi.org/10.1016/j.procs.2023.01.242.
Sharma, S., Diwakar, M., Singh, P., Singh, V., Kadry, S., & Kim, J. (2023). Machine translation systems based on classical-statistical-deep-learning approaches. Electronics, 12(1716). https://doi.org/10.3390/electronics12071716
Kandimalla, Akshara & Lohar, Pintu & Maji, Kumar & Way, Andy. (2022). Improving English-to-Indian Language Neural Machine Translation Systems. Information. 13. 245. 10.3390/info13050245.
Andrabi, S. B., & Wahid, A., et al. (2022). Machine translation system using deep learning for English to Urdu. [Journal Name]. https://doi.org/10.1155/2022/7873012
Saini, S., & Sahula, V. (2020). Setting up a neural machine translation system for English to Indian languages. In G. R. Sinha & J. S. Suri (Eds.), Cognitive informatics, computer modelling, and cognitive science (pp. 195–212). Academic Press. https://doi.org/10.1016/B978-0-12-819443-0.00011-8.
https://opus.nlpl.eu/results/en&te/corpus-result-table
https://huggingface.co/datasets/allenai/nllb
LAVUDIYA, N. S., & C.V.P.R Prasad. (2024). Enhancing Ophthalmological Diagnoses: An Adaptive Ensemble Learning Approach Using Fundus and OCT Imaging. International Journal of Computational and Experimental Science and Engineering, 10(4). https://doi.org/10.22399/ijcesen.678
P. Padma, & G. Siva Nageswara Rao. (2024). CBDC-Net: Recurrent Bidirectional LSTM Neural Networks Based Cyberbullying Detection with Synonym-Level N-Gram and TSR-SCSOFeatures. International Journal of Computational and Experimental Science and Engineering, 10(4). https://doi.org/10.22399/ijcesen.623
Rajani Kumari Inapagolla, & K . Kalyan Babu. (2025). Audio Fingerprinting to Achieve Greater Accuracy and Maximum Speed with Multi-Model CNN-RNN-LSTM in Speaker Identification: Speed with Multi-Model CNN-RNN-LSTM in Speaker Identification. International Journal of Computational and Experimental Science and Engineering, 11(1). https://doi.org/10.22399/ijcesen.1138
Mekala, B., Neelamadhab Padhy, & Kiran Kumar Reddy Penubaka. (2025). Brain Tumor Segmentation and Detection Utilizing Deep Learning Convolutional Neural Networks: Enhanced Medical Image for Precise Tumor Localization and Classification. International Journal of Computational and Experimental Science and Engineering, 11(1). https://doi.org/10.22399/ijcesen.1051
Rajitha Kotoju, B.N.V. Uma Shankar, Ravinder Reddy Baireddy, M. Aruna, Mohammed Abdullah Mohammed Alnaser, & Imad Hammood Sharqi. (2025). A Deep auto encoder based Framework for efficient weather forecasting. International Journal of Computational and Experimental Science and Engineering, 11(1). https://doi.org/10.22399/ijcesen.429
M. Kannan, & K.R. Ananthapadmanaban. (2025). Students Performance prediction by EDA analysis and Hybrid Deep Learning Algorithms. International Journal of Computational and Experimental Science and Engineering, 11(2). https://doi.org/10.22399/ijcesen.1524
Kumar, N., & T. Christopher. (2025). Enhanced hybrid classification model algorithm for medical dataset analysis. International Journal of Computational and Experimental Science and Engineering, 11(1). https://doi.org/10.22399/ijcesen.611
Olola, T. M., & Olatunde, T. I. (2025). Artificial Intelligence in Financial and Supply Chain Optimization: Predictive Analytics for Business Growth and Market Stability in The USA. International Journal of Applied Sciences and Radiation Research, 2(1). https://doi.org/10.22399/ijasrar.18
A, V., & J Avanija. (2025). AI-Driven Heart Disease Prediction Using Machine Learning and Deep Learning Techniques. International Journal of Computational and Experimental Science and Engineering, 11(2). https://doi.org/10.22399/ijcesen.1669
Ibeh, C. V., & Adegbola, A. (2025). AI and Machine Learning for Sustainable Energy: Predictive Modelling, Optimization and Socioeconomic Impact In The USA. International Journal of Applied Sciences and Radiation Research, 2(1). https://doi.org/10.22399/ijasrar.19
Polatoglu, A. (2024). Observation of the Long-Term Relationship Between Cosmic Rays and Solar Activity Parameters and Analysis of Cosmic Ray Data with Machine Learning. International Journal of Computational and Experimental Science and Engineering, 10(2). https://doi.org/10.22399/ijcesen.324
Fowowe, O. O., & Agboluaje, R. (2025). Leveraging Predictive Analytics for Customer Churn: A Cross-Industry Approach in the US Market. International Journal of Applied Sciences and Radiation Research , 2(1). https://doi.org/10.22399/ijasrar.20
Kumar, A., & Beniwal, S. (2025). Depression Sentiment Analysis using Machine Learning Techniques:A Review. International Journal of Computational and Experimental Science and Engineering, 11(1). https://doi.org/10.22399/ijcesen.851
Hafez, I. Y., & El-Mageed, A. A. A. (2025). Enhancing Digital Finance Security: AI-Based Approaches for Credit Card and Cryptocurrency Fraud Detection. International Journal of Applied Sciences and Radiation Research, 2(1). https://doi.org/10.22399/ijasrar.21
S, P., & A, P. (2024). Secured Fog-Body-Torrent : A Hybrid Symmetric Cryptography with Multi-layer Feed Forward Networks Tuned Chaotic Maps for Physiological Data Transmission in Fog-BAN Environment. International Journal of Computational and Experimental Science and Engineering, 10(4). https://doi.org/10.22399/ijcesen.490
García, R., Carlos Garzon, & Juan Estrella. (2025). Generative Artificial Intelligence to Optimize Lifting Lugs: Weight Reduction and Sustainability in AISI 304 Steel. International Journal of Applied Sciences and Radiation Research, 2(1). https://doi.org/10.22399/ijasrar.22
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2025 International Journal of Computational and Experimental Science and Engineering

This work is licensed under a Creative Commons Attribution 4.0 International License.