Neural Machine Translation of Spanish-English Food Recipes Using LSTM

Khen Dedes - Universitas Negeri Malang, Malang, East Java, Indonesia
Agung Bella Putra Utama - Universitas Negeri Malang, Malang, East Java, Indonesia
Aji Prasetya Wibawa - Universitas Negeri Malang, Malang, East Java, Indonesia
Arif Nur Afandi - Universitas Negeri Malang, Malang, East Java, Indonesia
Anik Nur Handayani - Universitas Negeri Malang, Malang, East Java, Indonesia
Leonel Hernandez - Instituciòn Universitaria ITSA, Colombia


Citation Format:



DOI: http://dx.doi.org/10.30630/joiv.6.2.804

Abstract


Nowadays, food is one of the things that has been globalized, and everyone from different parts of the world has been able to cook food from other countries through existing online recipes. Based on that, this study developed a translation formula using a neural machine translation (NMT). NMT is a recently proposed approach to machine translation. Unlike the traditional statistical machine translation, the neural machine translation aims at building a single neural network that can be jointly tuned to maximize the translation performance. The models proposed recently for neural machine translation often belong to a family of encoder–decoders. Our experiment led to novel insights and practical advice for building and extending NMT with the applied long short-term memory (LSTM) method to 47 bilingual food recipes between Spanish-English and English-Spanish. LSTM is one of the best machine learning methods for translating languages because it can retain memories for an extended period concurrently, grasp complicated connections between data, and provides highly useful information in deciding translation outcomes. The evaluation for this neural machine translation is to use BLEU. The comparing results show that the translation of recipes from Spanish-English has a better BLEU value of 0.998426 than English-Spanish with a data-sharing of 70%:30% during epoch 1000. Researchers can convert the country's popular cuisine recipes into another language for further research, allowing it to become more widely recognized abroad.

Keywords


Natural language processing; neural machine translation; food recipes; LSTM.

Full Text:

PDF

References


A. Salvador, M. Drozdzal, X. Giro-I-Nieto, and A. Romero, “Inverse cooking: Recipe generation from food images,†Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit., vol. 2019-June, pp. 10445–10454, 2019, doi: 10.1109/CVPR.2019.01070.

T. Sato, J. Harashima, and M. Komachi, “Japanese-English Machine Translation of Recipe Texts,†Wat, pp. 58–67, 2017.

S. Mori, T. Sasada, Y. Yamakata, and K. Yoshino, “A Machine Learning Approach to Recipe Text Processing,†Comput. Sci., 2012.

H. Tenzer and T. Schuster, Language Barriers in Different Forms of International Assignments, no. December. 2017.

A. Aguirregoitia-Martínez and M. D. Fernández-Poyatos, “The Gestation of Modern Gastronomy in Spain (1900-1936),†Cult. Hist. Digit. J., vol. 6, no. 2, p. 019, Nov. 2017, doi: 10.3989/chdj.2017.019.

M. J. Wild, “Eating Spain - National Cuisine Since 1900,†2015.

T. Prasad and R. B. Korrapati, “Application of Search & Sorting Techniques – in Natural Language Processing,†vol. 0869, no. 2, pp. 18–21, 2017.

K. Kim, “New canonical decomposition and composition processes for Hangeul,†Comput. Stand. Interfaces, vol. 24, no. 1, pp. 69–82, Mar. 2002, doi: 10.1016/S0920-5489(01)00098-8.

C. C. Emezue and F. P. B. Dossou, “FFR v1.1: Fon-French Neural Machine Translation,†pp. 83–87, 2020, doi: 10.18653/v1/2020.winlp-1.21.

J. J. Webster and C. Kit, “Tokenization as the initial phase in NLP,†no. January 1992, p. 1106, 1992, doi: 10.3115/992424.992434.

G. V. A. Gutiérrez, “A Comparative Study of NLP and Machine Learning Techniques for Sentiment Analysis and Topic Modeling on Amazon,†Int. J. Comput. Sci. Eng., vol. 9, no. 2, pp. 159–170, Apr. 2020, doi: 10.21817/ijcsenet/2020/v9i2/200902007.

M. IŞIK and H. DAĞ, “The impact of text preprocessing on the prediction of review ratings,†TURKISH J. Electr. Eng. Comput. Sci., vol. 28, no. 3, pp. 1405–1421, May 2020, doi: 10.3906/elk-1907-46.

D. P. Kingma and J. L. Ba, “Adam: A method for stochastic optimization,†3rd Int. Conf. Learn. Represent. ICLR 2015 - Conf. Track Proc., pp. 1–15, 2015.

K. Wołk and K. Marasek, “Enhanced Bilingual Evaluation Understudy,†Lect. Notes Inf. Theory, no. July, 2014, doi: 10.12720/lnit.2.2.191-197.

K. Papineni, S. Roukos, T. Ward, and W.-J. Zhu, “BLEU: a Method for Automatic Evaluation of Machine Translation,†in Proceedings of the 40th Annual Meeting on Association for Computational Linguistics - ACL ’02, 2001, p. 311, doi: 10.3115/1073083.1073135.