Show simple item record

dc.contributor.advisorNababan, Erna Budhiarti
dc.contributor.advisorMawengkang, Herman
dc.contributor.authorNasution, Nur Amalia
dc.date.accessioned2024-08-27T08:50:27Z
dc.date.available2024-08-27T08:50:27Z
dc.date.issued2024
dc.identifier.urihttps://repositori.usu.ac.id/handle/123456789/96206
dc.description.abstractThis research examines the performance of the Long Short-Term Memory (LSTM) algorithm in combination with two word embedding techniques, FastText and Word2Vec, for translating text between the Batak and English languages. LSTM, an advanced form of Recurrent Neural Networks (RNNs), is utilized for its capability to handle sequential data and maintain long-term dependencies. However, LSTM's effectiveness in translation tasks is significantly influenced by the quality of word embeddings, which provide low-dimensional vector representations of words, capturing their semantic and contextual relationships. This study conducted a comparative analysis of LSTM's performance using FastText and Word2Vec embeddings. Data comprising 28,420 Batak-English sentence pairs were collected from various sources, including the Lets Read Asia website and the "Kamus Batak Toba - Indonesia" dictionary. The sentences were then embedded using both FastText and Word2Vec techniques, and the resulting vectors were fed into the LSTM model. The LSTM model, incorporating encoder and decoder components, was trained over multiple epochs, and its performance was evaluated using the BLEU (Bilingual Evaluation Understudy) score. This metric compares n-grams of the predicted translations with reference translations, providing a measure of translation accuracy. The results indicate that the LSTM model with FastText embeddings consistently outperformed the model with Word2Vec embeddings. The FastText-based model achieved an average BLEU score of 0.9516, compared to 0.9389 for the Word2Vec-based model. This superior performance is attributed to FastText's ability to handle out-of-vocabulary words by leveraging subword information, thus providing more accurate and contextually relevant translations.en_US
dc.language.isoiden_US
dc.publisherUniversitas Sumatera Utaraen_US
dc.subjectLSTMen_US
dc.subjectFastTexten_US
dc.subjectWord2Vecen_US
dc.subjectMachine Translationen_US
dc.subjectBatak-Englishen_US
dc.subjectSDGsen_US
dc.titlePeningkatan Akurasi Long Short-Term Memori (LSTM) Menggunakan Word2Vec dan Fastext untuk Machine Translation Bahasa Batak-Inggrisen_US
dc.title.alternativeImproving The Accuracy of Long Short-Term Memory (LSTM) Using Word2Vec and Fastext for Batak-English Machine Translationen_US
dc.typeThesisen_US
dc.identifier.nimNIM207038026
dc.identifier.nidnNIDN0026106209
dc.identifier.nidnNIDN8859540017
dc.identifier.kodeprodiKODEPRODI55101#Teknik Informatika
dc.description.pages70 Pagesen_US
dc.description.typeTesis Magisteren_US


Files in this item

Thumbnail
Thumbnail

This item appears in the following Collection(s)

Show simple item record