Preprint Article Version 1 Preserved in Portico This version is not peer-reviewed

Learning Machine Translation with Linguistic Interpretation

Version 1 : Received: 12 December 2023 / Approved: 13 December 2023 / Online: 13 December 2023 (05:04:39 CET)

How to cite: Alkmim, K.; Patel, R.; Dave, D. Learning Machine Translation with Linguistic Interpretation. Preprints 2023, 2023120921. https://doi.org/10.20944/preprints202312.0921.v1 Alkmim, K.; Patel, R.; Dave, D. Learning Machine Translation with Linguistic Interpretation. Preprints 2023, 2023120921. https://doi.org/10.20944/preprints202312.0921.v1

Abstract

The Transformer architecture, while adept at capturing context through self-attention, falls short in encapsulating complex syntactic structures effectively. Addressing this gap, we introduce the Linguistic Structure through Graphical Interpretation with BERT (LSGIB) approach in Machine Translation (MT) frameworks. Combining the strengths of Graph Attention Network (GAT) and BERT, LSGIB intricately captures syntactic dependencies as explicit knowledge from the source language. This enhances the source language representation and aids in more accurate target language generation. Our empirical analysis leverages gold-standard syntax-annotated sentences and employs a Quality Estimation (QE) model. This approach enables us to assess translation improvements in terms of syntactic accuracy, extending beyond traditional BLEU score metrics. The LSGIB model demonstrates superior translation quality across diverse MT tasks, maintaining robust BLEU scores. Our study delves into the optimal sentence lengths benefiting from LSGIB and identifies which syntactic dependencies are more precisely captured. We observe that GAT's ability to learn specific dependency relations directly influences the translation quality of sentences with those relations. Additionally, we discover that incorporating syntactic structure into BERT's intermediate and lower layers offers a novel approach to modeling linguistic structure in source sentences.

Keywords

Machine Translation; Linguistic Interpretation; Attention Models

Subject

Computer Science and Mathematics, Artificial Intelligence and Machine Learning

Comments (0)

We encourage comments and feedback from a broad range of readers. See criteria for comments and our Diversity statement.

Leave a public comment
Send a private comment to the author(s)
* All users must log in before leaving a comment
Views 0
Downloads 0
Comments 0
Metrics 0


×
Alerts
Notify me about updates to this article or when a peer-reviewed version is published.
We use cookies on our website to ensure you get the best experience.
Read more about our cookies here.