Preprint
Article

Cross-Lingual Semantic Alignment in Large Language Models via Context-Aware Training

This version is not peer-reviewed.

Submitted:

12 March 2025

Posted:

13 March 2025

You are already at the latest version

Abstract
This paper introduces Context-Aware Cross-Modal Alignment Training (CACMAT), a novel multi-stage training paradigm to enhance translation capabilities of Large Language Models (LLMs) and Large Vision-Language Models (LVLMs). Current LLM translation models often struggle with contextual nuances and cross-lingual semantic alignment. CACMAT addresses this by incorporating three stages: secondary pre-training on target language monolingual data, continual pre-training with a contextual contrastive loss using Interlinear Text Format (ITF) data to improve cross-lingual alignment, and supervised fine-tuning on parallel translation datasets. Experiments on FLORES-200 and WMT datasets demonstrate that CACMAT significantly outperforms baseline models and achieves competitive results against state-of-the-art systems, as validated by both BLEU scores and human evaluations. Ablation studies confirm the crucial role of the contextual contrastive alignment stage. The results highlight CACMAT as an effective approach for improving translation quality by explicitly enhancing cross-lingual and cross-modal semantic alignment in LLMs and LVLMs.
Keywords: 
;  
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.

Downloads

36

Views

20

Comments

0

Subscription

Notify me about updates to this article or when a peer-reviewed version is published.

Email

Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

© 2025 MDPI (Basel, Switzerland) unless otherwise stated