Enhancing LLMs’ Translation Capabilities with ICL
This work builds on previous research that has focused on enhancing LLMs’ translation capabilities without additional training. A primary strategy involves leveraging LLMs’ prompt learning ability, which is an ability to learn from demonstrations or descriptions (Brown et al. 2020; Wei et al. 2022). Studies have explored selecting appropriate exemplars for few-shot learning and demonstrating linguistic knowledge(Agrawal et al. 2022; Vilar et al. 2023; Zhang et al. 2024). Besides providing a demonstration or a description, choosing the right temperature or prompting strategy has also been examined (Peng et al. 2023). Similar to previous researches, our method aim to improve LLMs’ MT capabilities without further fine-tuning, but rather focuses on eliciting models’ own capabilities rather than augmenting it with external knowledge sources.
Self-generated Prompts
Generating appropriate exemplars for in-context learning manually can be resource-intensive in many cases. To address this, previous works have explored enabling models to generate their own few-shot examples for tasks such as classification(Lyu et al. 2023; Kim et al. 2022) or other reasoning tasks(Zhang et al. 2022; Li et al. 2024). Our work is related to these efforts, as it also involves generating its own few-shot examples. The difference between our study and previous ones is that we take a different approach, which mitigates the potential noise of synthesized data by gradually expanding the example set with similar yet distinct examples.
Ablation
To find the optimal setup for the framework, we’ve conducted an ablation study on en-ko MT task with four different types of strategies we could choose: start sentence selection method, number of start sentences, method for aggregating gradual MT results, and method for choosing source sentences to translate with this framework. In this section, we briefly explain the effects of each strategies by averaging the QE scores. Full results with every different combination of strategies can be found in Appendix.
start sentence selection method
Start sentence selection method is a me
start sents number
aggregation strategy
end sent filtering
Characteristics of Interpolation
Basic Statistics
We also conducted a further analysis to see various characteristics of sentence interpolation. Below are the list of variables we have checked.
- Interval between sentences
- PCA Alignment
- Numb
- Maximum Interval between sentences
- Interval between last interpolated sentence and
Types of Interpolation