diff --git a/tower_eval/models/seq2seq/generator.py b/tower_eval/models/seq2seq/generator.py index 80dca4a..bf8b894 100644 --- a/tower_eval/models/seq2seq/generator.py +++ b/tower_eval/models/seq2seq/generator.py @@ -56,7 +56,7 @@ def _batch_generate(self, input_lines: List[str]) -> List[str]: """ # NLLB requires that source language be passed to the tokenizer and the target language be passed to the model if self.model_family == "nllb": - self.tokenizer.src_lang = self.source_language + self.tokenizer.src_lang = NLLB_LANGUAGE_CODES[self.source_language] inputs = self.tokenizer(input_lines, return_tensors="pt", padding=True).to( "cuda" )