Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -70,7 +70,6 @@ def translate_text(text, src_lang, tgt_lang, models):
|
|
70 |
if src_lang not in lang_map or tgt_lang not in lang_map:
|
71 |
return "Error: Unsupported language combination"
|
72 |
|
73 |
-
src_lang_code = lang_map[src_lang]
|
74 |
tgt_lang_code = lang_map[tgt_lang]
|
75 |
|
76 |
tokenizer, model = models["nllb"]
|
@@ -84,7 +83,8 @@ def translate_text(text, src_lang, tgt_lang, models):
|
|
84 |
|
85 |
for sentence in sentences:
|
86 |
if sentence.strip():
|
87 |
-
|
|
|
88 |
translated = model.generate(**inputs, forced_bos_token_id=tokenizer.get_lang_id(tgt_lang_code), max_length=512)
|
89 |
translated_sentence = tokenizer.decode(translated[0], skip_special_tokens=True)
|
90 |
translated_text += translated_sentence + "\n"
|
|
|
70 |
if src_lang not in lang_map or tgt_lang not in lang_map:
|
71 |
return "Error: Unsupported language combination"
|
72 |
|
|
|
73 |
tgt_lang_code = lang_map[tgt_lang]
|
74 |
|
75 |
tokenizer, model = models["nllb"]
|
|
|
83 |
|
84 |
for sentence in sentences:
|
85 |
if sentence.strip():
|
86 |
+
# Removed src_lang from tokenizer call
|
87 |
+
inputs = tokenizer(sentence, return_tensors="pt", padding=True, truncation=True, max_length=512)
|
88 |
translated = model.generate(**inputs, forced_bos_token_id=tokenizer.get_lang_id(tgt_lang_code), max_length=512)
|
89 |
translated_sentence = tokenizer.decode(translated[0], skip_special_tokens=True)
|
90 |
translated_text += translated_sentence + "\n"
|