Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -83,9 +83,13 @@ def translate_text(text, src_lang, tgt_lang, models):
|
|
83 |
|
84 |
for sentence in sentences:
|
85 |
if sentence.strip():
|
86 |
-
# Removed src_lang from tokenizer call
|
87 |
inputs = tokenizer(sentence, return_tensors="pt", padding=True, truncation=True, max_length=512)
|
88 |
-
|
|
|
|
|
|
|
|
|
|
|
89 |
translated_sentence = tokenizer.decode(translated[0], skip_special_tokens=True)
|
90 |
translated_text += translated_sentence + "\n"
|
91 |
|
|
|
83 |
|
84 |
for sentence in sentences:
|
85 |
if sentence.strip():
|
|
|
86 |
inputs = tokenizer(sentence, return_tensors="pt", padding=True, truncation=True, max_length=512)
|
87 |
+
# Use lang_code_to_id instead of get_lang_id
|
88 |
+
translated = model.generate(
|
89 |
+
**inputs,
|
90 |
+
forced_bos_token_id=tokenizer.lang_code_to_id[tgt_lang_code],
|
91 |
+
max_length=512
|
92 |
+
)
|
93 |
translated_sentence = tokenizer.decode(translated[0], skip_special_tokens=True)
|
94 |
translated_text += translated_sentence + "\n"
|
95 |
|