Update README.md
Browse files
README.md
CHANGED
@@ -82,7 +82,8 @@ model.eval()
|
|
82 |
gen_ids = model.forward_test(input)
|
83 |
gen_text = model.neck.tokenizer.batch_decode(gen_ids.clip(0))
|
84 |
# 4. Post-processing
|
85 |
-
#
|
|
|
86 |
gen_text = gen_text.split('<s>')[-1].split('\n')[0].strip()
|
87 |
gen_text = gen_text.replace(' in Chinese','')
|
88 |
gen_text = gen_text.replace(' Chinese','')
|
|
|
82 |
gen_ids = model.forward_test(input)
|
83 |
gen_text = model.neck.tokenizer.batch_decode(gen_ids.clip(0))
|
84 |
# 4. Post-processing
|
85 |
+
# Given that the training data may contain biases, the generated texts might need some straightforward post-processing to ensure accuracy.
|
86 |
+
# In future versions, we will enhance the quality of the data.
|
87 |
gen_text = gen_text.split('<s>')[-1].split('\n')[0].strip()
|
88 |
gen_text = gen_text.replace(' in Chinese','')
|
89 |
gen_text = gen_text.replace(' Chinese','')
|