You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

Model Card for Model ID

This is a fine-tuned model for question or statements rewrite task focused on Traditional Chinese specifically. In this version , we have adjusted the way the model calculates loss.
(The original training process (i.e. SFTTrainer class from trl) calculates CE on whole prompt template.)

In order to prevent the model from copying the original sentence, the total loss we use will be counted as three parts :

  1. Context Loss (from the beginning to <rephrased>)
  2. Answer Loss (from <rephrased> to </rephrased>)
  3. Variety Loss (VTLoss) , it calculates the IOU of orignal tokenized sentence and rewritten tokenized sentence , trying to encourage the model to generate as diverse text as possible.

Noted that the answer loss will take a larger weight than context loss since the answer is more important part that we shall take care of.

Model Details

the prompt template should be used as follow:

<task>  
你是一名熱於助人的AI小幫手,請將敘述語句或者問句變得更加通順與簡潔。  
</task>  

原始句子:  
<origin>  
{before}  
</origin>  

修改後:  
<rephrased>  
{after}  
</rephrased>  

Noted that {before} {after} are the original question/statement and rewritten question/statement respcetively.
Moreover , this model is not the best rewrite tool compared with many open source LLMs , it is a trial version.
But we'll still make improvements on it.

Model Description

This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.

  • Developed by: [--]
  • Funded by [optional]: [--]
  • Shared by [optional]: [--]
  • Model type: [--]
  • Language(s) (NLP): [Traditional Chinese]
  • License: [--]
  • Finetuned from model [optional]: [Taiwan LLM base v2.0]

Training Details

Training Data

Generate from GPT4o and artificial human feedback.
Custom Traditional Chinese BenchMark Dataset , with rewritten answers came from Gemini.
Also , the evaluation task is assigned to GPTo with custom rubrics.

[More Information Needed]

Training Procedure

Training Hyperparameters

  • Training regime: [QLoRA]

More Information [optional]

[--]

Model Card Authors [optional]

[--]

Model Card Contact

[--]

Downloads last month
1
Safetensors
Model size
6.74B params
Tensor type
FP16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.