Qwen2 Fine-Tuned on Parenthetical Terminology Translation (PTT) Dataset

Model Overview

This is a qwen2-1.5B model fine-tuned on the Parenthetical Terminology Translation (PTT) dataset. The PTT dataset focuses on translating technical terms accurately by placing the original English term in parentheses alongside its Korean translation, enhancing clarity and precision in specialized fields. This fine-tuned model is optimized for handling technical terminology in the Artificial Intelligence (AI) domain.

Example Usage

Hereโ€™s how to use this fine-tuned model with the Hugging Face transformers library:

import transformers
from transformers import AutoTokenizer, AutoModelForCausalLM

# Load Model and Tokenizer
model_name = "PrompTartLAB/llama3_8B_PTT_en_ko"
model = AutoModelForCausalLM.from_pretrained(
    model_name,
    torch_dtype="auto",
    device_map="auto",
)
tokenizer = AutoTokenizer.from_pretrained(model_name)

# Example sentence
text = "The model was fine-tuned using knowledge distillation techniques. The training dataset was created using a collaborative multi-agent framework powered by large language models."
prompt = f"Translate input sentence to Korean \n### Input: {text} \n### Translated:"

# Tokenize and generate translation
input_ids = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(**input_ids, max_new_tokens=1024)
out_message = tokenizer.decode(outputs[0][len(input_ids["input_ids"][0]):], skip_special_tokens=True)

# " ์ด ๋ชจ๋ธ์€ ์ง€์‹ ๋ถ„์‚ฐ ๊ธฐ๋ฒ•(knowledge distillation techniques)์„ ์‚ฌ์šฉํ•˜์—ฌ ๋ฏธ์„ธ ์กฐ์ •๋˜์—ˆ์Šต๋‹ˆ๋‹ค. ํ›ˆ๋ จ ๋ฐ์ดํ„ฐ์…‹์€ ๋Œ€ํ˜• ์–ธ์–ด ๋ชจ๋ธ(large language models)์„ ๊ธฐ๋ฐ˜์œผ๋กœ ํ•œ ํ˜‘๋ ฅ ๋‹ค์ค‘ ์—์ด์ „ํŠธ ํ”„๋ ˆ์ž„์›Œํฌ(collaborative multi-agent framework)๋ฅผ ํ†ตํ•ด ์ƒ์„ฑ๋˜์—ˆ์Šต๋‹ˆ๋‹ค."

Limitations

  • Out-of-Domain Accuracy: While the model generalizes to some extent, accuracy may vary in domains that were not part of the training set.
  • Incomplete Parenthetical Annotation: Not all technical terms are consistently displayed in parentheses; in some cases, terms may be omitted or not annotated as expected.

Citation

If you use this model in your research, please cite the original dataset and paper:

@inproceedings{jiyoon-etal-2024-efficient,
    title = "Efficient Technical Term Translation: A Knowledge Distillation Approach for Parenthetical Terminology Translation",
    author = "Jiyoon, Myung  and
      Park, Jihyeon  and
      Son, Jungki  and
      Lee, Kyungro  and
      Han, Joohyung",
    editor = "Haddow, Barry  and
      Kocmi, Tom  and
      Koehn, Philipp  and
      Monz, Christof",
    booktitle = "Proceedings of the Ninth Conference on Machine Translation",
    month = nov,
    year = "2024",
    address = "Miami, Florida, USA",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2024.wmt-1.129",
    doi = "10.18653/v1/2024.wmt-1.129",
    pages = "1410--1427",
    abstract = "This paper addresses the challenge of accurately translating technical terms, which are crucial for clear communication in specialized fields. We introduce the Parenthetical Terminology Translation (PTT) task, designed to mitigate potential inaccuracies by displaying the original term in parentheses alongside its translation. To implement this approach, we generated a representative PTT dataset using a collaborative approach with large language models and applied knowledge distillation to fine-tune traditional Neural Machine Translation (NMT) models and small-sized Large Language Models (sLMs). Additionally, we developed a novel evaluation metric to assess both overall translation accuracy and the correct parenthetical presentation of terms. Our findings indicate that sLMs did not consistently outperform NMT models, with fine-tuning proving more effective than few-shot prompting, particularly in models with continued pre-training in the target language. These insights contribute to the advancement of more reliable terminology translation methodologies.",
}

Contact

For questions or feedback, please contact [email protected].

Downloads last month
22
Safetensors
Model size
1.54B params
Tensor type
BF16
ยท
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for PrompTartLAB/qwen2_1.5B_PTT_en_ko

Base model

Qwen/Qwen2-1.5B
Finetuned
(30)
this model
Quantizations
1 model

Dataset used to train PrompTartLAB/qwen2_1.5B_PTT_en_ko