murawaki's picture
Update README.md
46ef41f
|
raw
history blame
3.17 kB
metadata
language: ja
license: cc-by-sa-4.0
library_name: transformers
tags:
  - gpt2
datasets:
  - wikipedia
  - cc100
  - oscar
widget:
  - text: 昨日私は京都で

Model Card for Japanese character-level GPT-2 Small

Model description

This is a Japanese character-level GPT-2 Small (90M parameters) language model pre-trained on Japanese Wikipedia, the Japanese portion of CC-100, and the Japanese portion of OSCAR.

How to use

You can use this model directly with a pipeline for text generation.

>>> from transformers import pipeline, set_seed
>>> generator = pipeline('text-generation', model='ku-nlp/gpt2-small-japanese-char')
>>> set_seed(5)
>>> generator("昨日私は京都で", max_length=30, do_sample=True, num_return_sequences=5)

[{'generated_text': '昨日私は京都であの日に、あんなに頑張ったのに…と思った。私は'},
 {'generated_text': '昨日私は京都で開かれた大阪市内で会場見学をしました。そしてそ'},
 {'generated_text': '昨日私は京都で行われました。その時はまだ若手が多数入学して何'},
 {'generated_text': '昨日私は京都では雪が解けるまで寝た様子があります・・・(;´'},
 {'generated_text': '昨日私は京都でこみ上げてきたものを写真撮るため、駅近くのセン'}]

You can also use this model to get the features of a given text.

Vocabulary

A character-level vocabulary of size 6K is used. To be precise, rare characters may be split into bytes because byte-level byte-pair encoding (BPE) is used. The BPE tokenizer was trained on a small subset of the training data. Since the data were converted into a one-character-per-line format, merge operations never go beyond character boundaries.

Note that the tokenizer maps U+0020 to [UNK] because preprocessing eliminated whitespace characters (U+0020) from training data. Use U+3000 (Ideographic Space) instead.

Training data

We used the following corpora for pre-training:

  • Japanese Wikipedia (as of 20221020, 3.2GB, 27M sentences, 1.3M documents)
  • Japanese portion of CC-100 (85GB, 619M sentences, 66M documents)
  • Japanese portion of OSCAR (54GB, 326M sentences, 25M documents)

Note that we filtered out documents annotated with "header", "footer", or "noisy" tags in OSCAR. Also note that Japanese Wikipedia was duplicated 10 times to make the total size of the corpus comparable to that of CC-100 and OSCAR. As a result, the total size of the training data is 171GB.

Training procedure

The training took about 3 months (with two interruptions) with a single NVIDIA A100 80GB GPU.

The following hyperparameters were used during pre-training:

  • learning_rate: 2e-4
  • per_device_train_batch_size: 36
  • gradient_accumulation_steps: 32
  • optimizer: AdamW with betas=(0.9, 0.999) and epsilon=1e-06
  • weight_decay: 0.01
  • lr_scheduler_type: linear
  • max_grad_norm: 1.0
  • max_steps: 500,000 (but terminated at 142,000 steps ~= 3.0 epochs)
  • warmup_steps: 10,000

The eval loss was 1.597 while the eval accuracy was 0.6359. The evaluation set consists of 5,000 randomly sampled documents from each of the training corpora.