|
|
|
|
|
|
|
|
|
""" |
|
## TODO: |
|
- i18 国际化 https://blog.csdn.net/qq_26212731/article/details/78457198 request.header中也有language |
|
- iter_vocab 的 warmup |
|
- 开关 |
|
- add_special_token 开关 |
|
- theme 开关 light/dark |
|
- token_id/tokens/bytes 开关 |
|
- 中文字词统计,是否要包括 _ G 等字符 |
|
- 评测 |
|
- OOV评测 |
|
- 通过 javascript 添加 hover_text |
|
- 英文 utf-8编码 |
|
- 词典支持下载,借用image下载的标签, |
|
- baichuan的单字数量怎么两万多个? |
|
- qwen: ValueError: Unclosed image token |
|
- 路径修改为全path meta-llama/Llama-2-13b-hf |
|
|
|
plots |
|
|
|
table |
|
|
|
## related demo |
|
- [](http://text-processing.com/demo/tokenize/) |
|
- [gpt-tokenizer](https://gpt-tokenizer.dev/) |
|
- [llama-tokenizer-js](https://belladoreai.github.io/llama-tokenizer-js/example-demo/build/) |
|
- [](https://huggingface.co/spaces/Xenova/the-tokenizer-playground) |
|
|
|
## 可视化 |
|
|
|
[ The, 2, QUICK, Brown, Foxes, jumped, over, the, lazy, dog's, bone ] |
|
""" |
|
|
|
import gradio as gr |
|
from vocab import all_tokenizers |
|
from util import * |
|
from examples import example_fn, example_types |
|
|
|
get_window_url_params = """ |
|
function(url_params) { |
|
const params = new URLSearchParams(window.location.search); |
|
url_params = JSON.stringify(Object.fromEntries(params)); |
|
return url_params; |
|
} |
|
""" |
|
|
|
with gr.Blocks() as demo: |
|
|
|
|
|
|
|
|
|
|
|
with gr.Row(): |
|
gr.Markdown("## Input Text") |
|
dropdown_examples = gr.Dropdown( |
|
example_types, |
|
value="Examples", |
|
type="index", |
|
show_label=False, |
|
container=False, |
|
scale=0, |
|
elem_classes="example-style" |
|
) |
|
user_input = gr.Textbox( |
|
|
|
label="Input Text", |
|
lines=5, |
|
show_label=False, |
|
) |
|
gr.Markdown("## Tokenization") |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
with gr.Row(): |
|
with gr.Column(scale=6): |
|
with gr.Group(): |
|
tokenizer_name_1 = gr.Dropdown( |
|
all_tokenizers, |
|
label="Tokenizer 1", |
|
) |
|
with gr.Group(): |
|
with gr.Row(): |
|
stats_vocab_size_1 = gr.TextArea( |
|
label="Vocab Size", |
|
lines=1, |
|
elem_classes="statistics" |
|
) |
|
stats_zh_token_size_1 = gr.TextArea( |
|
label="ZH char/word", |
|
lines=1, |
|
elem_classes="statistics", |
|
) |
|
|
|
|
|
|
|
|
|
|
|
stats_overlap_token_size_1 = gr.TextArea( |
|
|
|
label="Overlap Tokens", |
|
lines=1, |
|
elem_classes="statistics" |
|
) |
|
|
|
|
|
|
|
|
|
|
|
|
|
gr.Image("images/VS.svg", scale=1, show_label=False, |
|
show_download_button=False, container=False, |
|
show_share_button=False) |
|
with gr.Column(scale=6): |
|
with gr.Group(): |
|
tokenizer_name_2 = gr.Dropdown( |
|
all_tokenizers, |
|
label="Tokenizer 2", |
|
) |
|
with gr.Group(): |
|
with gr.Row(): |
|
stats_vocab_size_2 = gr.TextArea( |
|
label="VocabSize", |
|
lines=1, |
|
elem_classes="statistics" |
|
) |
|
stats_zh_token_size_2 = gr.TextArea( |
|
label="ZH char/word", |
|
lines=1, |
|
elem_classes="statistics", |
|
) |
|
|
|
|
|
|
|
|
|
|
|
stats_filtered_token_2 = gr.TextArea( |
|
label="filtered tokens", |
|
lines=1, |
|
elem_classes="statistics", |
|
visible=False |
|
) |
|
stats_overlap_token_size_2 = gr.TextArea( |
|
label="Overlap Tokens", |
|
lines=1, |
|
elem_classes="statistics" |
|
) |
|
|
|
|
|
with gr.Row(): |
|
|
|
with gr.Column(): |
|
output_text_1 = gr.Highlightedtext( |
|
show_legend=True, |
|
elem_classes="space-show" |
|
) |
|
with gr.Column(): |
|
output_text_2 = gr.Highlightedtext( |
|
show_legend=True, |
|
elem_classes="space-show" |
|
) |
|
|
|
with gr.Row(): |
|
output_table_1 = gr.Dataframe() |
|
output_table_2 = gr.Dataframe() |
|
|
|
|
|
|
|
|
|
|
|
tokenizer_name_1.change(tokenize, [user_input, tokenizer_name_1], |
|
[output_text_1, output_table_1]) |
|
tokenizer_name_1.change(basic_count, [tokenizer_name_1], [stats_vocab_size_1, stats_zh_token_size_1]) |
|
tokenizer_name_1.change(get_overlap_token_size, [tokenizer_name_1, tokenizer_name_2], |
|
[stats_overlap_token_size_1, stats_overlap_token_size_2]) |
|
|
|
|
|
|
|
|
|
user_input.change(tokenize_pair, |
|
[user_input, tokenizer_name_1, tokenizer_name_2], |
|
[output_text_1, output_table_1, output_text_2, output_table_2]) |
|
|
|
tokenizer_name_2.change(tokenize, [user_input, tokenizer_name_2], |
|
[output_text_2, output_table_2]) |
|
tokenizer_name_2.change(basic_count, [tokenizer_name_2], [stats_vocab_size_2, stats_zh_token_size_2]) |
|
tokenizer_name_2.change(get_overlap_token_size, [tokenizer_name_1, tokenizer_name_2], |
|
[stats_overlap_token_size_1, stats_overlap_token_size_2]) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
dropdown_examples.change( |
|
example_fn, |
|
dropdown_examples, |
|
[user_input, tokenizer_name_1, tokenizer_name_2] |
|
) |
|
|
|
demo.load(js=open("js/onload.js", "r", encoding="utf-8").read()) |
|
demo.load( |
|
fn=on_load, |
|
inputs=[user_input], |
|
outputs=[user_input, tokenizer_name_1, tokenizer_name_2], |
|
js=get_window_url_params |
|
) |
|
|
|
if __name__ == "__main__": |
|
|
|
demo.launch() |
|
|
|
|