Upload folder using huggingface_hub
Browse files- .gitattributes +1 -0
- .yml +9 -0
- SJMT_model.pth +3 -0
- app.py +25 -0
- code_generator/merges.txt +0 -0
- code_generator/special_tokens_map.json +5 -0
- code_generator/tokenizer.json +0 -0
- code_generator/tokenizer_config.json +20 -0
- code_generator/vocab.json +0 -0
- config.json +48 -0
- config.yml +73 -0
- nlp_encoder/special_tokens_map.json +7 -0
- nlp_encoder/tokenizer.json +0 -0
- nlp_encoder/tokenizer_config.json +56 -0
- nlp_encoder/vocab.txt +0 -0
- sample-15s.wav +3 -0
- speech_encoder/special_tokens_map.json +6 -0
- speech_encoder/tokenizer_config.json +50 -0
- speech_encoder/vocab.json +34 -0
- text_generator/merges.txt +0 -0
- text_generator/special_tokens_map.json +15 -0
- text_generator/tokenizer.json +0 -0
- text_generator/tokenizer_config.json +58 -0
- text_generator/vocab.json +0 -0
- vision_encoder/merges.txt +0 -0
- vision_encoder/special_tokens_map.json +30 -0
- vision_encoder/tokenizer.json +0 -0
- vision_encoder/tokenizer_config.json +31 -0
- vision_encoder/vocab.json +0 -0
.gitattributes
CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
sample-15s.wav filter=lfs diff=lfs merge=lfs -text
|
.yml
ADDED
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
speech_recognition:
|
2 |
+
waveform_path: "C:/Users/baby7/Desktop/multi_modal_model/sample-15s.wav"
|
3 |
+
task: "speech_recognition"
|
4 |
+
output_audio_key: "Transcription"
|
5 |
+
|
6 |
+
tests:
|
7 |
+
- name: speech_recognition_test
|
8 |
+
waveform_path: "C:/Users/baby7/Desktop/multi_modal_model/sample-15s.wav"
|
9 |
+
expected_output: "Expected transcription"
|
SJMT_model.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b83992690213c479a00be25af7fa6bfea7526861094cff5b6e97ee44d89f1cbb
|
3 |
+
size 2493181678
|
app.py
ADDED
@@ -0,0 +1,25 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
def mock_text_generation(input_text):
|
2 |
+
# 模拟文本生成逻辑
|
3 |
+
if input_text == "Tell me a joke.":
|
4 |
+
return "Why don't scientists trust atoms? Because they make up everything!"
|
5 |
+
return "I can come up with many ideas, but that request has stumped me!"
|
6 |
+
|
7 |
+
def mock_code_generation(input_code):
|
8 |
+
# 模拟代码生成逻辑
|
9 |
+
if input_code == "def greet(name):":
|
10 |
+
return "def greet(name):\n return f'Hello, {name}!'"
|
11 |
+
return "Hmm, I'm not sure how to complete that one."
|
12 |
+
|
13 |
+
# 测试文本生成功能
|
14 |
+
input_text = "Tell me a joke."
|
15 |
+
expected_output_text = "Why don't scientists trust atoms? Because they make up everything!"
|
16 |
+
generated_text = mock_text_generation(input_text)
|
17 |
+
assert generated_text == expected_output_text, f"Text generation test failed: expected {expected_output_text}, got {generated_text}"
|
18 |
+
print("Text generation test passed")
|
19 |
+
|
20 |
+
# 测试代码生成功能
|
21 |
+
input_code = "def greet(name):"
|
22 |
+
expected_output_code = "def greet(name):\n return f'Hello, {name}!'"
|
23 |
+
generated_code = mock_code_generation(input_code)
|
24 |
+
assert generated_code == expected_output_code, f"Code generation test failed: expected {expected_output_code}, got {generated_code}"
|
25 |
+
print("Code generation test passed")
|
code_generator/merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
code_generator/special_tokens_map.json
ADDED
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": "<|endoftext|>",
|
3 |
+
"eos_token": "<|endoftext|>",
|
4 |
+
"unk_token": "<|endoftext|>"
|
5 |
+
}
|
code_generator/tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
code_generator/tokenizer_config.json
ADDED
@@ -0,0 +1,20 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_prefix_space": false,
|
3 |
+
"added_tokens_decoder": {
|
4 |
+
"50256": {
|
5 |
+
"content": "<|endoftext|>",
|
6 |
+
"lstrip": false,
|
7 |
+
"normalized": true,
|
8 |
+
"rstrip": false,
|
9 |
+
"single_word": false,
|
10 |
+
"special": true
|
11 |
+
}
|
12 |
+
},
|
13 |
+
"bos_token": "<|endoftext|>",
|
14 |
+
"clean_up_tokenization_spaces": false,
|
15 |
+
"eos_token": "<|endoftext|>",
|
16 |
+
"extra_special_tokens": {},
|
17 |
+
"model_max_length": 1024,
|
18 |
+
"tokenizer_class": "GPT2Tokenizer",
|
19 |
+
"unk_token": "<|endoftext|>"
|
20 |
+
}
|
code_generator/vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
config.json
ADDED
@@ -0,0 +1,48 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"model_name": "Evolutionary Multi-Modal Model",
|
3 |
+
"model_type": "transformer",
|
4 |
+
"license": "mit",
|
5 |
+
"language": "en",
|
6 |
+
"datasets": [
|
7 |
+
"Custom"
|
8 |
+
],
|
9 |
+
"tags": [
|
10 |
+
"text-generation",
|
11 |
+
"code-generation",
|
12 |
+
"speech-recognition",
|
13 |
+
"multi-modal",
|
14 |
+
"evolutionary"
|
15 |
+
],
|
16 |
+
"base_model": "facebook/bart-base",
|
17 |
+
"finetuned_from": [
|
18 |
+
"gpt2",
|
19 |
+
"bert-base-uncased",
|
20 |
+
"facebook/wav2vec2-base-960h",
|
21 |
+
"openai/clip-vit-base-patch32"
|
22 |
+
],
|
23 |
+
"dataset": "Custom Multi-Modal Dataset",
|
24 |
+
"metrics": [
|
25 |
+
"perplexity",
|
26 |
+
"bleu",
|
27 |
+
"wer",
|
28 |
+
"cer"
|
29 |
+
],
|
30 |
+
"library_name": "transformers",
|
31 |
+
"pipeline_tag": "text-generation",
|
32 |
+
"inference": {
|
33 |
+
"parameters": {
|
34 |
+
"max_length": 50,
|
35 |
+
"top_k": 50,
|
36 |
+
"top_p": 0.95,
|
37 |
+
"temperature": 1.2,
|
38 |
+
"do_sample": true
|
39 |
+
}
|
40 |
+
},
|
41 |
+
"extra_info": {
|
42 |
+
"author": "zero",
|
43 |
+
"version": "1.0",
|
44 |
+
"description":"1.0",
|
45 |
+
"citation": "@1.0"
|
46 |
+
}
|
47 |
+
}
|
48 |
+
|
config.yml
ADDED
@@ -0,0 +1,73 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
model_name: Evolutionary Multi-Modal Model
|
2 |
+
model_type: transformer
|
3 |
+
license: mit
|
4 |
+
language: en zh
|
5 |
+
datasets:
|
6 |
+
- "Custom"
|
7 |
+
tags:
|
8 |
+
- text-generation
|
9 |
+
- code-generation
|
10 |
+
- speech-recognition
|
11 |
+
- multi-modal
|
12 |
+
- evolutionary
|
13 |
+
base_model: facebook/bart-base
|
14 |
+
finetuned_from: gpt2, bert-base-uncased, facebook/wav2vec2-base-960h, openai/clip-vit-base-patch32
|
15 |
+
dataset: Custom Multi-Modal Dataset
|
16 |
+
|
17 |
+
metrics:
|
18 |
+
- perplexity
|
19 |
+
- bleu
|
20 |
+
- wer
|
21 |
+
- cer
|
22 |
+
|
23 |
+
library_name: transformers
|
24 |
+
pipeline_tag: text-generation
|
25 |
+
inference:
|
26 |
+
parameters:
|
27 |
+
max_length: 50
|
28 |
+
top_k: 50
|
29 |
+
top_p: 0.95
|
30 |
+
temperature: 1.2
|
31 |
+
do_sample: true
|
32 |
+
|
33 |
+
speech_recognition:
|
34 |
+
waveform_path: "C:/Users/baby7/Desktop/权重参数/sample-15s.wav"
|
35 |
+
task: "speech_recognition"
|
36 |
+
output_audio_key: "Transcription"
|
37 |
+
|
38 |
+
text_generation:
|
39 |
+
input_text: "What is the future of AI?"
|
40 |
+
task: "text_generation"
|
41 |
+
output_text_key: "Generated Text"
|
42 |
+
|
43 |
+
code_generation:
|
44 |
+
input_code: "def add(a, b): return"
|
45 |
+
task: "code_generation"
|
46 |
+
output_code_key: "Generated Code"
|
47 |
+
|
48 |
+
tests:
|
49 |
+
- name: speech_recognition_test
|
50 |
+
waveform_path: "C:/Users/baby7/Desktop/权重参数/sample-15s.wav"
|
51 |
+
expected_output: "Expected transcription"
|
52 |
+
- name: text_generation_test
|
53 |
+
input_text: "What is the future of AI?"
|
54 |
+
expected_output: "Predicted text about AI"
|
55 |
+
- name: code_generation_test
|
56 |
+
input_code: "def add(a, b): return"
|
57 |
+
expected_output: "def add(a, b): return a + b"
|
58 |
+
|
59 |
+
extra_info:
|
60 |
+
author: zero
|
61 |
+
version: 1.0
|
62 |
+
description: |
|
63 |
+
This Evolutionary Multi-Modal Model is designed for tasks like text generation, code generation,
|
64 |
+
speech recognition, and vision understanding. It leverages the capabilities of multiple pre-trained
|
65 |
+
models and applies evolutionary techniques to optimize performance across these tasks.
|
66 |
+
citation:
|
67 |
+
- |
|
68 |
+
@article{your_reference_2025,
|
69 |
+
title={Evolutionary Multi-Modal Model for Enhanced Performance},
|
70 |
+
author={Your Name},
|
71 |
+
journal={Journal of AI Research},
|
72 |
+
year={2025}
|
73 |
+
}
|
nlp_encoder/special_tokens_map.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"cls_token": "[CLS]",
|
3 |
+
"mask_token": "[MASK]",
|
4 |
+
"pad_token": "[PAD]",
|
5 |
+
"sep_token": "[SEP]",
|
6 |
+
"unk_token": "[UNK]"
|
7 |
+
}
|
nlp_encoder/tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
nlp_encoder/tokenizer_config.json
ADDED
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"added_tokens_decoder": {
|
3 |
+
"0": {
|
4 |
+
"content": "[PAD]",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false,
|
9 |
+
"special": true
|
10 |
+
},
|
11 |
+
"100": {
|
12 |
+
"content": "[UNK]",
|
13 |
+
"lstrip": false,
|
14 |
+
"normalized": false,
|
15 |
+
"rstrip": false,
|
16 |
+
"single_word": false,
|
17 |
+
"special": true
|
18 |
+
},
|
19 |
+
"101": {
|
20 |
+
"content": "[CLS]",
|
21 |
+
"lstrip": false,
|
22 |
+
"normalized": false,
|
23 |
+
"rstrip": false,
|
24 |
+
"single_word": false,
|
25 |
+
"special": true
|
26 |
+
},
|
27 |
+
"102": {
|
28 |
+
"content": "[SEP]",
|
29 |
+
"lstrip": false,
|
30 |
+
"normalized": false,
|
31 |
+
"rstrip": false,
|
32 |
+
"single_word": false,
|
33 |
+
"special": true
|
34 |
+
},
|
35 |
+
"103": {
|
36 |
+
"content": "[MASK]",
|
37 |
+
"lstrip": false,
|
38 |
+
"normalized": false,
|
39 |
+
"rstrip": false,
|
40 |
+
"single_word": false,
|
41 |
+
"special": true
|
42 |
+
}
|
43 |
+
},
|
44 |
+
"clean_up_tokenization_spaces": false,
|
45 |
+
"cls_token": "[CLS]",
|
46 |
+
"do_lower_case": true,
|
47 |
+
"extra_special_tokens": {},
|
48 |
+
"mask_token": "[MASK]",
|
49 |
+
"model_max_length": 512,
|
50 |
+
"pad_token": "[PAD]",
|
51 |
+
"sep_token": "[SEP]",
|
52 |
+
"strip_accents": null,
|
53 |
+
"tokenize_chinese_chars": true,
|
54 |
+
"tokenizer_class": "BertTokenizer",
|
55 |
+
"unk_token": "[UNK]"
|
56 |
+
}
|
nlp_encoder/vocab.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
sample-15s.wav
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:33e2e7b2ffa021275a90a26704d923fe902d3600e4ffecf06253c57778a2a986
|
3 |
+
size 3382316
|
speech_encoder/special_tokens_map.json
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": "<s>",
|
3 |
+
"eos_token": "</s>",
|
4 |
+
"pad_token": "<pad>",
|
5 |
+
"unk_token": "<unk>"
|
6 |
+
}
|
speech_encoder/tokenizer_config.json
ADDED
@@ -0,0 +1,50 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"added_tokens_decoder": {
|
3 |
+
"0": {
|
4 |
+
"content": "<pad>",
|
5 |
+
"lstrip": true,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": true,
|
8 |
+
"single_word": false,
|
9 |
+
"special": false
|
10 |
+
},
|
11 |
+
"1": {
|
12 |
+
"content": "<s>",
|
13 |
+
"lstrip": true,
|
14 |
+
"normalized": false,
|
15 |
+
"rstrip": true,
|
16 |
+
"single_word": false,
|
17 |
+
"special": false
|
18 |
+
},
|
19 |
+
"2": {
|
20 |
+
"content": "</s>",
|
21 |
+
"lstrip": true,
|
22 |
+
"normalized": false,
|
23 |
+
"rstrip": true,
|
24 |
+
"single_word": false,
|
25 |
+
"special": false
|
26 |
+
},
|
27 |
+
"3": {
|
28 |
+
"content": "<unk>",
|
29 |
+
"lstrip": true,
|
30 |
+
"normalized": false,
|
31 |
+
"rstrip": true,
|
32 |
+
"single_word": false,
|
33 |
+
"special": false
|
34 |
+
}
|
35 |
+
},
|
36 |
+
"bos_token": "<s>",
|
37 |
+
"clean_up_tokenization_spaces": false,
|
38 |
+
"do_lower_case": false,
|
39 |
+
"do_normalize": true,
|
40 |
+
"eos_token": "</s>",
|
41 |
+
"extra_special_tokens": {},
|
42 |
+
"model_max_length": 1000000000000000019884624838656,
|
43 |
+
"pad_token": "<pad>",
|
44 |
+
"replace_word_delimiter_char": " ",
|
45 |
+
"return_attention_mask": false,
|
46 |
+
"target_lang": null,
|
47 |
+
"tokenizer_class": "Wav2Vec2CTCTokenizer",
|
48 |
+
"unk_token": "<unk>",
|
49 |
+
"word_delimiter_token": "|"
|
50 |
+
}
|
speech_encoder/vocab.json
ADDED
@@ -0,0 +1,34 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"'": 27,
|
3 |
+
"</s>": 2,
|
4 |
+
"<pad>": 0,
|
5 |
+
"<s>": 1,
|
6 |
+
"<unk>": 3,
|
7 |
+
"A": 7,
|
8 |
+
"B": 24,
|
9 |
+
"C": 19,
|
10 |
+
"D": 14,
|
11 |
+
"E": 5,
|
12 |
+
"F": 20,
|
13 |
+
"G": 21,
|
14 |
+
"H": 11,
|
15 |
+
"I": 10,
|
16 |
+
"J": 29,
|
17 |
+
"K": 26,
|
18 |
+
"L": 15,
|
19 |
+
"M": 17,
|
20 |
+
"N": 9,
|
21 |
+
"O": 8,
|
22 |
+
"P": 23,
|
23 |
+
"Q": 30,
|
24 |
+
"R": 13,
|
25 |
+
"S": 12,
|
26 |
+
"T": 6,
|
27 |
+
"U": 16,
|
28 |
+
"V": 25,
|
29 |
+
"W": 18,
|
30 |
+
"X": 28,
|
31 |
+
"Y": 22,
|
32 |
+
"Z": 31,
|
33 |
+
"|": 4
|
34 |
+
}
|
text_generator/merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
text_generator/special_tokens_map.json
ADDED
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": "<s>",
|
3 |
+
"cls_token": "<s>",
|
4 |
+
"eos_token": "</s>",
|
5 |
+
"mask_token": {
|
6 |
+
"content": "<mask>",
|
7 |
+
"lstrip": true,
|
8 |
+
"normalized": true,
|
9 |
+
"rstrip": false,
|
10 |
+
"single_word": false
|
11 |
+
},
|
12 |
+
"pad_token": "<pad>",
|
13 |
+
"sep_token": "</s>",
|
14 |
+
"unk_token": "<unk>"
|
15 |
+
}
|
text_generator/tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
text_generator/tokenizer_config.json
ADDED
@@ -0,0 +1,58 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_prefix_space": false,
|
3 |
+
"added_tokens_decoder": {
|
4 |
+
"0": {
|
5 |
+
"content": "<s>",
|
6 |
+
"lstrip": false,
|
7 |
+
"normalized": true,
|
8 |
+
"rstrip": false,
|
9 |
+
"single_word": false,
|
10 |
+
"special": true
|
11 |
+
},
|
12 |
+
"1": {
|
13 |
+
"content": "<pad>",
|
14 |
+
"lstrip": false,
|
15 |
+
"normalized": true,
|
16 |
+
"rstrip": false,
|
17 |
+
"single_word": false,
|
18 |
+
"special": true
|
19 |
+
},
|
20 |
+
"2": {
|
21 |
+
"content": "</s>",
|
22 |
+
"lstrip": false,
|
23 |
+
"normalized": true,
|
24 |
+
"rstrip": false,
|
25 |
+
"single_word": false,
|
26 |
+
"special": true
|
27 |
+
},
|
28 |
+
"3": {
|
29 |
+
"content": "<unk>",
|
30 |
+
"lstrip": false,
|
31 |
+
"normalized": true,
|
32 |
+
"rstrip": false,
|
33 |
+
"single_word": false,
|
34 |
+
"special": true
|
35 |
+
},
|
36 |
+
"50264": {
|
37 |
+
"content": "<mask>",
|
38 |
+
"lstrip": true,
|
39 |
+
"normalized": true,
|
40 |
+
"rstrip": false,
|
41 |
+
"single_word": false,
|
42 |
+
"special": true
|
43 |
+
}
|
44 |
+
},
|
45 |
+
"bos_token": "<s>",
|
46 |
+
"clean_up_tokenization_spaces": false,
|
47 |
+
"cls_token": "<s>",
|
48 |
+
"eos_token": "</s>",
|
49 |
+
"errors": "replace",
|
50 |
+
"extra_special_tokens": {},
|
51 |
+
"mask_token": "<mask>",
|
52 |
+
"model_max_length": 1000000000000000019884624838656,
|
53 |
+
"pad_token": "<pad>",
|
54 |
+
"sep_token": "</s>",
|
55 |
+
"tokenizer_class": "BartTokenizer",
|
56 |
+
"trim_offsets": true,
|
57 |
+
"unk_token": "<unk>"
|
58 |
+
}
|
text_generator/vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
vision_encoder/merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
vision_encoder/special_tokens_map.json
ADDED
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": {
|
3 |
+
"content": "<|startoftext|>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": true,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"eos_token": {
|
10 |
+
"content": "<|endoftext|>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"pad_token": {
|
17 |
+
"content": "<|endoftext|>",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": false,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
+
},
|
23 |
+
"unk_token": {
|
24 |
+
"content": "<|endoftext|>",
|
25 |
+
"lstrip": false,
|
26 |
+
"normalized": false,
|
27 |
+
"rstrip": false,
|
28 |
+
"single_word": false
|
29 |
+
}
|
30 |
+
}
|
vision_encoder/tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
vision_encoder/tokenizer_config.json
ADDED
@@ -0,0 +1,31 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_prefix_space": false,
|
3 |
+
"added_tokens_decoder": {
|
4 |
+
"49406": {
|
5 |
+
"content": "<|startoftext|>",
|
6 |
+
"lstrip": false,
|
7 |
+
"normalized": true,
|
8 |
+
"rstrip": false,
|
9 |
+
"single_word": false,
|
10 |
+
"special": true
|
11 |
+
},
|
12 |
+
"49407": {
|
13 |
+
"content": "<|endoftext|>",
|
14 |
+
"lstrip": false,
|
15 |
+
"normalized": false,
|
16 |
+
"rstrip": false,
|
17 |
+
"single_word": false,
|
18 |
+
"special": true
|
19 |
+
}
|
20 |
+
},
|
21 |
+
"bos_token": "<|startoftext|>",
|
22 |
+
"clean_up_tokenization_spaces": false,
|
23 |
+
"do_lower_case": true,
|
24 |
+
"eos_token": "<|endoftext|>",
|
25 |
+
"errors": "replace",
|
26 |
+
"extra_special_tokens": {},
|
27 |
+
"model_max_length": 77,
|
28 |
+
"pad_token": "<|endoftext|>",
|
29 |
+
"tokenizer_class": "CLIPTokenizer",
|
30 |
+
"unk_token": "<|endoftext|>"
|
31 |
+
}
|
vision_encoder/vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|