koichi12 commited on
Commit
cc325c2
·
verified ·
1 Parent(s): 04522a2

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. lib/transformers/src/transformers.egg-info/PKG-INFO +766 -0
  2. lib/transformers/src/transformers.egg-info/SOURCES.txt +0 -0
  3. lib/transformers/src/transformers.egg-info/dependency_links.txt +1 -0
  4. lib/transformers/src/transformers.egg-info/entry_points.txt +2 -0
  5. lib/transformers/src/transformers.egg-info/not-zip-safe +1 -0
  6. lib/transformers/src/transformers.egg-info/requires.txt +437 -0
  7. lib/transformers/src/transformers.egg-info/top_level.txt +1 -0
  8. lib/transformers/src/transformers/data/__pycache__/__init__.cpython-310.pyc +0 -0
  9. lib/transformers/src/transformers/data/__pycache__/data_collator.cpython-310.pyc +0 -0
  10. lib/transformers/src/transformers/data/metrics/__pycache__/__init__.cpython-310.pyc +0 -0
  11. lib/transformers/src/transformers/data/metrics/__pycache__/squad_metrics.cpython-310.pyc +0 -0
  12. lib/transformers/src/transformers/data/processors/__pycache__/__init__.cpython-310.pyc +0 -0
  13. lib/transformers/src/transformers/data/processors/__pycache__/glue.cpython-310.pyc +0 -0
  14. lib/transformers/src/transformers/data/processors/__pycache__/squad.cpython-310.pyc +0 -0
  15. lib/transformers/src/transformers/data/processors/__pycache__/utils.cpython-310.pyc +0 -0
  16. lib/transformers/src/transformers/data/processors/__pycache__/xnli.cpython-310.pyc +0 -0
  17. lib/transformers/src/transformers/generation/__pycache__/__init__.cpython-310.pyc +0 -0
  18. lib/transformers/src/transformers/generation/__pycache__/beam_constraints.cpython-310.pyc +0 -0
  19. lib/transformers/src/transformers/generation/__pycache__/beam_search.cpython-310.pyc +0 -0
  20. lib/transformers/src/transformers/generation/__pycache__/candidate_generator.cpython-310.pyc +0 -0
  21. lib/transformers/src/transformers/generation/__pycache__/configuration_utils.cpython-310.pyc +0 -0
  22. lib/transformers/src/transformers/generation/__pycache__/logits_process.cpython-310.pyc +0 -0
  23. lib/transformers/src/transformers/generation/__pycache__/stopping_criteria.cpython-310.pyc +0 -0
  24. lib/transformers/src/transformers/generation/__pycache__/utils.cpython-310.pyc +0 -0
  25. lib/transformers/src/transformers/pipelines/__pycache__/audio_utils.cpython-310.pyc +0 -0
  26. lib/transformers/src/transformers/pipelines/__pycache__/document_question_answering.cpython-310.pyc +0 -0
  27. lib/transformers/src/transformers/pipelines/__pycache__/image_classification.cpython-310.pyc +0 -0
  28. lib/transformers/src/transformers/pipelines/__pycache__/question_answering.cpython-310.pyc +0 -0
  29. megatron_lm/megatron/__pycache__/global_vars.cpython-310.pyc +0 -0
  30. megatron_lm/megatron/core/datasets/__pycache__/blended_dataset.cpython-310.pyc +0 -0
  31. megatron_lm/megatron/core/datasets/__pycache__/blended_megatron_dataset_builder.cpython-310.pyc +0 -0
  32. megatron_lm/megatron/core/datasets/__pycache__/blended_megatron_dataset_config.cpython-310.pyc +0 -0
  33. megatron_lm/megatron/core/datasets/__pycache__/gpt_dataset.cpython-310.pyc +0 -0
  34. megatron_lm/megatron/core/datasets/__pycache__/indexed_dataset.cpython-310.pyc +0 -0
  35. megatron_lm/megatron/core/datasets/__pycache__/megatron_dataset.cpython-310.pyc +0 -0
  36. megatron_lm/megatron/core/datasets/__pycache__/megatron_tokenizer.cpython-310.pyc +0 -0
  37. megatron_lm/megatron/core/datasets/__pycache__/utils.cpython-310.pyc +0 -0
  38. megatron_lm/megatron/core/datasets/helpers.cpython-310-x86_64-linux-gnu.so +0 -0
  39. megatron_lm/megatron/data/__pycache__/data_samplers.cpython-310.pyc +0 -0
  40. megatron_lm/megatron/tokenizer/__pycache__/__init__.cpython-310.pyc +0 -0
  41. megatron_lm/megatron/tokenizer/__pycache__/tokenizer.cpython-310.pyc +0 -0
  42. src/llama_recipes/__pycache__/arguments.cpython-310.pyc +0 -0
  43. src/llama_recipes/__pycache__/finetuning.cpython-310.pyc +0 -0
  44. src/llama_recipes/__pycache__/get_fsdp.cpython-310.pyc +0 -0
  45. src/llama_recipes/__pycache__/get_model_decoder_layer.cpython-310.pyc +0 -0
  46. src/llama_recipes/__pycache__/get_models.cpython-310.pyc +0 -0
  47. src/llama_recipes/__pycache__/optimizer.cpython-310.pyc +0 -0
  48. src/llama_recipes/datasets/__pycache__/pretrain_dataset.cpython-310.pyc +0 -0
  49. src/llama_recipes/policies/__pycache__/__init__.cpython-310.pyc +0 -0
  50. src/llama_recipes/policies/__pycache__/activation_checkpointing_functions.cpython-310.pyc +0 -0
lib/transformers/src/transformers.egg-info/PKG-INFO ADDED
@@ -0,0 +1,766 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Metadata-Version: 2.1
2
+ Name: transformers
3
+ Version: 4.43.3
4
+ Summary: State-of-the-art Machine Learning for JAX, PyTorch and TensorFlow
5
+ Home-page: https://github.com/huggingface/transformers
6
+ Author: The Hugging Face team (past and future) with the help of all our contributors (https://github.com/huggingface/transformers/graphs/contributors)
7
+ Author-email: [email protected]
8
+ License: Apache 2.0 License
9
+ Keywords: NLP vision speech deep learning transformer pytorch tensorflow jax BERT GPT-2 Wav2Vec2 ViT
10
+ Classifier: Development Status :: 5 - Production/Stable
11
+ Classifier: Intended Audience :: Developers
12
+ Classifier: Intended Audience :: Education
13
+ Classifier: Intended Audience :: Science/Research
14
+ Classifier: License :: OSI Approved :: Apache Software License
15
+ Classifier: Operating System :: OS Independent
16
+ Classifier: Programming Language :: Python :: 3
17
+ Classifier: Programming Language :: Python :: 3.8
18
+ Classifier: Programming Language :: Python :: 3.9
19
+ Classifier: Programming Language :: Python :: 3.10
20
+ Classifier: Topic :: Scientific/Engineering :: Artificial Intelligence
21
+ Requires-Python: >=3.8.0
22
+ Description-Content-Type: text/markdown
23
+ License-File: LICENSE
24
+ Requires-Dist: filelock
25
+ Requires-Dist: huggingface-hub<1.0,>=0.23.2
26
+ Requires-Dist: numpy>=1.17
27
+ Requires-Dist: packaging>=20.0
28
+ Requires-Dist: pyyaml>=5.1
29
+ Requires-Dist: regex!=2019.12.17
30
+ Requires-Dist: requests
31
+ Requires-Dist: tokenizers<0.20,>=0.19
32
+ Requires-Dist: safetensors>=0.4.1
33
+ Requires-Dist: tqdm>=4.27
34
+ Provides-Extra: ja
35
+ Requires-Dist: fugashi>=1.0; extra == "ja"
36
+ Requires-Dist: ipadic<2.0,>=1.0.0; extra == "ja"
37
+ Requires-Dist: unidic_lite>=1.0.7; extra == "ja"
38
+ Requires-Dist: unidic>=1.0.2; extra == "ja"
39
+ Requires-Dist: sudachipy>=0.6.6; extra == "ja"
40
+ Requires-Dist: sudachidict_core>=20220729; extra == "ja"
41
+ Requires-Dist: rhoknp<1.3.1,>=1.1.0; extra == "ja"
42
+ Provides-Extra: sklearn
43
+ Requires-Dist: scikit-learn; extra == "sklearn"
44
+ Provides-Extra: tf
45
+ Requires-Dist: tensorflow<2.16,>2.9; extra == "tf"
46
+ Requires-Dist: onnxconverter-common; extra == "tf"
47
+ Requires-Dist: tf2onnx; extra == "tf"
48
+ Requires-Dist: tensorflow-text<2.16; extra == "tf"
49
+ Requires-Dist: keras-nlp<0.14.0,>=0.3.1; extra == "tf"
50
+ Provides-Extra: tf-cpu
51
+ Requires-Dist: keras<2.16,>2.9; extra == "tf-cpu"
52
+ Requires-Dist: tensorflow-cpu<2.16,>2.9; extra == "tf-cpu"
53
+ Requires-Dist: onnxconverter-common; extra == "tf-cpu"
54
+ Requires-Dist: tf2onnx; extra == "tf-cpu"
55
+ Requires-Dist: tensorflow-text<2.16; extra == "tf-cpu"
56
+ Requires-Dist: keras-nlp<0.14.0,>=0.3.1; extra == "tf-cpu"
57
+ Requires-Dist: tensorflow-probability<0.24; extra == "tf-cpu"
58
+ Provides-Extra: torch
59
+ Requires-Dist: torch; extra == "torch"
60
+ Requires-Dist: accelerate>=0.21.0; extra == "torch"
61
+ Provides-Extra: accelerate
62
+ Requires-Dist: accelerate>=0.21.0; extra == "accelerate"
63
+ Provides-Extra: retrieval
64
+ Requires-Dist: faiss-cpu; extra == "retrieval"
65
+ Requires-Dist: datasets!=2.5.0; extra == "retrieval"
66
+ Provides-Extra: flax
67
+ Requires-Dist: jax<=0.4.13,>=0.4.1; extra == "flax"
68
+ Requires-Dist: jaxlib<=0.4.13,>=0.4.1; extra == "flax"
69
+ Requires-Dist: flax<=0.7.0,>=0.4.1; extra == "flax"
70
+ Requires-Dist: optax<=0.1.4,>=0.0.8; extra == "flax"
71
+ Requires-Dist: scipy<1.13.0; extra == "flax"
72
+ Provides-Extra: tokenizers
73
+ Requires-Dist: tokenizers<0.20,>=0.19; extra == "tokenizers"
74
+ Provides-Extra: ftfy
75
+ Requires-Dist: ftfy; extra == "ftfy"
76
+ Provides-Extra: onnxruntime
77
+ Requires-Dist: onnxruntime>=1.4.0; extra == "onnxruntime"
78
+ Requires-Dist: onnxruntime-tools>=1.4.2; extra == "onnxruntime"
79
+ Provides-Extra: onnx
80
+ Requires-Dist: onnxconverter-common; extra == "onnx"
81
+ Requires-Dist: tf2onnx; extra == "onnx"
82
+ Requires-Dist: onnxruntime>=1.4.0; extra == "onnx"
83
+ Requires-Dist: onnxruntime-tools>=1.4.2; extra == "onnx"
84
+ Provides-Extra: modelcreation
85
+ Requires-Dist: cookiecutter==1.7.3; extra == "modelcreation"
86
+ Provides-Extra: sagemaker
87
+ Requires-Dist: sagemaker>=2.31.0; extra == "sagemaker"
88
+ Provides-Extra: deepspeed
89
+ Requires-Dist: deepspeed>=0.9.3; extra == "deepspeed"
90
+ Requires-Dist: accelerate>=0.21.0; extra == "deepspeed"
91
+ Provides-Extra: optuna
92
+ Requires-Dist: optuna; extra == "optuna"
93
+ Provides-Extra: ray
94
+ Requires-Dist: ray[tune]>=2.7.0; extra == "ray"
95
+ Provides-Extra: sigopt
96
+ Requires-Dist: sigopt; extra == "sigopt"
97
+ Provides-Extra: integrations
98
+ Requires-Dist: optuna; extra == "integrations"
99
+ Requires-Dist: ray[tune]>=2.7.0; extra == "integrations"
100
+ Requires-Dist: sigopt; extra == "integrations"
101
+ Provides-Extra: serving
102
+ Requires-Dist: pydantic; extra == "serving"
103
+ Requires-Dist: uvicorn; extra == "serving"
104
+ Requires-Dist: fastapi; extra == "serving"
105
+ Requires-Dist: starlette; extra == "serving"
106
+ Provides-Extra: audio
107
+ Requires-Dist: librosa; extra == "audio"
108
+ Requires-Dist: pyctcdecode>=0.4.0; extra == "audio"
109
+ Requires-Dist: phonemizer; extra == "audio"
110
+ Requires-Dist: kenlm; extra == "audio"
111
+ Provides-Extra: speech
112
+ Requires-Dist: torchaudio; extra == "speech"
113
+ Requires-Dist: librosa; extra == "speech"
114
+ Requires-Dist: pyctcdecode>=0.4.0; extra == "speech"
115
+ Requires-Dist: phonemizer; extra == "speech"
116
+ Requires-Dist: kenlm; extra == "speech"
117
+ Provides-Extra: torch-speech
118
+ Requires-Dist: torchaudio; extra == "torch-speech"
119
+ Requires-Dist: librosa; extra == "torch-speech"
120
+ Requires-Dist: pyctcdecode>=0.4.0; extra == "torch-speech"
121
+ Requires-Dist: phonemizer; extra == "torch-speech"
122
+ Requires-Dist: kenlm; extra == "torch-speech"
123
+ Provides-Extra: tf-speech
124
+ Requires-Dist: librosa; extra == "tf-speech"
125
+ Requires-Dist: pyctcdecode>=0.4.0; extra == "tf-speech"
126
+ Requires-Dist: phonemizer; extra == "tf-speech"
127
+ Requires-Dist: kenlm; extra == "tf-speech"
128
+ Provides-Extra: flax-speech
129
+ Requires-Dist: librosa; extra == "flax-speech"
130
+ Requires-Dist: pyctcdecode>=0.4.0; extra == "flax-speech"
131
+ Requires-Dist: phonemizer; extra == "flax-speech"
132
+ Requires-Dist: kenlm; extra == "flax-speech"
133
+ Provides-Extra: vision
134
+ Requires-Dist: Pillow<=15.0,>=10.0.1; extra == "vision"
135
+ Provides-Extra: timm
136
+ Requires-Dist: timm<=0.9.16; extra == "timm"
137
+ Provides-Extra: torch-vision
138
+ Requires-Dist: torchvision; extra == "torch-vision"
139
+ Requires-Dist: Pillow<=15.0,>=10.0.1; extra == "torch-vision"
140
+ Provides-Extra: natten
141
+ Requires-Dist: natten<0.15.0,>=0.14.6; extra == "natten"
142
+ Provides-Extra: codecarbon
143
+ Requires-Dist: codecarbon==1.2.0; extra == "codecarbon"
144
+ Provides-Extra: video
145
+ Requires-Dist: decord==0.6.0; extra == "video"
146
+ Requires-Dist: av==9.2.0; extra == "video"
147
+ Provides-Extra: sentencepiece
148
+ Requires-Dist: sentencepiece!=0.1.92,>=0.1.91; extra == "sentencepiece"
149
+ Requires-Dist: protobuf; extra == "sentencepiece"
150
+ Provides-Extra: testing
151
+ Requires-Dist: pytest<8.0.0,>=7.2.0; extra == "testing"
152
+ Requires-Dist: pytest-rich; extra == "testing"
153
+ Requires-Dist: pytest-xdist; extra == "testing"
154
+ Requires-Dist: timeout-decorator; extra == "testing"
155
+ Requires-Dist: parameterized; extra == "testing"
156
+ Requires-Dist: psutil; extra == "testing"
157
+ Requires-Dist: datasets!=2.5.0; extra == "testing"
158
+ Requires-Dist: dill<0.3.5; extra == "testing"
159
+ Requires-Dist: evaluate>=0.2.0; extra == "testing"
160
+ Requires-Dist: pytest-timeout; extra == "testing"
161
+ Requires-Dist: ruff==0.4.4; extra == "testing"
162
+ Requires-Dist: sacrebleu<2.0.0,>=1.4.12; extra == "testing"
163
+ Requires-Dist: rouge-score!=0.0.7,!=0.0.8,!=0.1,!=0.1.1; extra == "testing"
164
+ Requires-Dist: nltk; extra == "testing"
165
+ Requires-Dist: GitPython<3.1.19; extra == "testing"
166
+ Requires-Dist: sacremoses; extra == "testing"
167
+ Requires-Dist: rjieba; extra == "testing"
168
+ Requires-Dist: beautifulsoup4; extra == "testing"
169
+ Requires-Dist: tensorboard; extra == "testing"
170
+ Requires-Dist: pydantic; extra == "testing"
171
+ Requires-Dist: sentencepiece!=0.1.92,>=0.1.91; extra == "testing"
172
+ Requires-Dist: faiss-cpu; extra == "testing"
173
+ Requires-Dist: datasets!=2.5.0; extra == "testing"
174
+ Requires-Dist: cookiecutter==1.7.3; extra == "testing"
175
+ Provides-Extra: deepspeed-testing
176
+ Requires-Dist: deepspeed>=0.9.3; extra == "deepspeed-testing"
177
+ Requires-Dist: accelerate>=0.21.0; extra == "deepspeed-testing"
178
+ Requires-Dist: pytest<8.0.0,>=7.2.0; extra == "deepspeed-testing"
179
+ Requires-Dist: pytest-rich; extra == "deepspeed-testing"
180
+ Requires-Dist: pytest-xdist; extra == "deepspeed-testing"
181
+ Requires-Dist: timeout-decorator; extra == "deepspeed-testing"
182
+ Requires-Dist: parameterized; extra == "deepspeed-testing"
183
+ Requires-Dist: psutil; extra == "deepspeed-testing"
184
+ Requires-Dist: datasets!=2.5.0; extra == "deepspeed-testing"
185
+ Requires-Dist: dill<0.3.5; extra == "deepspeed-testing"
186
+ Requires-Dist: evaluate>=0.2.0; extra == "deepspeed-testing"
187
+ Requires-Dist: pytest-timeout; extra == "deepspeed-testing"
188
+ Requires-Dist: ruff==0.4.4; extra == "deepspeed-testing"
189
+ Requires-Dist: sacrebleu<2.0.0,>=1.4.12; extra == "deepspeed-testing"
190
+ Requires-Dist: rouge-score!=0.0.7,!=0.0.8,!=0.1,!=0.1.1; extra == "deepspeed-testing"
191
+ Requires-Dist: nltk; extra == "deepspeed-testing"
192
+ Requires-Dist: GitPython<3.1.19; extra == "deepspeed-testing"
193
+ Requires-Dist: sacremoses; extra == "deepspeed-testing"
194
+ Requires-Dist: rjieba; extra == "deepspeed-testing"
195
+ Requires-Dist: beautifulsoup4; extra == "deepspeed-testing"
196
+ Requires-Dist: tensorboard; extra == "deepspeed-testing"
197
+ Requires-Dist: pydantic; extra == "deepspeed-testing"
198
+ Requires-Dist: sentencepiece!=0.1.92,>=0.1.91; extra == "deepspeed-testing"
199
+ Requires-Dist: faiss-cpu; extra == "deepspeed-testing"
200
+ Requires-Dist: datasets!=2.5.0; extra == "deepspeed-testing"
201
+ Requires-Dist: cookiecutter==1.7.3; extra == "deepspeed-testing"
202
+ Requires-Dist: optuna; extra == "deepspeed-testing"
203
+ Requires-Dist: sentencepiece!=0.1.92,>=0.1.91; extra == "deepspeed-testing"
204
+ Requires-Dist: protobuf; extra == "deepspeed-testing"
205
+ Provides-Extra: ruff
206
+ Requires-Dist: ruff==0.4.4; extra == "ruff"
207
+ Provides-Extra: quality
208
+ Requires-Dist: datasets!=2.5.0; extra == "quality"
209
+ Requires-Dist: isort>=5.5.4; extra == "quality"
210
+ Requires-Dist: ruff==0.4.4; extra == "quality"
211
+ Requires-Dist: GitPython<3.1.19; extra == "quality"
212
+ Requires-Dist: urllib3<2.0.0; extra == "quality"
213
+ Provides-Extra: all
214
+ Requires-Dist: tensorflow<2.16,>2.9; extra == "all"
215
+ Requires-Dist: onnxconverter-common; extra == "all"
216
+ Requires-Dist: tf2onnx; extra == "all"
217
+ Requires-Dist: tensorflow-text<2.16; extra == "all"
218
+ Requires-Dist: keras-nlp<0.14.0,>=0.3.1; extra == "all"
219
+ Requires-Dist: torch; extra == "all"
220
+ Requires-Dist: accelerate>=0.21.0; extra == "all"
221
+ Requires-Dist: jax<=0.4.13,>=0.4.1; extra == "all"
222
+ Requires-Dist: jaxlib<=0.4.13,>=0.4.1; extra == "all"
223
+ Requires-Dist: flax<=0.7.0,>=0.4.1; extra == "all"
224
+ Requires-Dist: optax<=0.1.4,>=0.0.8; extra == "all"
225
+ Requires-Dist: scipy<1.13.0; extra == "all"
226
+ Requires-Dist: sentencepiece!=0.1.92,>=0.1.91; extra == "all"
227
+ Requires-Dist: protobuf; extra == "all"
228
+ Requires-Dist: tokenizers<0.20,>=0.19; extra == "all"
229
+ Requires-Dist: torchaudio; extra == "all"
230
+ Requires-Dist: librosa; extra == "all"
231
+ Requires-Dist: pyctcdecode>=0.4.0; extra == "all"
232
+ Requires-Dist: phonemizer; extra == "all"
233
+ Requires-Dist: kenlm; extra == "all"
234
+ Requires-Dist: Pillow<=15.0,>=10.0.1; extra == "all"
235
+ Requires-Dist: optuna; extra == "all"
236
+ Requires-Dist: ray[tune]>=2.7.0; extra == "all"
237
+ Requires-Dist: sigopt; extra == "all"
238
+ Requires-Dist: timm<=0.9.16; extra == "all"
239
+ Requires-Dist: torchvision; extra == "all"
240
+ Requires-Dist: Pillow<=15.0,>=10.0.1; extra == "all"
241
+ Requires-Dist: codecarbon==1.2.0; extra == "all"
242
+ Requires-Dist: accelerate>=0.21.0; extra == "all"
243
+ Requires-Dist: decord==0.6.0; extra == "all"
244
+ Requires-Dist: av==9.2.0; extra == "all"
245
+ Provides-Extra: dev-torch
246
+ Requires-Dist: pytest<8.0.0,>=7.2.0; extra == "dev-torch"
247
+ Requires-Dist: pytest-rich; extra == "dev-torch"
248
+ Requires-Dist: pytest-xdist; extra == "dev-torch"
249
+ Requires-Dist: timeout-decorator; extra == "dev-torch"
250
+ Requires-Dist: parameterized; extra == "dev-torch"
251
+ Requires-Dist: psutil; extra == "dev-torch"
252
+ Requires-Dist: datasets!=2.5.0; extra == "dev-torch"
253
+ Requires-Dist: dill<0.3.5; extra == "dev-torch"
254
+ Requires-Dist: evaluate>=0.2.0; extra == "dev-torch"
255
+ Requires-Dist: pytest-timeout; extra == "dev-torch"
256
+ Requires-Dist: ruff==0.4.4; extra == "dev-torch"
257
+ Requires-Dist: sacrebleu<2.0.0,>=1.4.12; extra == "dev-torch"
258
+ Requires-Dist: rouge-score!=0.0.7,!=0.0.8,!=0.1,!=0.1.1; extra == "dev-torch"
259
+ Requires-Dist: nltk; extra == "dev-torch"
260
+ Requires-Dist: GitPython<3.1.19; extra == "dev-torch"
261
+ Requires-Dist: sacremoses; extra == "dev-torch"
262
+ Requires-Dist: rjieba; extra == "dev-torch"
263
+ Requires-Dist: beautifulsoup4; extra == "dev-torch"
264
+ Requires-Dist: tensorboard; extra == "dev-torch"
265
+ Requires-Dist: pydantic; extra == "dev-torch"
266
+ Requires-Dist: sentencepiece!=0.1.92,>=0.1.91; extra == "dev-torch"
267
+ Requires-Dist: faiss-cpu; extra == "dev-torch"
268
+ Requires-Dist: datasets!=2.5.0; extra == "dev-torch"
269
+ Requires-Dist: cookiecutter==1.7.3; extra == "dev-torch"
270
+ Requires-Dist: torch; extra == "dev-torch"
271
+ Requires-Dist: accelerate>=0.21.0; extra == "dev-torch"
272
+ Requires-Dist: sentencepiece!=0.1.92,>=0.1.91; extra == "dev-torch"
273
+ Requires-Dist: protobuf; extra == "dev-torch"
274
+ Requires-Dist: tokenizers<0.20,>=0.19; extra == "dev-torch"
275
+ Requires-Dist: torchaudio; extra == "dev-torch"
276
+ Requires-Dist: librosa; extra == "dev-torch"
277
+ Requires-Dist: pyctcdecode>=0.4.0; extra == "dev-torch"
278
+ Requires-Dist: phonemizer; extra == "dev-torch"
279
+ Requires-Dist: kenlm; extra == "dev-torch"
280
+ Requires-Dist: Pillow<=15.0,>=10.0.1; extra == "dev-torch"
281
+ Requires-Dist: optuna; extra == "dev-torch"
282
+ Requires-Dist: ray[tune]>=2.7.0; extra == "dev-torch"
283
+ Requires-Dist: sigopt; extra == "dev-torch"
284
+ Requires-Dist: timm<=0.9.16; extra == "dev-torch"
285
+ Requires-Dist: torchvision; extra == "dev-torch"
286
+ Requires-Dist: Pillow<=15.0,>=10.0.1; extra == "dev-torch"
287
+ Requires-Dist: codecarbon==1.2.0; extra == "dev-torch"
288
+ Requires-Dist: datasets!=2.5.0; extra == "dev-torch"
289
+ Requires-Dist: isort>=5.5.4; extra == "dev-torch"
290
+ Requires-Dist: ruff==0.4.4; extra == "dev-torch"
291
+ Requires-Dist: GitPython<3.1.19; extra == "dev-torch"
292
+ Requires-Dist: urllib3<2.0.0; extra == "dev-torch"
293
+ Requires-Dist: fugashi>=1.0; extra == "dev-torch"
294
+ Requires-Dist: ipadic<2.0,>=1.0.0; extra == "dev-torch"
295
+ Requires-Dist: unidic_lite>=1.0.7; extra == "dev-torch"
296
+ Requires-Dist: unidic>=1.0.2; extra == "dev-torch"
297
+ Requires-Dist: sudachipy>=0.6.6; extra == "dev-torch"
298
+ Requires-Dist: sudachidict_core>=20220729; extra == "dev-torch"
299
+ Requires-Dist: rhoknp<1.3.1,>=1.1.0; extra == "dev-torch"
300
+ Requires-Dist: scikit-learn; extra == "dev-torch"
301
+ Requires-Dist: cookiecutter==1.7.3; extra == "dev-torch"
302
+ Requires-Dist: onnxruntime>=1.4.0; extra == "dev-torch"
303
+ Requires-Dist: onnxruntime-tools>=1.4.2; extra == "dev-torch"
304
+ Provides-Extra: dev-tensorflow
305
+ Requires-Dist: pytest<8.0.0,>=7.2.0; extra == "dev-tensorflow"
306
+ Requires-Dist: pytest-rich; extra == "dev-tensorflow"
307
+ Requires-Dist: pytest-xdist; extra == "dev-tensorflow"
308
+ Requires-Dist: timeout-decorator; extra == "dev-tensorflow"
309
+ Requires-Dist: parameterized; extra == "dev-tensorflow"
310
+ Requires-Dist: psutil; extra == "dev-tensorflow"
311
+ Requires-Dist: datasets!=2.5.0; extra == "dev-tensorflow"
312
+ Requires-Dist: dill<0.3.5; extra == "dev-tensorflow"
313
+ Requires-Dist: evaluate>=0.2.0; extra == "dev-tensorflow"
314
+ Requires-Dist: pytest-timeout; extra == "dev-tensorflow"
315
+ Requires-Dist: ruff==0.4.4; extra == "dev-tensorflow"
316
+ Requires-Dist: sacrebleu<2.0.0,>=1.4.12; extra == "dev-tensorflow"
317
+ Requires-Dist: rouge-score!=0.0.7,!=0.0.8,!=0.1,!=0.1.1; extra == "dev-tensorflow"
318
+ Requires-Dist: nltk; extra == "dev-tensorflow"
319
+ Requires-Dist: GitPython<3.1.19; extra == "dev-tensorflow"
320
+ Requires-Dist: sacremoses; extra == "dev-tensorflow"
321
+ Requires-Dist: rjieba; extra == "dev-tensorflow"
322
+ Requires-Dist: beautifulsoup4; extra == "dev-tensorflow"
323
+ Requires-Dist: tensorboard; extra == "dev-tensorflow"
324
+ Requires-Dist: pydantic; extra == "dev-tensorflow"
325
+ Requires-Dist: sentencepiece!=0.1.92,>=0.1.91; extra == "dev-tensorflow"
326
+ Requires-Dist: faiss-cpu; extra == "dev-tensorflow"
327
+ Requires-Dist: datasets!=2.5.0; extra == "dev-tensorflow"
328
+ Requires-Dist: cookiecutter==1.7.3; extra == "dev-tensorflow"
329
+ Requires-Dist: tensorflow<2.16,>2.9; extra == "dev-tensorflow"
330
+ Requires-Dist: onnxconverter-common; extra == "dev-tensorflow"
331
+ Requires-Dist: tf2onnx; extra == "dev-tensorflow"
332
+ Requires-Dist: tensorflow-text<2.16; extra == "dev-tensorflow"
333
+ Requires-Dist: keras-nlp<0.14.0,>=0.3.1; extra == "dev-tensorflow"
334
+ Requires-Dist: sentencepiece!=0.1.92,>=0.1.91; extra == "dev-tensorflow"
335
+ Requires-Dist: protobuf; extra == "dev-tensorflow"
336
+ Requires-Dist: tokenizers<0.20,>=0.19; extra == "dev-tensorflow"
337
+ Requires-Dist: Pillow<=15.0,>=10.0.1; extra == "dev-tensorflow"
338
+ Requires-Dist: datasets!=2.5.0; extra == "dev-tensorflow"
339
+ Requires-Dist: isort>=5.5.4; extra == "dev-tensorflow"
340
+ Requires-Dist: ruff==0.4.4; extra == "dev-tensorflow"
341
+ Requires-Dist: GitPython<3.1.19; extra == "dev-tensorflow"
342
+ Requires-Dist: urllib3<2.0.0; extra == "dev-tensorflow"
343
+ Requires-Dist: scikit-learn; extra == "dev-tensorflow"
344
+ Requires-Dist: cookiecutter==1.7.3; extra == "dev-tensorflow"
345
+ Requires-Dist: onnxconverter-common; extra == "dev-tensorflow"
346
+ Requires-Dist: tf2onnx; extra == "dev-tensorflow"
347
+ Requires-Dist: onnxruntime>=1.4.0; extra == "dev-tensorflow"
348
+ Requires-Dist: onnxruntime-tools>=1.4.2; extra == "dev-tensorflow"
349
+ Requires-Dist: librosa; extra == "dev-tensorflow"
350
+ Requires-Dist: pyctcdecode>=0.4.0; extra == "dev-tensorflow"
351
+ Requires-Dist: phonemizer; extra == "dev-tensorflow"
352
+ Requires-Dist: kenlm; extra == "dev-tensorflow"
353
+ Provides-Extra: dev
354
+ Requires-Dist: tensorflow<2.16,>2.9; extra == "dev"
355
+ Requires-Dist: onnxconverter-common; extra == "dev"
356
+ Requires-Dist: tf2onnx; extra == "dev"
357
+ Requires-Dist: tensorflow-text<2.16; extra == "dev"
358
+ Requires-Dist: keras-nlp<0.14.0,>=0.3.1; extra == "dev"
359
+ Requires-Dist: torch; extra == "dev"
360
+ Requires-Dist: accelerate>=0.21.0; extra == "dev"
361
+ Requires-Dist: jax<=0.4.13,>=0.4.1; extra == "dev"
362
+ Requires-Dist: jaxlib<=0.4.13,>=0.4.1; extra == "dev"
363
+ Requires-Dist: flax<=0.7.0,>=0.4.1; extra == "dev"
364
+ Requires-Dist: optax<=0.1.4,>=0.0.8; extra == "dev"
365
+ Requires-Dist: scipy<1.13.0; extra == "dev"
366
+ Requires-Dist: sentencepiece!=0.1.92,>=0.1.91; extra == "dev"
367
+ Requires-Dist: protobuf; extra == "dev"
368
+ Requires-Dist: tokenizers<0.20,>=0.19; extra == "dev"
369
+ Requires-Dist: torchaudio; extra == "dev"
370
+ Requires-Dist: librosa; extra == "dev"
371
+ Requires-Dist: pyctcdecode>=0.4.0; extra == "dev"
372
+ Requires-Dist: phonemizer; extra == "dev"
373
+ Requires-Dist: kenlm; extra == "dev"
374
+ Requires-Dist: Pillow<=15.0,>=10.0.1; extra == "dev"
375
+ Requires-Dist: optuna; extra == "dev"
376
+ Requires-Dist: ray[tune]>=2.7.0; extra == "dev"
377
+ Requires-Dist: sigopt; extra == "dev"
378
+ Requires-Dist: timm<=0.9.16; extra == "dev"
379
+ Requires-Dist: torchvision; extra == "dev"
380
+ Requires-Dist: Pillow<=15.0,>=10.0.1; extra == "dev"
381
+ Requires-Dist: codecarbon==1.2.0; extra == "dev"
382
+ Requires-Dist: accelerate>=0.21.0; extra == "dev"
383
+ Requires-Dist: decord==0.6.0; extra == "dev"
384
+ Requires-Dist: av==9.2.0; extra == "dev"
385
+ Requires-Dist: pytest<8.0.0,>=7.2.0; extra == "dev"
386
+ Requires-Dist: pytest-rich; extra == "dev"
387
+ Requires-Dist: pytest-xdist; extra == "dev"
388
+ Requires-Dist: timeout-decorator; extra == "dev"
389
+ Requires-Dist: parameterized; extra == "dev"
390
+ Requires-Dist: psutil; extra == "dev"
391
+ Requires-Dist: datasets!=2.5.0; extra == "dev"
392
+ Requires-Dist: dill<0.3.5; extra == "dev"
393
+ Requires-Dist: evaluate>=0.2.0; extra == "dev"
394
+ Requires-Dist: pytest-timeout; extra == "dev"
395
+ Requires-Dist: ruff==0.4.4; extra == "dev"
396
+ Requires-Dist: sacrebleu<2.0.0,>=1.4.12; extra == "dev"
397
+ Requires-Dist: rouge-score!=0.0.7,!=0.0.8,!=0.1,!=0.1.1; extra == "dev"
398
+ Requires-Dist: nltk; extra == "dev"
399
+ Requires-Dist: GitPython<3.1.19; extra == "dev"
400
+ Requires-Dist: sacremoses; extra == "dev"
401
+ Requires-Dist: rjieba; extra == "dev"
402
+ Requires-Dist: beautifulsoup4; extra == "dev"
403
+ Requires-Dist: tensorboard; extra == "dev"
404
+ Requires-Dist: pydantic; extra == "dev"
405
+ Requires-Dist: sentencepiece!=0.1.92,>=0.1.91; extra == "dev"
406
+ Requires-Dist: faiss-cpu; extra == "dev"
407
+ Requires-Dist: datasets!=2.5.0; extra == "dev"
408
+ Requires-Dist: cookiecutter==1.7.3; extra == "dev"
409
+ Requires-Dist: datasets!=2.5.0; extra == "dev"
410
+ Requires-Dist: isort>=5.5.4; extra == "dev"
411
+ Requires-Dist: ruff==0.4.4; extra == "dev"
412
+ Requires-Dist: GitPython<3.1.19; extra == "dev"
413
+ Requires-Dist: urllib3<2.0.0; extra == "dev"
414
+ Requires-Dist: fugashi>=1.0; extra == "dev"
415
+ Requires-Dist: ipadic<2.0,>=1.0.0; extra == "dev"
416
+ Requires-Dist: unidic_lite>=1.0.7; extra == "dev"
417
+ Requires-Dist: unidic>=1.0.2; extra == "dev"
418
+ Requires-Dist: sudachipy>=0.6.6; extra == "dev"
419
+ Requires-Dist: sudachidict_core>=20220729; extra == "dev"
420
+ Requires-Dist: rhoknp<1.3.1,>=1.1.0; extra == "dev"
421
+ Requires-Dist: scikit-learn; extra == "dev"
422
+ Requires-Dist: cookiecutter==1.7.3; extra == "dev"
423
+ Provides-Extra: torchhub
424
+ Requires-Dist: filelock; extra == "torchhub"
425
+ Requires-Dist: huggingface-hub<1.0,>=0.23.2; extra == "torchhub"
426
+ Requires-Dist: importlib_metadata; extra == "torchhub"
427
+ Requires-Dist: numpy>=1.17; extra == "torchhub"
428
+ Requires-Dist: packaging>=20.0; extra == "torchhub"
429
+ Requires-Dist: protobuf; extra == "torchhub"
430
+ Requires-Dist: regex!=2019.12.17; extra == "torchhub"
431
+ Requires-Dist: requests; extra == "torchhub"
432
+ Requires-Dist: sentencepiece!=0.1.92,>=0.1.91; extra == "torchhub"
433
+ Requires-Dist: torch; extra == "torchhub"
434
+ Requires-Dist: tokenizers<0.20,>=0.19; extra == "torchhub"
435
+ Requires-Dist: tqdm>=4.27; extra == "torchhub"
436
+ Provides-Extra: agents
437
+ Requires-Dist: diffusers; extra == "agents"
438
+ Requires-Dist: accelerate>=0.21.0; extra == "agents"
439
+ Requires-Dist: datasets!=2.5.0; extra == "agents"
440
+ Requires-Dist: torch; extra == "agents"
441
+ Requires-Dist: sentencepiece!=0.1.92,>=0.1.91; extra == "agents"
442
+ Requires-Dist: opencv-python; extra == "agents"
443
+ Requires-Dist: Pillow<=15.0,>=10.0.1; extra == "agents"
444
+ Provides-Extra: benchmark
445
+ Requires-Dist: optimum-benchmark>=0.2.0; extra == "benchmark"
446
+
447
+ <!---
448
+ Copyright 2020 The HuggingFace Team. All rights reserved.
449
+
450
+ Licensed under the Apache License, Version 2.0 (the "License");
451
+ you may not use this file except in compliance with the License.
452
+ You may obtain a copy of the License at
453
+
454
+ http://www.apache.org/licenses/LICENSE-2.0
455
+
456
+ Unless required by applicable law or agreed to in writing, software
457
+ distributed under the License is distributed on an "AS IS" BASIS,
458
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
459
+ See the License for the specific language governing permissions and
460
+ limitations under the License.
461
+ -->
462
+
463
+ <p align="center">
464
+ <picture>
465
+ <source media="(prefers-color-scheme: dark)" srcset="https://huggingface.co/datasets/huggingface/documentation-images/raw/main/transformers-logo-dark.svg">
466
+ <source media="(prefers-color-scheme: light)" srcset="https://huggingface.co/datasets/huggingface/documentation-images/raw/main/transformers-logo-light.svg">
467
+ <img alt="Hugging Face Transformers Library" src="https://huggingface.co/datasets/huggingface/documentation-images/raw/main/transformers-logo-light.svg" width="352" height="59" style="max-width: 100%;">
468
+ </picture>
469
+ <br/>
470
+ <br/>
471
+ </p>
472
+
473
+ <p align="center">
474
+ <a href="https://circleci.com/gh/huggingface/transformers"><img alt="Build" src="https://img.shields.io/circleci/build/github/huggingface/transformers/main"></a>
475
+ <a href="https://github.com/huggingface/transformers/blob/main/LICENSE"><img alt="GitHub" src="https://img.shields.io/github/license/huggingface/transformers.svg?color=blue"></a>
476
+ <a href="https://huggingface.co/docs/transformers/index"><img alt="Documentation" src="https://img.shields.io/website/http/huggingface.co/docs/transformers/index.svg?down_color=red&down_message=offline&up_message=online"></a>
477
+ <a href="https://github.com/huggingface/transformers/releases"><img alt="GitHub release" src="https://img.shields.io/github/release/huggingface/transformers.svg"></a>
478
+ <a href="https://github.com/huggingface/transformers/blob/main/CODE_OF_CONDUCT.md"><img alt="Contributor Covenant" src="https://img.shields.io/badge/Contributor%20Covenant-v2.0%20adopted-ff69b4.svg"></a>
479
+ <a href="https://zenodo.org/badge/latestdoi/155220641"><img src="https://zenodo.org/badge/155220641.svg" alt="DOI"></a>
480
+ </p>
481
+
482
+ <h4 align="center">
483
+ <p>
484
+ <b>English</b> |
485
+ <a href="https://github.com/huggingface/transformers/blob/main/i18n/README_zh-hans.md">简体中文</a> |
486
+ <a href="https://github.com/huggingface/transformers/blob/main/i18n/README_zh-hant.md">繁體中文</a> |
487
+ <a href="https://github.com/huggingface/transformers/blob/main/i18n/README_ko.md">한국어</a> |
488
+ <a href="https://github.com/huggingface/transformers/blob/main/i18n/README_es.md">Español</a> |
489
+ <a href="https://github.com/huggingface/transformers/blob/main/i18n/README_ja.md">日本語</a> |
490
+ <a href="https://github.com/huggingface/transformers/blob/main/i18n/README_hd.md">हिन्दी</a> |
491
+ <a href="https://github.com/huggingface/transformers/blob/main/i18n/README_ru.md">Русский</a> |
492
+ <a href="https://github.com/huggingface/transformers/blob/main/i18n/README_pt-br.md">Рortuguês</a> |
493
+ <a href="https://github.com/huggingface/transformers/blob/main/i18n/README_te.md">తెలుగు</a> |
494
+ <a href="https://github.com/huggingface/transformers/blob/main/i18n/README_fr.md">Français</a> |
495
+ <a href="https://github.com/huggingface/transformers/blob/main/i18n/README_de.md">Deutsch</a> |
496
+ <a href="https://github.com/huggingface/transformers/blob/main/i18n/README_vi.md">Tiếng Việt</a> |
497
+ </p>
498
+ </h4>
499
+
500
+ <h3 align="center">
501
+ <p>State-of-the-art Machine Learning for JAX, PyTorch and TensorFlow</p>
502
+ </h3>
503
+
504
+ <h3 align="center">
505
+ <a href="https://hf.co/course"><img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/course_banner.png"></a>
506
+ </h3>
507
+
508
+ 🤗 Transformers provides thousands of pretrained models to perform tasks on different modalities such as text, vision, and audio.
509
+
510
+ These models can be applied on:
511
+
512
+ * 📝 Text, for tasks like text classification, information extraction, question answering, summarization, translation, and text generation, in over 100 languages.
513
+ * 🖼️ Images, for tasks like image classification, object detection, and segmentation.
514
+ * 🗣️ Audio, for tasks like speech recognition and audio classification.
515
+
516
+ Transformer models can also perform tasks on **several modalities combined**, such as table question answering, optical character recognition, information extraction from scanned documents, video classification, and visual question answering.
517
+
518
+ 🤗 Transformers provides APIs to quickly download and use those pretrained models on a given text, fine-tune them on your own datasets and then share them with the community on our [model hub](https://huggingface.co/models). At the same time, each python module defining an architecture is fully standalone and can be modified to enable quick research experiments.
519
+
520
+ 🤗 Transformers is backed by the three most popular deep learning libraries — [Jax](https://jax.readthedocs.io/en/latest/), [PyTorch](https://pytorch.org/) and [TensorFlow](https://www.tensorflow.org/) — with a seamless integration between them. It's straightforward to train your models with one before loading them for inference with the other.
521
+
522
+ ## Online demos
523
+
524
+ You can test most of our models directly on their pages from the [model hub](https://huggingface.co/models). We also offer [private model hosting, versioning, & an inference API](https://huggingface.co/pricing) for public and private models.
525
+
526
+ Here are a few examples:
527
+
528
+ In Natural Language Processing:
529
+ - [Masked word completion with BERT](https://huggingface.co/google-bert/bert-base-uncased?text=Paris+is+the+%5BMASK%5D+of+France)
530
+ - [Named Entity Recognition with Electra](https://huggingface.co/dbmdz/electra-large-discriminator-finetuned-conll03-english?text=My+name+is+Sarah+and+I+live+in+London+city)
531
+ - [Text generation with Mistral](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2)
532
+ - [Natural Language Inference with RoBERTa](https://huggingface.co/FacebookAI/roberta-large-mnli?text=The+dog+was+lost.+Nobody+lost+any+animal)
533
+ - [Summarization with BART](https://huggingface.co/facebook/bart-large-cnn?text=The+tower+is+324+metres+%281%2C063+ft%29+tall%2C+about+the+same+height+as+an+81-storey+building%2C+and+the+tallest+structure+in+Paris.+Its+base+is+square%2C+measuring+125+metres+%28410+ft%29+on+each+side.+During+its+construction%2C+the+Eiffel+Tower+surpassed+the+Washington+Monument+to+become+the+tallest+man-made+structure+in+the+world%2C+a+title+it+held+for+41+years+until+the+Chrysler+Building+in+New+York+City+was+finished+in+1930.+It+was+the+first+structure+to+reach+a+height+of+300+metres.+Due+to+the+addition+of+a+broadcasting+aerial+at+the+top+of+the+tower+in+1957%2C+it+is+now+taller+than+the+Chrysler+Building+by+5.2+metres+%2817+ft%29.+Excluding+transmitters%2C+the+Eiffel+Tower+is+the+second+tallest+free-standing+structure+in+France+after+the+Millau+Viaduct)
534
+ - [Question answering with DistilBERT](https://huggingface.co/distilbert/distilbert-base-uncased-distilled-squad?text=Which+name+is+also+used+to+describe+the+Amazon+rainforest+in+English%3F&context=The+Amazon+rainforest+%28Portuguese%3A+Floresta+Amaz%C3%B4nica+or+Amaz%C3%B4nia%3B+Spanish%3A+Selva+Amaz%C3%B3nica%2C+Amazon%C3%ADa+or+usually+Amazonia%3B+French%3A+For%C3%AAt+amazonienne%3B+Dutch%3A+Amazoneregenwoud%29%2C+also+known+in+English+as+Amazonia+or+the+Amazon+Jungle%2C+is+a+moist+broadleaf+forest+that+covers+most+of+the+Amazon+basin+of+South+America.+This+basin+encompasses+7%2C000%2C000+square+kilometres+%282%2C700%2C000+sq+mi%29%2C+of+which+5%2C500%2C000+square+kilometres+%282%2C100%2C000+sq+mi%29+are+covered+by+the+rainforest.+This+region+includes+territory+belonging+to+nine+nations.+The+majority+of+the+forest+is+contained+within+Brazil%2C+with+60%25+of+the+rainforest%2C+followed+by+Peru+with+13%25%2C+Colombia+with+10%25%2C+and+with+minor+amounts+in+Venezuela%2C+Ecuador%2C+Bolivia%2C+Guyana%2C+Suriname+and+French+Guiana.+States+or+departments+in+four+nations+contain+%22Amazonas%22+in+their+names.+The+Amazon+represents+over+half+of+the+planet%27s+remaining+rainforests%2C+and+comprises+the+largest+and+most+biodiverse+tract+of+tropical+rainforest+in+the+world%2C+with+an+estimated+390+billion+individual+trees+divided+into+16%2C000+species)
535
+ - [Translation with T5](https://huggingface.co/google-t5/t5-base?text=My+name+is+Wolfgang+and+I+live+in+Berlin)
536
+
537
+ In Computer Vision:
538
+ - [Image classification with ViT](https://huggingface.co/google/vit-base-patch16-224)
539
+ - [Object Detection with DETR](https://huggingface.co/facebook/detr-resnet-50)
540
+ - [Semantic Segmentation with SegFormer](https://huggingface.co/nvidia/segformer-b0-finetuned-ade-512-512)
541
+ - [Panoptic Segmentation with Mask2Former](https://huggingface.co/facebook/mask2former-swin-large-coco-panoptic)
542
+ - [Depth Estimation with Depth Anything](https://huggingface.co/docs/transformers/main/model_doc/depth_anything)
543
+ - [Video Classification with VideoMAE](https://huggingface.co/docs/transformers/model_doc/videomae)
544
+ - [Universal Segmentation with OneFormer](https://huggingface.co/shi-labs/oneformer_ade20k_dinat_large)
545
+
546
+ In Audio:
547
+ - [Automatic Speech Recognition with Whisper](https://huggingface.co/openai/whisper-large-v3)
548
+ - [Keyword Spotting with Wav2Vec2](https://huggingface.co/superb/wav2vec2-base-superb-ks)
549
+ - [Audio Classification with Audio Spectrogram Transformer](https://huggingface.co/MIT/ast-finetuned-audioset-10-10-0.4593)
550
+
551
+ In Multimodal tasks:
552
+ - [Table Question Answering with TAPAS](https://huggingface.co/google/tapas-base-finetuned-wtq)
553
+ - [Visual Question Answering with ViLT](https://huggingface.co/dandelin/vilt-b32-finetuned-vqa)
554
+ - [Image captioning with LLaVa](https://huggingface.co/llava-hf/llava-1.5-7b-hf)
555
+ - [Zero-shot Image Classification with SigLIP](https://huggingface.co/google/siglip-so400m-patch14-384)
556
+ - [Document Question Answering with LayoutLM](https://huggingface.co/impira/layoutlm-document-qa)
557
+ - [Zero-shot Video Classification with X-CLIP](https://huggingface.co/docs/transformers/model_doc/xclip)
558
+ - [Zero-shot Object Detection with OWLv2](https://huggingface.co/docs/transformers/en/model_doc/owlv2)
559
+ - [Zero-shot Image Segmentation with CLIPSeg](https://huggingface.co/docs/transformers/model_doc/clipseg)
560
+ - [Automatic Mask Generation with SAM](https://huggingface.co/docs/transformers/model_doc/sam)
561
+
562
+
563
+ ## 100 projects using Transformers
564
+
565
+ Transformers is more than a toolkit to use pretrained models: it's a community of projects built around it and the
566
+ Hugging Face Hub. We want Transformers to enable developers, researchers, students, professors, engineers, and anyone
567
+ else to build their dream projects.
568
+
569
+ In order to celebrate the 100,000 stars of transformers, we have decided to put the spotlight on the
570
+ community, and we have created the [awesome-transformers](./awesome-transformers.md) page which lists 100
571
+ incredible projects built in the vicinity of transformers.
572
+
573
+ If you own or use a project that you believe should be part of the list, please open a PR to add it!
574
+
575
+ ## If you are looking for custom support from the Hugging Face team
576
+
577
+ <a target="_blank" href="https://huggingface.co/support">
578
+ <img alt="HuggingFace Expert Acceleration Program" src="https://cdn-media.huggingface.co/marketing/transformers/new-support-improved.png" style="max-width: 600px; border: 1px solid #eee; border-radius: 4px; box-shadow: 0 1px 2px 0 rgba(0, 0, 0, 0.05);">
579
+ </a><br>
580
+
581
+ ## Quick tour
582
+
583
+ To immediately use a model on a given input (text, image, audio, ...), we provide the `pipeline` API. Pipelines group together a pretrained model with the preprocessing that was used during that model's training. Here is how to quickly use a pipeline to classify positive versus negative texts:
584
+
585
+ ```python
586
+ >>> from transformers import pipeline
587
+
588
+ # Allocate a pipeline for sentiment-analysis
589
+ >>> classifier = pipeline('sentiment-analysis')
590
+ >>> classifier('We are very happy to introduce pipeline to the transformers repository.')
591
+ [{'label': 'POSITIVE', 'score': 0.9996980428695679}]
592
+ ```
593
+
594
+ The second line of code downloads and caches the pretrained model used by the pipeline, while the third evaluates it on the given text. Here, the answer is "positive" with a confidence of 99.97%.
595
+
596
+ Many tasks have a pre-trained `pipeline` ready to go, in NLP but also in computer vision and speech. For example, we can easily extract detected objects in an image:
597
+
598
+ ``` python
599
+ >>> import requests
600
+ >>> from PIL import Image
601
+ >>> from transformers import pipeline
602
+
603
+ # Download an image with cute cats
604
+ >>> url = "https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/coco_sample.png"
605
+ >>> image_data = requests.get(url, stream=True).raw
606
+ >>> image = Image.open(image_data)
607
+
608
+ # Allocate a pipeline for object detection
609
+ >>> object_detector = pipeline('object-detection')
610
+ >>> object_detector(image)
611
+ [{'score': 0.9982201457023621,
612
+ 'label': 'remote',
613
+ 'box': {'xmin': 40, 'ymin': 70, 'xmax': 175, 'ymax': 117}},
614
+ {'score': 0.9960021376609802,
615
+ 'label': 'remote',
616
+ 'box': {'xmin': 333, 'ymin': 72, 'xmax': 368, 'ymax': 187}},
617
+ {'score': 0.9954745173454285,
618
+ 'label': 'couch',
619
+ 'box': {'xmin': 0, 'ymin': 1, 'xmax': 639, 'ymax': 473}},
620
+ {'score': 0.9988006353378296,
621
+ 'label': 'cat',
622
+ 'box': {'xmin': 13, 'ymin': 52, 'xmax': 314, 'ymax': 470}},
623
+ {'score': 0.9986783862113953,
624
+ 'label': 'cat',
625
+ 'box': {'xmin': 345, 'ymin': 23, 'xmax': 640, 'ymax': 368}}]
626
+ ```
627
+
628
+ Here, we get a list of objects detected in the image, with a box surrounding the object and a confidence score. Here is the original image on the left, with the predictions displayed on the right:
629
+
630
+ <h3 align="center">
631
+ <a><img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/coco_sample.png" width="400"></a>
632
+ <a><img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/coco_sample_post_processed.png" width="400"></a>
633
+ </h3>
634
+
635
+ You can learn more about the tasks supported by the `pipeline` API in [this tutorial](https://huggingface.co/docs/transformers/task_summary).
636
+
637
+ In addition to `pipeline`, to download and use any of the pretrained models on your given task, all it takes is three lines of code. Here is the PyTorch version:
638
+ ```python
639
+ >>> from transformers import AutoTokenizer, AutoModel
640
+
641
+ >>> tokenizer = AutoTokenizer.from_pretrained("google-bert/bert-base-uncased")
642
+ >>> model = AutoModel.from_pretrained("google-bert/bert-base-uncased")
643
+
644
+ >>> inputs = tokenizer("Hello world!", return_tensors="pt")
645
+ >>> outputs = model(**inputs)
646
+ ```
647
+
648
+ And here is the equivalent code for TensorFlow:
649
+ ```python
650
+ >>> from transformers import AutoTokenizer, TFAutoModel
651
+
652
+ >>> tokenizer = AutoTokenizer.from_pretrained("google-bert/bert-base-uncased")
653
+ >>> model = TFAutoModel.from_pretrained("google-bert/bert-base-uncased")
654
+
655
+ >>> inputs = tokenizer("Hello world!", return_tensors="tf")
656
+ >>> outputs = model(**inputs)
657
+ ```
658
+
659
+ The tokenizer is responsible for all the preprocessing the pretrained model expects and can be called directly on a single string (as in the above examples) or a list. It will output a dictionary that you can use in downstream code or simply directly pass to your model using the ** argument unpacking operator.
660
+
661
+ The model itself is a regular [Pytorch `nn.Module`](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) or a [TensorFlow `tf.keras.Model`](https://www.tensorflow.org/api_docs/python/tf/keras/Model) (depending on your backend) which you can use as usual. [This tutorial](https://huggingface.co/docs/transformers/training) explains how to integrate such a model into a classic PyTorch or TensorFlow training loop, or how to use our `Trainer` API to quickly fine-tune on a new dataset.
662
+
663
+ ## Why should I use transformers?
664
+
665
+ 1. Easy-to-use state-of-the-art models:
666
+ - High performance on natural language understanding & generation, computer vision, and audio tasks.
667
+ - Low barrier to entry for educators and practitioners.
668
+ - Few user-facing abstractions with just three classes to learn.
669
+ - A unified API for using all our pretrained models.
670
+
671
+ 1. Lower compute costs, smaller carbon footprint:
672
+ - Researchers can share trained models instead of always retraining.
673
+ - Practitioners can reduce compute time and production costs.
674
+ - Dozens of architectures with over 400,000 pretrained models across all modalities.
675
+
676
+ 1. Choose the right framework for every part of a model's lifetime:
677
+ - Train state-of-the-art models in 3 lines of code.
678
+ - Move a single model between TF2.0/PyTorch/JAX frameworks at will.
679
+ - Seamlessly pick the right framework for training, evaluation, and production.
680
+
681
+ 1. Easily customize a model or an example to your needs:
682
+ - We provide examples for each architecture to reproduce the results published by its original authors.
683
+ - Model internals are exposed as consistently as possible.
684
+ - Model files can be used independently of the library for quick experiments.
685
+
686
+ ## Why shouldn't I use transformers?
687
+
688
+ - This library is not a modular toolbox of building blocks for neural nets. The code in the model files is not refactored with additional abstractions on purpose, so that researchers can quickly iterate on each of the models without diving into additional abstractions/files.
689
+ - The training API is not intended to work on any model but is optimized to work with the models provided by the library. For generic machine learning loops, you should use another library (possibly, [Accelerate](https://huggingface.co/docs/accelerate)).
690
+ - While we strive to present as many use cases as possible, the scripts in our [examples folder](https://github.com/huggingface/transformers/tree/main/examples) are just that: examples. It is expected that they won't work out-of-the-box on your specific problem and that you will be required to change a few lines of code to adapt them to your needs.
691
+
692
+ ## Installation
693
+
694
+ ### With pip
695
+
696
+ This repository is tested on Python 3.8+, Flax 0.4.1+, PyTorch 1.11+, and TensorFlow 2.6+.
697
+
698
+ You should install 🤗 Transformers in a [virtual environment](https://docs.python.org/3/library/venv.html). If you're unfamiliar with Python virtual environments, check out the [user guide](https://packaging.python.org/guides/installing-using-pip-and-virtual-environments/).
699
+
700
+ First, create a virtual environment with the version of Python you're going to use and activate it.
701
+
702
+ Then, you will need to install at least one of Flax, PyTorch, or TensorFlow.
703
+ Please refer to [TensorFlow installation page](https://www.tensorflow.org/install/), [PyTorch installation page](https://pytorch.org/get-started/locally/#start-locally) and/or [Flax](https://github.com/google/flax#quick-install) and [Jax](https://github.com/google/jax#installation) installation pages regarding the specific installation command for your platform.
704
+
705
+ When one of those backends has been installed, 🤗 Transformers can be installed using pip as follows:
706
+
707
+ ```bash
708
+ pip install transformers
709
+ ```
710
+
711
+ If you'd like to play with the examples or need the bleeding edge of the code and can't wait for a new release, you must [install the library from source](https://huggingface.co/docs/transformers/installation#installing-from-source).
712
+
713
+ ### With conda
714
+
715
+ 🤗 Transformers can be installed using conda as follows:
716
+
717
+ ```shell script
718
+ conda install conda-forge::transformers
719
+ ```
720
+
721
+ > **_NOTE:_** Installing `transformers` from the `huggingface` channel is deprecated.
722
+
723
+ Follow the installation pages of Flax, PyTorch or TensorFlow to see how to install them with conda.
724
+
725
+ > **_NOTE:_** On Windows, you may be prompted to activate Developer Mode in order to benefit from caching. If this is not an option for you, please let us know in [this issue](https://github.com/huggingface/huggingface_hub/issues/1062).
726
+
727
+ ## Model architectures
728
+
729
+ **[All the model checkpoints](https://huggingface.co/models)** provided by 🤗 Transformers are seamlessly integrated from the huggingface.co [model hub](https://huggingface.co/models), where they are uploaded directly by [users](https://huggingface.co/users) and [organizations](https://huggingface.co/organizations).
730
+
731
+ Current number of checkpoints: ![](https://img.shields.io/endpoint?url=https://huggingface.co/api/shields/models&color=brightgreen)
732
+
733
+ 🤗 Transformers currently provides the following architectures: see [here](https://huggingface.co/docs/transformers/model_summary) for a high-level summary of each them.
734
+
735
+ To check if each model has an implementation in Flax, PyTorch or TensorFlow, or has an associated tokenizer backed by the 🤗 Tokenizers library, refer to [this table](https://huggingface.co/docs/transformers/index#supported-frameworks).
736
+
737
+ These implementations have been tested on several datasets (see the example scripts) and should match the performance of the original implementations. You can find more details on performance in the Examples section of the [documentation](https://github.com/huggingface/transformers/tree/main/examples).
738
+
739
+
740
+ ## Learn more
741
+
742
+ | Section | Description |
743
+ |-|-|
744
+ | [Documentation](https://huggingface.co/docs/transformers/) | Full API documentation and tutorials |
745
+ | [Task summary](https://huggingface.co/docs/transformers/task_summary) | Tasks supported by 🤗 Transformers |
746
+ | [Preprocessing tutorial](https://huggingface.co/docs/transformers/preprocessing) | Using the `Tokenizer` class to prepare data for the models |
747
+ | [Training and fine-tuning](https://huggingface.co/docs/transformers/training) | Using the models provided by 🤗 Transformers in a PyTorch/TensorFlow training loop and the `Trainer` API |
748
+ | [Quick tour: Fine-tuning/usage scripts](https://github.com/huggingface/transformers/tree/main/examples) | Example scripts for fine-tuning models on a wide range of tasks |
749
+ | [Model sharing and uploading](https://huggingface.co/docs/transformers/model_sharing) | Upload and share your fine-tuned models with the community |
750
+
751
+ ## Citation
752
+
753
+ We now have a [paper](https://www.aclweb.org/anthology/2020.emnlp-demos.6/) you can cite for the 🤗 Transformers library:
754
+ ```bibtex
755
+ @inproceedings{wolf-etal-2020-transformers,
756
+ title = "Transformers: State-of-the-Art Natural Language Processing",
757
+ author = "Thomas Wolf and Lysandre Debut and Victor Sanh and Julien Chaumond and Clement Delangue and Anthony Moi and Pierric Cistac and Tim Rault and Rémi Louf and Morgan Funtowicz and Joe Davison and Sam Shleifer and Patrick von Platen and Clara Ma and Yacine Jernite and Julien Plu and Canwen Xu and Teven Le Scao and Sylvain Gugger and Mariama Drame and Quentin Lhoest and Alexander M. Rush",
758
+ booktitle = "Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations",
759
+ month = oct,
760
+ year = "2020",
761
+ address = "Online",
762
+ publisher = "Association for Computational Linguistics",
763
+ url = "https://www.aclweb.org/anthology/2020.emnlp-demos.6",
764
+ pages = "38--45"
765
+ }
766
+ ```
lib/transformers/src/transformers.egg-info/SOURCES.txt ADDED
The diff for this file is too large to render. See raw diff
 
lib/transformers/src/transformers.egg-info/dependency_links.txt ADDED
@@ -0,0 +1 @@
 
 
1
+
lib/transformers/src/transformers.egg-info/entry_points.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ [console_scripts]
2
+ transformers-cli = transformers.commands.transformers_cli:main
lib/transformers/src/transformers.egg-info/not-zip-safe ADDED
@@ -0,0 +1 @@
 
 
1
+
lib/transformers/src/transformers.egg-info/requires.txt ADDED
@@ -0,0 +1,437 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ filelock
2
+ huggingface-hub<1.0,>=0.23.2
3
+ numpy>=1.17
4
+ packaging>=20.0
5
+ pyyaml>=5.1
6
+ regex!=2019.12.17
7
+ requests
8
+ tokenizers<0.20,>=0.19
9
+ safetensors>=0.4.1
10
+ tqdm>=4.27
11
+
12
+ [accelerate]
13
+ accelerate>=0.21.0
14
+
15
+ [agents]
16
+ diffusers
17
+ accelerate>=0.21.0
18
+ datasets!=2.5.0
19
+ torch
20
+ sentencepiece!=0.1.92,>=0.1.91
21
+ opencv-python
22
+ Pillow<=15.0,>=10.0.1
23
+
24
+ [all]
25
+ tensorflow<2.16,>2.9
26
+ onnxconverter-common
27
+ tf2onnx
28
+ tensorflow-text<2.16
29
+ keras-nlp<0.14.0,>=0.3.1
30
+ torch
31
+ accelerate>=0.21.0
32
+ jax<=0.4.13,>=0.4.1
33
+ jaxlib<=0.4.13,>=0.4.1
34
+ flax<=0.7.0,>=0.4.1
35
+ optax<=0.1.4,>=0.0.8
36
+ scipy<1.13.0
37
+ sentencepiece!=0.1.92,>=0.1.91
38
+ protobuf
39
+ tokenizers<0.20,>=0.19
40
+ torchaudio
41
+ librosa
42
+ pyctcdecode>=0.4.0
43
+ phonemizer
44
+ kenlm
45
+ Pillow<=15.0,>=10.0.1
46
+ optuna
47
+ ray[tune]>=2.7.0
48
+ sigopt
49
+ timm<=0.9.16
50
+ torchvision
51
+ codecarbon==1.2.0
52
+ decord==0.6.0
53
+ av==9.2.0
54
+
55
+ [audio]
56
+ librosa
57
+ pyctcdecode>=0.4.0
58
+ phonemizer
59
+ kenlm
60
+
61
+ [benchmark]
62
+ optimum-benchmark>=0.2.0
63
+
64
+ [codecarbon]
65
+ codecarbon==1.2.0
66
+
67
+ [deepspeed]
68
+ deepspeed>=0.9.3
69
+ accelerate>=0.21.0
70
+
71
+ [deepspeed-testing]
72
+ deepspeed>=0.9.3
73
+ accelerate>=0.21.0
74
+ pytest<8.0.0,>=7.2.0
75
+ pytest-rich
76
+ pytest-xdist
77
+ timeout-decorator
78
+ parameterized
79
+ psutil
80
+ datasets!=2.5.0
81
+ dill<0.3.5
82
+ evaluate>=0.2.0
83
+ pytest-timeout
84
+ ruff==0.4.4
85
+ sacrebleu<2.0.0,>=1.4.12
86
+ rouge-score!=0.0.7,!=0.0.8,!=0.1,!=0.1.1
87
+ nltk
88
+ GitPython<3.1.19
89
+ sacremoses
90
+ rjieba
91
+ beautifulsoup4
92
+ tensorboard
93
+ pydantic
94
+ sentencepiece!=0.1.92,>=0.1.91
95
+ faiss-cpu
96
+ cookiecutter==1.7.3
97
+ optuna
98
+ protobuf
99
+
100
+ [dev]
101
+ tensorflow<2.16,>2.9
102
+ onnxconverter-common
103
+ tf2onnx
104
+ tensorflow-text<2.16
105
+ keras-nlp<0.14.0,>=0.3.1
106
+ torch
107
+ accelerate>=0.21.0
108
+ jax<=0.4.13,>=0.4.1
109
+ jaxlib<=0.4.13,>=0.4.1
110
+ flax<=0.7.0,>=0.4.1
111
+ optax<=0.1.4,>=0.0.8
112
+ scipy<1.13.0
113
+ sentencepiece!=0.1.92,>=0.1.91
114
+ protobuf
115
+ tokenizers<0.20,>=0.19
116
+ torchaudio
117
+ librosa
118
+ pyctcdecode>=0.4.0
119
+ phonemizer
120
+ kenlm
121
+ Pillow<=15.0,>=10.0.1
122
+ optuna
123
+ ray[tune]>=2.7.0
124
+ sigopt
125
+ timm<=0.9.16
126
+ torchvision
127
+ codecarbon==1.2.0
128
+ decord==0.6.0
129
+ av==9.2.0
130
+ pytest<8.0.0,>=7.2.0
131
+ pytest-rich
132
+ pytest-xdist
133
+ timeout-decorator
134
+ parameterized
135
+ psutil
136
+ datasets!=2.5.0
137
+ dill<0.3.5
138
+ evaluate>=0.2.0
139
+ pytest-timeout
140
+ ruff==0.4.4
141
+ sacrebleu<2.0.0,>=1.4.12
142
+ rouge-score!=0.0.7,!=0.0.8,!=0.1,!=0.1.1
143
+ nltk
144
+ GitPython<3.1.19
145
+ sacremoses
146
+ rjieba
147
+ beautifulsoup4
148
+ tensorboard
149
+ pydantic
150
+ faiss-cpu
151
+ cookiecutter==1.7.3
152
+ isort>=5.5.4
153
+ urllib3<2.0.0
154
+ fugashi>=1.0
155
+ ipadic<2.0,>=1.0.0
156
+ unidic_lite>=1.0.7
157
+ unidic>=1.0.2
158
+ sudachipy>=0.6.6
159
+ sudachidict_core>=20220729
160
+ rhoknp<1.3.1,>=1.1.0
161
+ scikit-learn
162
+
163
+ [dev-tensorflow]
164
+ pytest<8.0.0,>=7.2.0
165
+ pytest-rich
166
+ pytest-xdist
167
+ timeout-decorator
168
+ parameterized
169
+ psutil
170
+ datasets!=2.5.0
171
+ dill<0.3.5
172
+ evaluate>=0.2.0
173
+ pytest-timeout
174
+ ruff==0.4.4
175
+ sacrebleu<2.0.0,>=1.4.12
176
+ rouge-score!=0.0.7,!=0.0.8,!=0.1,!=0.1.1
177
+ nltk
178
+ GitPython<3.1.19
179
+ sacremoses
180
+ rjieba
181
+ beautifulsoup4
182
+ tensorboard
183
+ pydantic
184
+ sentencepiece!=0.1.92,>=0.1.91
185
+ faiss-cpu
186
+ cookiecutter==1.7.3
187
+ tensorflow<2.16,>2.9
188
+ onnxconverter-common
189
+ tf2onnx
190
+ tensorflow-text<2.16
191
+ keras-nlp<0.14.0,>=0.3.1
192
+ protobuf
193
+ tokenizers<0.20,>=0.19
194
+ Pillow<=15.0,>=10.0.1
195
+ isort>=5.5.4
196
+ urllib3<2.0.0
197
+ scikit-learn
198
+ onnxruntime>=1.4.0
199
+ onnxruntime-tools>=1.4.2
200
+ librosa
201
+ pyctcdecode>=0.4.0
202
+ phonemizer
203
+ kenlm
204
+
205
+ [dev-torch]
206
+ pytest<8.0.0,>=7.2.0
207
+ pytest-rich
208
+ pytest-xdist
209
+ timeout-decorator
210
+ parameterized
211
+ psutil
212
+ datasets!=2.5.0
213
+ dill<0.3.5
214
+ evaluate>=0.2.0
215
+ pytest-timeout
216
+ ruff==0.4.4
217
+ sacrebleu<2.0.0,>=1.4.12
218
+ rouge-score!=0.0.7,!=0.0.8,!=0.1,!=0.1.1
219
+ nltk
220
+ GitPython<3.1.19
221
+ sacremoses
222
+ rjieba
223
+ beautifulsoup4
224
+ tensorboard
225
+ pydantic
226
+ sentencepiece!=0.1.92,>=0.1.91
227
+ faiss-cpu
228
+ cookiecutter==1.7.3
229
+ torch
230
+ accelerate>=0.21.0
231
+ protobuf
232
+ tokenizers<0.20,>=0.19
233
+ torchaudio
234
+ librosa
235
+ pyctcdecode>=0.4.0
236
+ phonemizer
237
+ kenlm
238
+ Pillow<=15.0,>=10.0.1
239
+ optuna
240
+ ray[tune]>=2.7.0
241
+ sigopt
242
+ timm<=0.9.16
243
+ torchvision
244
+ codecarbon==1.2.0
245
+ isort>=5.5.4
246
+ urllib3<2.0.0
247
+ fugashi>=1.0
248
+ ipadic<2.0,>=1.0.0
249
+ unidic_lite>=1.0.7
250
+ unidic>=1.0.2
251
+ sudachipy>=0.6.6
252
+ sudachidict_core>=20220729
253
+ rhoknp<1.3.1,>=1.1.0
254
+ scikit-learn
255
+ onnxruntime>=1.4.0
256
+ onnxruntime-tools>=1.4.2
257
+
258
+ [flax]
259
+ jax<=0.4.13,>=0.4.1
260
+ jaxlib<=0.4.13,>=0.4.1
261
+ flax<=0.7.0,>=0.4.1
262
+ optax<=0.1.4,>=0.0.8
263
+ scipy<1.13.0
264
+
265
+ [flax-speech]
266
+ librosa
267
+ pyctcdecode>=0.4.0
268
+ phonemizer
269
+ kenlm
270
+
271
+ [ftfy]
272
+ ftfy
273
+
274
+ [integrations]
275
+ optuna
276
+ ray[tune]>=2.7.0
277
+ sigopt
278
+
279
+ [ja]
280
+ fugashi>=1.0
281
+ ipadic<2.0,>=1.0.0
282
+ unidic_lite>=1.0.7
283
+ unidic>=1.0.2
284
+ sudachipy>=0.6.6
285
+ sudachidict_core>=20220729
286
+ rhoknp<1.3.1,>=1.1.0
287
+
288
+ [modelcreation]
289
+ cookiecutter==1.7.3
290
+
291
+ [natten]
292
+ natten<0.15.0,>=0.14.6
293
+
294
+ [onnx]
295
+ onnxconverter-common
296
+ tf2onnx
297
+ onnxruntime>=1.4.0
298
+ onnxruntime-tools>=1.4.2
299
+
300
+ [onnxruntime]
301
+ onnxruntime>=1.4.0
302
+ onnxruntime-tools>=1.4.2
303
+
304
+ [optuna]
305
+ optuna
306
+
307
+ [quality]
308
+ datasets!=2.5.0
309
+ isort>=5.5.4
310
+ ruff==0.4.4
311
+ GitPython<3.1.19
312
+ urllib3<2.0.0
313
+
314
+ [ray]
315
+ ray[tune]>=2.7.0
316
+
317
+ [retrieval]
318
+ faiss-cpu
319
+ datasets!=2.5.0
320
+
321
+ [ruff]
322
+ ruff==0.4.4
323
+
324
+ [sagemaker]
325
+ sagemaker>=2.31.0
326
+
327
+ [sentencepiece]
328
+ sentencepiece!=0.1.92,>=0.1.91
329
+ protobuf
330
+
331
+ [serving]
332
+ pydantic
333
+ uvicorn
334
+ fastapi
335
+ starlette
336
+
337
+ [sigopt]
338
+ sigopt
339
+
340
+ [sklearn]
341
+ scikit-learn
342
+
343
+ [speech]
344
+ torchaudio
345
+ librosa
346
+ pyctcdecode>=0.4.0
347
+ phonemizer
348
+ kenlm
349
+
350
+ [testing]
351
+ pytest<8.0.0,>=7.2.0
352
+ pytest-rich
353
+ pytest-xdist
354
+ timeout-decorator
355
+ parameterized
356
+ psutil
357
+ datasets!=2.5.0
358
+ dill<0.3.5
359
+ evaluate>=0.2.0
360
+ pytest-timeout
361
+ ruff==0.4.4
362
+ sacrebleu<2.0.0,>=1.4.12
363
+ rouge-score!=0.0.7,!=0.0.8,!=0.1,!=0.1.1
364
+ nltk
365
+ GitPython<3.1.19
366
+ sacremoses
367
+ rjieba
368
+ beautifulsoup4
369
+ tensorboard
370
+ pydantic
371
+ sentencepiece!=0.1.92,>=0.1.91
372
+ faiss-cpu
373
+ cookiecutter==1.7.3
374
+
375
+ [tf]
376
+ tensorflow<2.16,>2.9
377
+ onnxconverter-common
378
+ tf2onnx
379
+ tensorflow-text<2.16
380
+ keras-nlp<0.14.0,>=0.3.1
381
+
382
+ [tf-cpu]
383
+ keras<2.16,>2.9
384
+ tensorflow-cpu<2.16,>2.9
385
+ onnxconverter-common
386
+ tf2onnx
387
+ tensorflow-text<2.16
388
+ keras-nlp<0.14.0,>=0.3.1
389
+ tensorflow-probability<0.24
390
+
391
+ [tf-speech]
392
+ librosa
393
+ pyctcdecode>=0.4.0
394
+ phonemizer
395
+ kenlm
396
+
397
+ [timm]
398
+ timm<=0.9.16
399
+
400
+ [tokenizers]
401
+ tokenizers<0.20,>=0.19
402
+
403
+ [torch]
404
+ torch
405
+ accelerate>=0.21.0
406
+
407
+ [torch-speech]
408
+ torchaudio
409
+ librosa
410
+ pyctcdecode>=0.4.0
411
+ phonemizer
412
+ kenlm
413
+
414
+ [torch-vision]
415
+ torchvision
416
+ Pillow<=15.0,>=10.0.1
417
+
418
+ [torchhub]
419
+ filelock
420
+ huggingface-hub<1.0,>=0.23.2
421
+ importlib_metadata
422
+ numpy>=1.17
423
+ packaging>=20.0
424
+ protobuf
425
+ regex!=2019.12.17
426
+ requests
427
+ sentencepiece!=0.1.92,>=0.1.91
428
+ torch
429
+ tokenizers<0.20,>=0.19
430
+ tqdm>=4.27
431
+
432
+ [video]
433
+ decord==0.6.0
434
+ av==9.2.0
435
+
436
+ [vision]
437
+ Pillow<=15.0,>=10.0.1
lib/transformers/src/transformers.egg-info/top_level.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ transformers
lib/transformers/src/transformers/data/__pycache__/__init__.cpython-310.pyc ADDED
Binary file (1.1 kB). View file
 
lib/transformers/src/transformers/data/__pycache__/data_collator.cpython-310.pyc ADDED
Binary file (47.7 kB). View file
 
lib/transformers/src/transformers/data/metrics/__pycache__/__init__.cpython-310.pyc ADDED
Binary file (2.51 kB). View file
 
lib/transformers/src/transformers/data/metrics/__pycache__/squad_metrics.cpython-310.pyc ADDED
Binary file (16 kB). View file
 
lib/transformers/src/transformers/data/processors/__pycache__/__init__.cpython-310.pyc ADDED
Binary file (718 Bytes). View file
 
lib/transformers/src/transformers/data/processors/__pycache__/glue.cpython-310.pyc ADDED
Binary file (17.7 kB). View file
 
lib/transformers/src/transformers/data/processors/__pycache__/squad.cpython-310.pyc ADDED
Binary file (20 kB). View file
 
lib/transformers/src/transformers/data/processors/__pycache__/utils.cpython-310.pyc ADDED
Binary file (12.1 kB). View file
 
lib/transformers/src/transformers/data/processors/__pycache__/xnli.cpython-310.pyc ADDED
Binary file (2.5 kB). View file
 
lib/transformers/src/transformers/generation/__pycache__/__init__.cpython-310.pyc ADDED
Binary file (6.14 kB). View file
 
lib/transformers/src/transformers/generation/__pycache__/beam_constraints.cpython-310.pyc ADDED
Binary file (15.9 kB). View file
 
lib/transformers/src/transformers/generation/__pycache__/beam_search.cpython-310.pyc ADDED
Binary file (28.9 kB). View file
 
lib/transformers/src/transformers/generation/__pycache__/candidate_generator.cpython-310.pyc ADDED
Binary file (13 kB). View file
 
lib/transformers/src/transformers/generation/__pycache__/configuration_utils.cpython-310.pyc ADDED
Binary file (55.2 kB). View file
 
lib/transformers/src/transformers/generation/__pycache__/logits_process.cpython-310.pyc ADDED
Binary file (107 kB). View file
 
lib/transformers/src/transformers/generation/__pycache__/stopping_criteria.cpython-310.pyc ADDED
Binary file (24 kB). View file
 
lib/transformers/src/transformers/generation/__pycache__/utils.cpython-310.pyc ADDED
Binary file (128 kB). View file
 
lib/transformers/src/transformers/pipelines/__pycache__/audio_utils.cpython-310.pyc ADDED
Binary file (7.29 kB). View file
 
lib/transformers/src/transformers/pipelines/__pycache__/document_question_answering.cpython-310.pyc ADDED
Binary file (17 kB). View file
 
lib/transformers/src/transformers/pipelines/__pycache__/image_classification.cpython-310.pyc ADDED
Binary file (8.55 kB). View file
 
lib/transformers/src/transformers/pipelines/__pycache__/question_answering.cpython-310.pyc ADDED
Binary file (20.6 kB). View file
 
megatron_lm/megatron/__pycache__/global_vars.cpython-310.pyc ADDED
Binary file (2.21 kB). View file
 
megatron_lm/megatron/core/datasets/__pycache__/blended_dataset.cpython-310.pyc ADDED
Binary file (5.64 kB). View file
 
megatron_lm/megatron/core/datasets/__pycache__/blended_megatron_dataset_builder.cpython-310.pyc ADDED
Binary file (9.16 kB). View file
 
megatron_lm/megatron/core/datasets/__pycache__/blended_megatron_dataset_config.cpython-310.pyc ADDED
Binary file (6.1 kB). View file
 
megatron_lm/megatron/core/datasets/__pycache__/gpt_dataset.cpython-310.pyc ADDED
Binary file (13.4 kB). View file
 
megatron_lm/megatron/core/datasets/__pycache__/indexed_dataset.cpython-310.pyc ADDED
Binary file (19.2 kB). View file
 
megatron_lm/megatron/core/datasets/__pycache__/megatron_dataset.cpython-310.pyc ADDED
Binary file (4.74 kB). View file
 
megatron_lm/megatron/core/datasets/__pycache__/megatron_tokenizer.cpython-310.pyc ADDED
Binary file (4.47 kB). View file
 
megatron_lm/megatron/core/datasets/__pycache__/utils.cpython-310.pyc ADDED
Binary file (1.89 kB). View file
 
megatron_lm/megatron/core/datasets/helpers.cpython-310-x86_64-linux-gnu.so ADDED
Binary file (219 kB). View file
 
megatron_lm/megatron/data/__pycache__/data_samplers.cpython-310.pyc ADDED
Binary file (5.65 kB). View file
 
megatron_lm/megatron/tokenizer/__pycache__/__init__.cpython-310.pyc ADDED
Binary file (193 Bytes). View file
 
megatron_lm/megatron/tokenizer/__pycache__/tokenizer.cpython-310.pyc ADDED
Binary file (18 kB). View file
 
src/llama_recipes/__pycache__/arguments.cpython-310.pyc ADDED
Binary file (10.5 kB). View file
 
src/llama_recipes/__pycache__/finetuning.cpython-310.pyc ADDED
Binary file (6.93 kB). View file
 
src/llama_recipes/__pycache__/get_fsdp.cpython-310.pyc ADDED
Binary file (644 Bytes). View file
 
src/llama_recipes/__pycache__/get_model_decoder_layer.cpython-310.pyc ADDED
Binary file (1.27 kB). View file
 
src/llama_recipes/__pycache__/get_models.cpython-310.pyc ADDED
Binary file (2.65 kB). View file
 
src/llama_recipes/__pycache__/optimizer.cpython-310.pyc ADDED
Binary file (1.73 kB). View file
 
src/llama_recipes/datasets/__pycache__/pretrain_dataset.cpython-310.pyc ADDED
Binary file (2.47 kB). View file
 
src/llama_recipes/policies/__pycache__/__init__.cpython-310.pyc ADDED
Binary file (431 Bytes). View file
 
src/llama_recipes/policies/__pycache__/activation_checkpointing_functions.cpython-310.pyc ADDED
Binary file (1.18 kB). View file