Export model 'openai/clip-vit-base-patch32', on 2025-01-29 03:23:33 JST
Browse files
README.md
CHANGED
@@ -2,6 +2,7 @@
|
|
2 |
pipeline_tag: zero-shot-classification
|
3 |
base_model:
|
4 |
- openai/clip-vit-base-patch16
|
|
|
5 |
language:
|
6 |
- en
|
7 |
tags:
|
@@ -16,9 +17,10 @@ ONNX exported version of CLIP models.
|
|
16 |
|
17 |
# Models
|
18 |
|
19 |
-
|
20 |
|
21 |
| Name | Image (Params/FLOPS) | Image Size | Image Width (Enc/Emb) | Text (Params/FLOPS) | Text Width (Enc/Emb) | Created At |
|
22 |
|:------------------------------------------------------------------------------------|:-----------------------|-------------:|:------------------------|:----------------------|:-----------------------|:-------------|
|
23 |
| [openai/clip-vit-base-patch16](https://huggingface.co/openai/clip-vit-base-patch16) | 85.6M / 16.9G | 224 | 768 / 512 | 37.8M / 529.2M | 512 / 512 | 2022-03-03 |
|
|
|
24 |
|
|
|
2 |
pipeline_tag: zero-shot-classification
|
3 |
base_model:
|
4 |
- openai/clip-vit-base-patch16
|
5 |
+
- openai/clip-vit-base-patch32
|
6 |
language:
|
7 |
- en
|
8 |
tags:
|
|
|
17 |
|
18 |
# Models
|
19 |
|
20 |
+
2 models exported in total.
|
21 |
|
22 |
| Name | Image (Params/FLOPS) | Image Size | Image Width (Enc/Emb) | Text (Params/FLOPS) | Text Width (Enc/Emb) | Created At |
|
23 |
|:------------------------------------------------------------------------------------|:-----------------------|-------------:|:------------------------|:----------------------|:-----------------------|:-------------|
|
24 |
| [openai/clip-vit-base-patch16](https://huggingface.co/openai/clip-vit-base-patch16) | 85.6M / 16.9G | 224 | 768 / 512 | 37.8M / 529.2M | 512 / 512 | 2022-03-03 |
|
25 |
+
| [openai/clip-vit-base-patch32](https://huggingface.co/openai/clip-vit-base-patch32) | 87.4M / 4.4G | 224 | 768 / 512 | 37.8M / 529.2M | 512 / 512 | 2022-03-03 |
|
26 |
|
models.parquet
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:654807b800475bff877b4b1ab225d20be8c7ff24f9996b85ec2bfaa338823528
|
3 |
+
size 8468
|
openai/clip-vit-base-patch32/image_encode.onnx
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:58773462a749e7122def90d16eb159c20dff1f40b210f40ba834161ddd076f2e
|
3 |
+
size 351650753
|
openai/clip-vit-base-patch32/meta.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4d227887f338e5137c56fd72343d1899e91073eafc6f4677ce925e8141b4dbc3
|
3 |
+
size 443
|
openai/clip-vit-base-patch32/preprocessor.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7205c8fb09a2889e5abac115d76f9a949626ef911417c18a848d7aa81f34820d
|
3 |
+
size 826
|
openai/clip-vit-base-patch32/text_encode.onnx
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2f07758ed6a9f05f8af341a80c486fcb547db499e8a5f5c79d8bc7e38d298156
|
3 |
+
size 253990000
|
openai/clip-vit-base-patch32/tokenizer.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:358e995b0cfd26ae9243ad4027fb5aa92bfa6c46ed22fe1adfd0fed53a9baeac
|
3 |
+
size 3642240
|