ai-forever
commited on
Commit
•
42a0887
1
Parent(s):
8135a8f
Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,63 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# ruclip-vit-large-patch14-336
|
2 |
+
|
3 |
+
**RuCLIP** (**Ru**ssian **C**ontrastive **L**anguage–**I**mage **P**retraining) is a multimodal model
|
4 |
+
for obtaining images and text similarities and rearranging captions and pictures.
|
5 |
+
RuCLIP builds on a large body of work on zero-shot transfer, computer vision, natural language processing and
|
6 |
+
multimodal learning.
|
7 |
+
|
8 |
+
Model was trained by [Sber AI](https://github.com/sberbank-ai) and [SberDevices](https://sberdevices.ru/) teams.
|
9 |
+
* Task: `text ranking`; `image ranking`; `zero-shot image classification`;
|
10 |
+
* Type: `encoder`
|
11 |
+
* Num Parameters: `430M`
|
12 |
+
* Training Data Volume: `240 million text-image pairs`
|
13 |
+
* Language: `Russian`
|
14 |
+
* Context Length: `77`
|
15 |
+
* Transformer Layers: `12`
|
16 |
+
* Transformer Width: `768`
|
17 |
+
* Transformer Heads: `12`
|
18 |
+
* Image Size: `336`
|
19 |
+
* Vision Layers: `24`
|
20 |
+
* Vision Width: `1024`
|
21 |
+
* Vision Patch Size: `14`
|
22 |
+
|
23 |
+
## Usage [Github](https://github.com/sberbank-ai/ru-clip)
|
24 |
+
|
25 |
+
```
|
26 |
+
pip install ruclip
|
27 |
+
```
|
28 |
+
|
29 |
+
```python
|
30 |
+
clip, processor = ruclip.load("ruclip-vit-large-patch14-336", device="cuda")
|
31 |
+
```
|
32 |
+
|
33 |
+
## Performance
|
34 |
+
We have evaluated the performance on the following datasets:
|
35 |
+
|
36 |
+
| Dataset | Metric Name | Metric Result |
|
37 |
+
|:--------------|:---------------|:--------------------|
|
38 |
+
| Food101 | acc | 0.712 |
|
39 |
+
| CIFAR10 | acc | 0.906 |
|
40 |
+
| CIFAR100 | acc | 0.591 |
|
41 |
+
| Birdsnap | acc | 0.213 |
|
42 |
+
| SUN397 | acc | 0.523 |
|
43 |
+
| Stanford Cars | acc | 0.659 |
|
44 |
+
| DTD | acc | 0.408 |
|
45 |
+
| MNIST | acc | 0.242 |
|
46 |
+
| STL10 | acc | 0.956 |
|
47 |
+
| PCam | acc | 0.554 |
|
48 |
+
| CLEVR | acc | 0.142 |
|
49 |
+
| Rendered SST2 | acc | 0.539 |
|
50 |
+
| ImageNet | acc | 0.488 |
|
51 |
+
| FGVC Aircraft | mean-per-class | 0.075 |
|
52 |
+
| Oxford Pets | mean-per-class | 0.546 |
|
53 |
+
| Caltech101 | mean-per-class | 0.835 |
|
54 |
+
| Flowers102 | mean-per-class | 0.517 |
|
55 |
+
| HatefulMemes | roc-auc | 0.519 |
|
56 |
+
|
57 |
+
|
58 |
+
# Authors
|
59 |
+
|
60 |
+
+ Alex Shonenkov: [Github](https://github.com/shonenkov), [Kaggle GM](https://www.kaggle.com/shonenkov)
|
61 |
+
+ Daniil Chesakov: [Github](https://github.com/Danyache)
|
62 |
+
+ Denis Dimitrov: [Github](https://github.com/denndimitrov)
|
63 |
+
+ Igor Pavlov: [Github](https://github.com/boomb0om)
|