Update README.md
Browse files
README.md
CHANGED
@@ -15,7 +15,7 @@ Retrieval](https://arxiv.org/pdf/2104.08860.pdf) by Lou et el, and implemented i
|
|
15 |
|
16 |
The training process involved 150,000 videos obtained from the [WebVid Dataset](https://m-bain.github.io/webvid-dataset/), a comprehensive collection of short videos with corresponding textual descriptions sourced from the web.
|
17 |
|
18 |
-
To adapt the clip model obtained during training
|
19 |
|
20 |
### Use with Transformers
|
21 |
### Extracting Text Embeddings:
|
@@ -45,11 +45,10 @@ For video embedding there is an extra [notebook](https://huggingface.co/Diangle/
|
|
45 |
|
46 |
## Model Intended Use
|
47 |
|
48 |
-
This model is intended to use for video
|
49 |
|
50 |
-
## Performance and Limitations
|
51 |
|
52 |
-
|
53 |
|
54 |
We have evaluated the performance of differenet models on the last 10k video clips from Webvid database.
|
55 |
|
|
|
15 |
|
16 |
The training process involved 150,000 videos obtained from the [WebVid Dataset](https://m-bain.github.io/webvid-dataset/), a comprehensive collection of short videos with corresponding textual descriptions sourced from the web.
|
17 |
|
18 |
+
To adapt the clip model obtained during training into the implementation of [clip-vit-base-patch32](https://huggingface.co/openai/clip-vit-base-patch32), we have made modifications to the weights.
|
19 |
|
20 |
### Use with Transformers
|
21 |
### Extracting Text Embeddings:
|
|
|
45 |
|
46 |
## Model Intended Use
|
47 |
|
48 |
+
This model is intended to use for video retrieval, look for example this [**SPACE**](https://huggingface.co/spaces/Diangle/Clip4Clip-webvid).
|
49 |
|
|
|
50 |
|
51 |
+
## Performance
|
52 |
|
53 |
We have evaluated the performance of differenet models on the last 10k video clips from Webvid database.
|
54 |
|