Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,70 @@
|
|
1 |
-
---
|
2 |
-
license: apache-2.0
|
3 |
-
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
---
|
4 |
+
|
5 |
+
<p align="center">
|
6 |
+
<img src="https://github.com/ermongroup/TEOChat/blob/main/static/logo.png" width="150" style="margin-bottom: 0.2;"/>
|
7 |
+
<p>
|
8 |
+
<h2 align="center"> <a href="http://arxiv.org/abs/2410.06234">TEOChat: Large Language and Vision Assistant for Temporal Earth Observation Data</a></h2>
|
9 |
+
<h5 align="center"> If you like our project, please give us a star โญ on GitHub for latest update. </h2>
|
10 |
+
|
11 |
+
|
12 |
+
## ๐ฎ Highlights
|
13 |
+
|
14 |
+
**TEOChat** is the first language and vision assistant that can engage in conversation about sequences of temporal earth observation imagery, and exhibits impressive performance on multiple temporal instruction-following tasks.
|
15 |
+
|
16 |
+
### ๐ TEOChatlas: A new instruction-following dataset for temporal EO data
|
17 |
+
We introduce a new instruction-following dataset for temporal EO data called **TEOChatlas** which we use to train TEOChat. TEOChatlas contains 554,071 examples spanning dozens of temporal instruction-following tasks.
|
18 |
+
|
19 |
+
### ๐ค TEOChat: A new vision-language model for temporal EO data
|
20 |
+
We design TEOChat to use a LLaVA-style architecture, combining a temporally shared vision encoder with a LLaMA 2 LLM connected through an MLP vision-language projector
|
21 |
+
|
22 |
+
## ๐ค Demo
|
23 |
+
|
24 |
+
### Gradio Web UI
|
25 |
+
|
26 |
+
We provide an [online demo](https://huggingface.co/spaces/jirvin16/TEOChat) in Huggingface Spaces.
|
27 |
+
|
28 |
+
You can also run the demo locally by running the following command:
|
29 |
+
```bash
|
30 |
+
python videollava/serve/teochat_demo.py
|
31 |
+
```
|
32 |
+
|
33 |
+
|
34 |
+
## ๐ ๏ธ Requirements and Installation
|
35 |
+
* Python >= 3.9
|
36 |
+
* Pytorch == 2.2.1
|
37 |
+
* CUDA Version >= 12.1
|
38 |
+
* Install required packages:
|
39 |
+
```bash
|
40 |
+
git clone https://github.com/ermongroup/TEOChat.git
|
41 |
+
cd Video-LLaVA
|
42 |
+
conda create -n teochat python=3.9 -y
|
43 |
+
conda activate teochat
|
44 |
+
pip install --upgrade pip # enable PEP 660 support
|
45 |
+
pip install -r requirements.txt
|
46 |
+
```
|
47 |
+
|
48 |
+
## ๐๏ธ Training & Validating
|
49 |
+
The training & validating instructions are in [TRAIN_AND_VALIDATE.md](https://github.com/ermongroup/TEOChat/blob/main/TRAIN_AND_VALIDATE.md).
|
50 |
+
|
51 |
+
## ๐ Acknowledgement
|
52 |
+
* [Video-LLaVA](https://github.com/PKU-YuanGroup/Video-LLaVA) The codebase and model we built upon.
|
53 |
+
* [GeoChat](https://github.com/mbzuai-oryx/geochat) The single image instruction-following dataset we included in TEOChatlas.
|
54 |
+
|
55 |
+
## ๐ License
|
56 |
+
* The majority of this project is released under the Apache 2.0 license as found in the [LICENSE](https://github.com/ermongroup/TEOChat/blob/main/LICENSE) file.
|
57 |
+
* The service is a research preview intended for non-commercial use only, subject to the model [License](https://github.com/facebookresearch/llama/blob/main/MODEL_CARD.md) of LLaMA, [Terms of Use](https://openai.com/policies/terms-of-use) of the data generated by OpenAI, and [Privacy Practices](https://chrome.google.com/webstore/detail/sharegpt-share-your-chatg/daiacboceoaocpibfodeljbdfacokfjb) of ShareGPT. Please contact us if you find any potential violation.
|
58 |
+
|
59 |
+
## โ๏ธ Citation
|
60 |
+
If you find our paper and code useful in your research, please consider giving a star :star: and citation :pencil:.
|
61 |
+
|
62 |
+
```BibTeX
|
63 |
+
@article{irvin2024teochat,
|
64 |
+
title={TEOChat: A Large Vision-Language Assistant for Temporal Earth Observation Data},
|
65 |
+
author={Liu, Emily Ruoyu and Chen, Joyce Chuyi and Dormoy, Ines and Kim, Jinyoung and Khanna, Samar and Zheng, Zhuo and Ermon, Stefano},
|
66 |
+
journal={arXiv preprint arXiv:2410.06234},
|
67 |
+
year={2024}
|
68 |
+
}
|
69 |
+
```
|
70 |
+
|