jirvin16 commited on
Commit
0c22592
ยท
verified ยท
1 Parent(s): 77ceac8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +70 -3
README.md CHANGED
@@ -1,3 +1,70 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ ---
4
+
5
+ <p align="center">
6
+ <img src="https://github.com/ermongroup/TEOChat/blob/main/static/logo.png" width="150" style="margin-bottom: 0.2;"/>
7
+ <p>
8
+ <h2 align="center"> <a href="http://arxiv.org/abs/2410.06234">TEOChat: Large Language and Vision Assistant for Temporal Earth Observation Data</a></h2>
9
+ <h5 align="center"> If you like our project, please give us a star โญ on GitHub for latest update. </h2>
10
+
11
+
12
+ ## ๐Ÿ˜ฎ Highlights
13
+
14
+ **TEOChat** is the first language and vision assistant that can engage in conversation about sequences of temporal earth observation imagery, and exhibits impressive performance on multiple temporal instruction-following tasks.
15
+
16
+ ### ๐Ÿ“š TEOChatlas: A new instruction-following dataset for temporal EO data
17
+ We introduce a new instruction-following dataset for temporal EO data called **TEOChatlas** which we use to train TEOChat. TEOChatlas contains 554,071 examples spanning dozens of temporal instruction-following tasks.
18
+
19
+ ### ๐Ÿค– TEOChat: A new vision-language model for temporal EO data
20
+ We design TEOChat to use a LLaVA-style architecture, combining a temporally shared vision encoder with a LLaMA 2 LLM connected through an MLP vision-language projector
21
+
22
+ ## ๐Ÿค— Demo
23
+
24
+ ### Gradio Web UI
25
+
26
+ We provide an [online demo](https://huggingface.co/spaces/jirvin16/TEOChat) in Huggingface Spaces.
27
+
28
+ You can also run the demo locally by running the following command:
29
+ ```bash
30
+ python videollava/serve/teochat_demo.py
31
+ ```
32
+
33
+
34
+ ## ๐Ÿ› ๏ธ Requirements and Installation
35
+ * Python >= 3.9
36
+ * Pytorch == 2.2.1
37
+ * CUDA Version >= 12.1
38
+ * Install required packages:
39
+ ```bash
40
+ git clone https://github.com/ermongroup/TEOChat.git
41
+ cd Video-LLaVA
42
+ conda create -n teochat python=3.9 -y
43
+ conda activate teochat
44
+ pip install --upgrade pip # enable PEP 660 support
45
+ pip install -r requirements.txt
46
+ ```
47
+
48
+ ## ๐Ÿ—๏ธ Training & Validating
49
+ The training & validating instructions are in [TRAIN_AND_VALIDATE.md](https://github.com/ermongroup/TEOChat/blob/main/TRAIN_AND_VALIDATE.md).
50
+
51
+ ## ๐Ÿ‘ Acknowledgement
52
+ * [Video-LLaVA](https://github.com/PKU-YuanGroup/Video-LLaVA) The codebase and model we built upon.
53
+ * [GeoChat](https://github.com/mbzuai-oryx/geochat) The single image instruction-following dataset we included in TEOChatlas.
54
+
55
+ ## ๐Ÿ”’ License
56
+ * The majority of this project is released under the Apache 2.0 license as found in the [LICENSE](https://github.com/ermongroup/TEOChat/blob/main/LICENSE) file.
57
+ * The service is a research preview intended for non-commercial use only, subject to the model [License](https://github.com/facebookresearch/llama/blob/main/MODEL_CARD.md) of LLaMA, [Terms of Use](https://openai.com/policies/terms-of-use) of the data generated by OpenAI, and [Privacy Practices](https://chrome.google.com/webstore/detail/sharegpt-share-your-chatg/daiacboceoaocpibfodeljbdfacokfjb) of ShareGPT. Please contact us if you find any potential violation.
58
+
59
+ ## โœ๏ธ Citation
60
+ If you find our paper and code useful in your research, please consider giving a star :star: and citation :pencil:.
61
+
62
+ ```BibTeX
63
+ @article{irvin2024teochat,
64
+ title={TEOChat: A Large Vision-Language Assistant for Temporal Earth Observation Data},
65
+ author={Liu, Emily Ruoyu and Chen, Joyce Chuyi and Dormoy, Ines and Kim, Jinyoung and Khanna, Samar and Zheng, Zhuo and Ermon, Stefano},
66
+ journal={arXiv preprint arXiv:2410.06234},
67
+ year={2024}
68
+ }
69
+ ```
70
+