Text Generation
Transformers
Safetensors
English
llama
climate
conversational
text-generation-inference
Inference Endpoints
dthulke commited on
Commit
32a05f1
1 Parent(s): d1fc2d9

adds chat template and update README

Browse files
Files changed (2) hide show
  1. README.md +8 -2
  2. tokenizer_config.json +2 -1
README.md CHANGED
@@ -12,6 +12,9 @@ co2_eq_emissions:
12
  training_type: "pre-training"
13
  geographical_location: "Washington, USA"
14
  hardware_used: "8x NVIDIA H100 HBM"
 
 
 
15
  ---
16
  # ClimateGPT-13B
17
 
@@ -32,7 +35,7 @@ The model is designed to be used together with retrieval augmentation to extend
32
  - **Context length:** 4K tokens
33
  - **Input:** Text-only data
34
  - **Output:** Model generates text only
35
- - **Paper:** [Download](https://shareddatastgacct.blob.core.windows.net/shared-data/climategpt-v1-publication.pdf)
36
  - **Website:** [eci.io](https://eci.io)
37
 
38
  ## Uses
@@ -71,7 +74,7 @@ The model was trained using ChatML so the following format should be followed wh
71
 
72
  ## Evaluation
73
 
74
- Detailed evaluation results are presented in our [paper](https://shareddatastgacct.blob.core.windows.net/shared-data/climategpt-v1-publication.pdf) on our model card website: [eci.io/model-card](https://eci.io/model-card)
75
 
76
  ## Environmental Impact
77
  - **Hardware Type:** 8x NVIDIA H100 HBM
@@ -91,5 +94,8 @@ If you find ClimateGPT is useful in your work, please cite it with:
91
  title={ClimateGPT: Towards AI Synthesizing Interdisciplinary Research on Climate Change},
92
  author={David Thulke and Yingbo Gao and Petrus Pelser and Rein Brune and Rricha Jalota and Floris Fok and Michael Ramos and Ian van Wyk and Abdallah Nasir and Hayden Goldstein and Taylor Tragemann and Katie Nguyen and Ariana Fowler and Andrew Stanco and Jon Gabriel and Jordan Taylor and Dean Moro and Evgenii Tsymbalov and Juliette de Waal and Evgeny Matusov and Mudar Yaghi and Mohammad Shihadah and Hermann Ney and Christian Dugast and Jonathan Dotan and Daniel Erasmus},
93
  year={2024},
 
 
 
94
  }
95
  ```
 
12
  training_type: "pre-training"
13
  geographical_location: "Washington, USA"
14
  hardware_used: "8x NVIDIA H100 HBM"
15
+ license: other
16
+ license_name: climategpt-community-license
17
+ license_link: LICENSE.txt
18
  ---
19
  # ClimateGPT-13B
20
 
 
35
  - **Context length:** 4K tokens
36
  - **Input:** Text-only data
37
  - **Output:** Model generates text only
38
+ - **Paper:** [arXiv:2401.09646](https://arxiv.org/abs/2401.09646)
39
  - **Website:** [eci.io](https://eci.io)
40
 
41
  ## Uses
 
74
 
75
  ## Evaluation
76
 
77
+ Detailed evaluation results are presented in our [paper](https://arxiv.org/abs/2401.09646) on our model card website: [eci.io/model-card](https://eci.io/model-card)
78
 
79
  ## Environmental Impact
80
  - **Hardware Type:** 8x NVIDIA H100 HBM
 
94
  title={ClimateGPT: Towards AI Synthesizing Interdisciplinary Research on Climate Change},
95
  author={David Thulke and Yingbo Gao and Petrus Pelser and Rein Brune and Rricha Jalota and Floris Fok and Michael Ramos and Ian van Wyk and Abdallah Nasir and Hayden Goldstein and Taylor Tragemann and Katie Nguyen and Ariana Fowler and Andrew Stanco and Jon Gabriel and Jordan Taylor and Dean Moro and Evgenii Tsymbalov and Juliette de Waal and Evgeny Matusov and Mudar Yaghi and Mohammad Shihadah and Hermann Ney and Christian Dugast and Jonathan Dotan and Daniel Erasmus},
96
  year={2024},
97
+ eprint={2401.09646},
98
+ archivePrefix={arXiv},
99
+ primaryClass={cs.LG}
100
  }
101
  ```
tokenizer_config.json CHANGED
@@ -145,5 +145,6 @@
145
  "spaces_between_special_tokens": false,
146
  "tokenizer_class": "LlamaTokenizer",
147
  "unk_token": "<unk>",
148
- "use_default_system_prompt": false
 
149
  }
 
145
  "spaces_between_special_tokens": false,
146
  "tokenizer_class": "LlamaTokenizer",
147
  "unk_token": "<unk>",
148
+ "use_default_system_prompt": false,
149
+ "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}"
150
  }