File size: 2,757 Bytes
b1c6cc2
 
 
 
 
 
 
 
 
 
 
 
 
755f72a
b1c6cc2
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f8886c5
b1c6cc2
 
 
 
 
 
 
 
 
 
 
 
4ec18d4
 
 
b1c6cc2
f8886c5
 
 
 
b1c6cc2
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
---
license: llama2
model_type: llama
tags:
  - facebook
  - meta
  - pytorch
  - llama
  - llama-2
---

# GOAT-70B-STORYTELLING model

![GOAT-70B-STORYTELLING](https://assets.adapt.ws/files/20231117_ehznrqludevtapck.png)

GOAT-70B-STORYTELLING model is supervised finetuned (SFT) version of LLaMA 2 developed by GOAT.AI lab for story-writing. 


# GOAT-STORYTELLING-AGENT
The GOAT-70B-STORYTELLING model has been developed as an integral component within the GOAT-STORYTELLING-AGENT Framework. This framework is designed to facilitate the generation of high-quality, cohesive, and captivating narratives, including stories and books. It achieves this by utilizing inputs such as plot outlines, character profiles, their interrelationships, and other relevant details. Example is provided below.

# Model description
 - **Base Architecture:** LLaMA 2 70B 
 - **License:** llama2
 - **Context window length:** 4096 tokens

### Learn more

- **Blog:** TBA
- **Framework:** github.com/GOAT-STORYTELLING-AGENT (TBA)
## Uses

The main purpose of GOAT-70B-STORYTELLING is to generate books, novels, moviescripts and etc. as an agent in cope with our GOAT-STORYTELLING-AGENT Framework. It is specifically designed for storywriters.

## Usage


Usage can be either self-hosted via `transformers` or used with Spaces

```python
import torch

from transformers import AutoTokenizer, AutoModelForCausalLM

model_name = "GOAT-AI/GOAT-70B-STORYTELLING"

tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(
    model_name,
    torch_dtype=torch.bfloat16
)
```
Currently, we support LLM endpoint generation, where you need to send a post request to generation endpoint (we recommend using Text Generation Inference by HuggingFace)
First, modify config.py and add your generation endpoint.
Then you can use it inside via GOAT-STORYTELLING-AGENT framework:

```python
from goat_storytelling_agent.story_processor.prompt_manager import generate_story

novel_scenes = generate_story('never too much coffee', form='novel')
```
## License

GOAT-70B-STORYTELLING model is based on [Meta's LLaMA-2-70b-hf](https://huggingface.co/meta-llama/Llama-2-70b-hf), and using own datasets.  

GOAT-70B-STORYTELLING model weights are available under LLAMA-2 license. Note that the GOAT-70B-STORYTELLING model weights require access to the LLaMA-2 model weighs. The GOAT-70B-STORYTELLING model is based on LLaMA-2 and should be used according to the LLaMA-2 license. 

### Risks and Biases 

GOAT-70B-STORYTELLING model can produce factually incorrect output and should not be relied on to deliver factually accurate information. Therefore, the GOAT-70B-STORYTELLING model could possibly generate wrong, biased, or otherwise offensive outputs.