power-greg commited on
Commit
bdca489
·
1 Parent(s): debb252

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -7,7 +7,6 @@ sdk: static
7
  pinned: false
8
  ---
9
 
10
- ![Lamini Prompt vs Engine](promp-vs-engine.png "Lamini Prompt vs LLM Engine")
11
 
12
 
13
  # Introducing Lamini, the LLM Engine for Rapid Customization
@@ -18,6 +17,7 @@ Today, you can try out our open dataset generator for training instruction-follo
18
 
19
  [Sign up](https://lamini.ai/contact) for early access to our full LLM training module, including enterprise features like cloud prem deployments.
20
 
 
21
 
22
  # Training LLMs should be as easy as prompt-tuning 🦾
23
  Why is writing a prompt so easy, but training an LLM from a base model still so hard? Iteration cycles for finetuning on modest datasets are measured in months because it takes significant time to figure out why finetuned models fail. Conversely, prompt-tuning iterations are on the order of seconds, but performance plateaus in a matter of hours. Only a limited amount of data can be crammed into the prompt, not the terabytes of data in a warehouse.
 
7
  pinned: false
8
  ---
9
 
 
10
 
11
 
12
  # Introducing Lamini, the LLM Engine for Rapid Customization
 
17
 
18
  [Sign up](https://lamini.ai/contact) for early access to our full LLM training module, including enterprise features like cloud prem deployments.
19
 
20
+ ![Lamini Prompt vs Engine](promp-vs-engine.png "Lamini Prompt vs LLM Engine")
21
 
22
  # Training LLMs should be as easy as prompt-tuning 🦾
23
  Why is writing a prompt so easy, but training an LLM from a base model still so hard? Iteration cycles for finetuning on modest datasets are measured in months because it takes significant time to figure out why finetuned models fail. Conversely, prompt-tuning iterations are on the order of seconds, but performance plateaus in a matter of hours. Only a limited amount of data can be crammed into the prompt, not the terabytes of data in a warehouse.