draganjovanovich commited on
Commit
7b966b8
โ€ข
1 Parent(s): e7859ca

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +62 -0
README.md ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ - sr
6
+ - hr
7
+ - bs
8
+ datasets:
9
+ - teknium/OpenHermes-2.5
10
+ - WizardLM/WizardLM_evol_instruct_V2_196k
11
+ - draganjovanovich/airoboros-3.0-serbian
12
+ - Open-Orca/SlimOrca
13
+ - draganjovanovich/prodigy-instruct-reason-function-calling
14
+ - iskonai/chatislav-instruct
15
+ - mlabonne/FineTome-100k
16
+ ---
17
+ # Prodigy SM Instruct v0.1
18
+
19
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/617bbeec14572ebe9e6ea83f/5sl1RHE054rNkJMA4Z0dp.png" width="70%" height="70%">
20
+
21
+ **Prodigy SM Instruct v0.1** is an instruction-tuned version of our [Prodigy SM Base v0.1](https://huggingface.co/iskonai/prodigy-sm-base-v0.1) model, optimized for instruction following in Serbian, Croatian, Bosnian and English languages. The model demonstrates exceptional capabilities in function calling and tool usage while maintaining strong performance across all supported languages.
22
+
23
+ Model was shared as part of the presenatation at DSC Europe 2024 / Belgrade (BREAKOUT SESSIONS & WORKSHOPS).
24
+
25
+ # ๐ŸŽฎ DEMO TIME! ๐Ÿš€ For next couple of weeks you can test model at: [https://chat.iskon.ai](https://chat.iskon.ai)
26
+ or you can download it and use **SGLang**(recomended as it have superior structured output sampling), vLLM, HF text-generation-inference etc.
27
+ proper model card coming soon when we upload final checkpoint, and remove **draft** from the name :)
28
+
29
+ # Training
30
+
31
+ The model was instruction-tuned using a carefully curated mix of high-quality datasets in ChatML format.
32
+
33
+ # Features
34
+ - Precise **system prompt** following capabilities in **Serbian** (trained on really large system prompts: 500+ tokens.)
35
+ - Strong instruction following capabilities in Serbian, Croatian, Bosnian and English
36
+ - Enhanced function calling and tool usage abilities
37
+ - Maintains the base model's strong performance in all supported languages
38
+ - Uses ChatML format for conversations
39
+
40
+ # Use Cases
41
+
42
+ The model is particularly well-suited for:
43
+ - Multilingual applications requiring Serbian, Croatian, Bosnian and English support
44
+ - Applications requiring precise instruction following
45
+ - Systems utilizing function calling and tool usage
46
+ - Conversational AI applications
47
+ - Text generation in supported languages
48
+
49
+ # Limitations
50
+
51
+ As with all language models:
52
+ - Outputs should be verified for critical applications
53
+ - Performance in languages other than Serbian, Croatian, Bosnian and English may vary
54
+ - Users should be aware of general LLM limitations regarding hallucination and factual accuracy
55
+
56
+ # What's Next
57
+
58
+ Stay tuned for our upcoming releases:
59
+ - [prodigy-sm-base (llama-3.1)]() **COMING SOON**
60
+ - [prodigy-sm-instruct (llama-3.1)]() **COMING SOON**
61
+
62
+ ๐Ÿ“ข We're excited to announce that [iskon.ai](https://Iskon.ai) will soon launch an API platform featuring advanced **Prodigy** series of models, advanced AI tools and much more! ๐Ÿš€