HoangHa commited on
Commit
bb861f1
·
verified ·
1 Parent(s): 2bb0007

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -9
README.md CHANGED
@@ -3,15 +3,14 @@ license: mit
3
  ---
4
  ## Overview
5
 
6
- **DeepSeek** developed and released the [DeepSeek R1 Distill Qwen 1.5B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B) model, a distilled version of the Qwen 1.5B language model. It is fine-tuned for high-performance text generation and optimized for dialogue and information-seeking tasks. This model achieves a balance of efficiency and accuracy while maintaining a smaller footprint compared to the original Qwen 1.5B.
7
-
8
- The model is designed for applications in customer support, conversational AI, and research, prioritizing both helpfulness and safety.
9
 
 
10
  ## Variants
11
 
12
  | No | Variant | Cortex CLI command |
13
  | --- | --- | --- |
14
- | 1 | [gguf](https://huggingface.co/cortexso/deepseek-r1-distill-qwen-1.5b/tree/main) | `cortex run deepseek-r1-distill-qwen-1.5b` |
15
 
16
 
17
  ## Use it with Jan (UI)
@@ -19,18 +18,18 @@ The model is designed for applications in customer support, conversational AI, a
19
  1. Install **Jan** using [Quickstart](https://jan.ai/docs/quickstart)
20
  2. Use in Jan model Hub:
21
  ```text
22
- cortexso/deepseek-r1-distill-qwen-1.5b
23
  ```
24
  ## Use it with Cortex (CLI)
25
 
26
  1. Install **Cortex** using [Quickstart](https://cortex.jan.ai/docs/quickstart)
27
  2. Run the model with command:
28
  ```bash
29
- cortex run deepseek-r1-distill-qwen-1.5b
30
  ```
31
  ## Credits
32
 
33
- - **Author:** DeepSeek
34
  - **Converter:** [Homebrew](https://www.homebrew.ltd/)
35
- - **Original License:** [License](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B#7-license)
36
- - **Papers:** [DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning](https://arxiv.org/html/2501.12948v1)
 
3
  ---
4
  ## Overview
5
 
6
+ This model is a supervised fine-tuned (SFT) version of homebrewltd/Ichigo-llama3.1-s-base-v0.3, trained on over 1 billion tokens from the Instruction Speech WhisperVQ v4 dataset. This dataset builds upon Instruction Speech WhisperVQ v3, incorporating multi-turn speech conversations and advanced noise rejection capabilities.
 
 
7
 
8
+ As a result, the Ichigo-llama3s models offer enhanced robustness to noisy environmental inputs and improved handling of multi-turn conversations, ensuring greater reliability and performance in real-world applications.
9
  ## Variants
10
 
11
  | No | Variant | Cortex CLI command |
12
  | --- | --- | --- |
13
+ | 1 | [gguf](https://huggingface.co/cortexso/ichigo-0.5/tree/main) | `cortex run deepseek-r1-distill-qwen-1.5b` |
14
 
15
 
16
  ## Use it with Jan (UI)
 
18
  1. Install **Jan** using [Quickstart](https://jan.ai/docs/quickstart)
19
  2. Use in Jan model Hub:
20
  ```text
21
+ cortexso/ichigo-0.5
22
  ```
23
  ## Use it with Cortex (CLI)
24
 
25
  1. Install **Cortex** using [Quickstart](https://cortex.jan.ai/docs/quickstart)
26
  2. Run the model with command:
27
  ```bash
28
+ cortex run ichigo-0.5
29
  ```
30
  ## Credits
31
 
32
+ - **Author:** HomebrewResearch
33
  - **Converter:** [Homebrew](https://www.homebrew.ltd/)
34
+ - **Original License:** [License]()
35
+ - **Papers:** [Papers]()