Removing Sub 500M models ? - What wind.surf will do ?
Browse files- app.py +0 -10
- models.lst +5 -0
app.py
CHANGED
@@ -76,16 +76,6 @@ MODELS = {
|
|
76 |
"repo_id": "Qwen/Qwen2.5-0.5B-Instruct-GGUF",
|
77 |
"filename": "*Q4_K_M.gguf",
|
78 |
"chat_format": "chatml"
|
79 |
-
},
|
80 |
-
"MobileLLM-350M": {
|
81 |
-
"repo_id": "pjh64/MobileLLM-350M-GGUF",
|
82 |
-
"filename": "*Q4_K_M.gguf",
|
83 |
-
"chat_format": "chatml"
|
84 |
-
},
|
85 |
-
"MobileLLM-125M": {
|
86 |
-
"repo_id": "pjh64/MobileLLM-125M-GGUF",
|
87 |
-
"filename": "*Q4_K_M.gguf",
|
88 |
-
"chat_format": "chatml"
|
89 |
}
|
90 |
}
|
91 |
|
|
|
76 |
"repo_id": "Qwen/Qwen2.5-0.5B-Instruct-GGUF",
|
77 |
"filename": "*Q4_K_M.gguf",
|
78 |
"chat_format": "chatml"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
79 |
}
|
80 |
}
|
81 |
|
models.lst
CHANGED
@@ -26,3 +26,8 @@ https://huggingface.co/pjh64/MobileLLM-350M-GGUF
|
|
26 |
|
27 |
https://huggingface.co/pjh64/MobileLLM-125M-GGUF
|
28 |
|
|
|
|
|
|
|
|
|
|
|
|
26 |
|
27 |
https://huggingface.co/pjh64/MobileLLM-125M-GGUF
|
28 |
|
29 |
+
https://huggingface.co/spaces/ggml-org/gguf-my-repo
|
30 |
+
ggml-org/gguf-my-repo
|
31 |
+
|
32 |
+
https://huggingface.co/flair/bueble-lm-2b
|
33 |
+
|