munish0838's picture
Upload README.md with huggingface_hub
3f68638 verified
|
raw
history blame
No virus
1.73 kB
---
license: afl-3.0
tags:
- medical
---
![](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)
# QuantFactory/ClinicalGPT-base-zh-GGUF
This is quantized version of [medicalai/ClinicalGPT-base-zh](https://huggingface.co/medicalai/ClinicalGPT-base-zh) created using llama.cpp
# Original Model Card
# ClinicalGPT
This model card introduces ClinicalGPT model, a large language model designed and optimized for clinical scenarios. ClinicalGPT is fine-tuned on extensive and diverse medical datasets, including medical records, domain-specific knowledge, and multi-round dialogue consultations. The model is undergoing ongoing and continuous updates.
## Model Fine-tuning
We set the learning rate to 5e-5, with a batch size of 128 and a maximum length of 1,024, training across 3 epochs.
## How to use the model
Load the model via the transformers library:
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("medicalai/ClinicalGPT-base-zh")
model = AutoModelForCausalLM.from_pretrained("medicalai/ClinicalGPT-base-zh")
```
## Limitations
The project is intended for research purposes only and restricted from commercial or clinical use. The generated content by the model is subject to factors such as model computations, randomness, misinterpretation, and biases, and this project cannot guarantee its accuracy. This project assumes no legal liability for any content produced by the model. Users are advised to exercise caution and independently verify the generated results.