File size: 4,315 Bytes
89c4ca6
 
89839a5
 
 
 
 
4b70a99
4cfc657
 
89c4ca6
 
89839a5
89c4ca6
 
 
 
89839a5
89c4ca6
7f36ff2
89839a5
 
 
 
 
4b70a99
89839a5
 
 
 
 
89c4ca6
 
89839a5
 
 
 
 
 
 
 
 
 
 
89c4ca6
 
89839a5
 
 
 
 
89c4ca6
 
 
89839a5
 
 
 
89c4ca6
89839a5
 
 
 
89c4ca6
 
89839a5
 
 
89c4ca6
 
 
 
89839a5
 
89c4ca6
 
89839a5
 
 
 
 
 
 
 
 
 
 
 
5254c0c
89839a5
 
 
 
4b1ac70
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
89839a5
4b1ac70
89839a5
 
 
 
5254c0c
 
 
4cfc657
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
---
library_name: transformers
tags:
- medical-qa
- healthcare
- llama
- fine-tuned
license: llama3.2
datasets:
- ruslanmv/ai-medical-chatbot
---

# Model Card: Llama-3.2-3B-Chat-Doctor

## Model Details

### Model Description
Llama-3.2-3B-Chat-Doctor is a specialized medical question-answering model based on the Llama 3.2 3B architecture. This model has been fine-tuned specifically for providing accurate and helpful responses to medical-related queries.

- **Developed by:** Ellbendl Satria
- **Model type:** Language Model (Conversational AI)
- **Language:** English
- **Base Model:** Meta Llama-3.2-3B-Instruct
- **Model Size:** 3 Billion Parameters
- **Specialization:** Medical Question Answering
- **License:** llama3.2

### Model Capabilities
- Provides informative responses to medical questions
- Assists in understanding medical terminology and health-related concepts
- Offers preliminary medical information (not a substitute for professional medical advice)

### Direct Use
This model can be used for:
- Providing general medical information
- Explaining medical conditions and symptoms
- Offering basic health-related guidance
- Supporting medical education and patient communication

### Limitations and Important Disclaimers
⚠️ **CRITICAL WARNINGS:**
- **NOT A MEDICAL PROFESSIONAL:** This model is NOT a substitute for professional medical advice, diagnosis, or treatment.
- Always consult a qualified healthcare provider for medical concerns.
- The model's responses should be treated as informational only and not as medical recommendations.

### Out-of-Scope Use
The model SHOULD NOT be used for:
- Providing emergency medical advice
- Diagnosing specific medical conditions
- Replacing professional medical consultation
- Making critical healthcare decisions

## Bias, Risks, and Limitations

### Potential Biases
- May reflect biases present in the training data
- Responses might not account for individual patient variations
- Limited by the comprehensiveness of the training dataset

### Technical Limitations
- Accuracy is limited to the knowledge in the training data
- May not capture the most recent medical research or developments
- Cannot perform physical examinations or medical tests

### Recommendations
- Always verify medical information with professional healthcare providers
- Use the model as a supplementary information source
- Be aware of potential inaccuracies or incomplete information

## Training Details

### Training Data
- **Source Dataset:** [ruslanmv/ai-medical-chatbot](https://huggingface.co/datasets/ruslanmv/ai-medical-chatbot)
- **Base Model:** [Meta Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct)

### Training Procedure
[Provide details about the fine-tuning process, if available]
- Fine-tuning approach
- Computational resources used
- Training duration
- Specific techniques applied during fine-tuning

## How to Use the Model

### Hugging Face Transformers
```python
from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "Ellbendls/llama-3.2-3b-chat-doctor"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)

# Example usage
input_text = "I had a surgery which ended up with some failures. What can I do to fix it?"

# Prepare inputs with explicit padding and attention mask
inputs = tokenizer(input_text, return_tensors="pt", padding=True, truncation=True)

# Generate response with more explicit parameters
outputs = model.generate(
    input_ids=inputs['input_ids'], 
    attention_mask=inputs['attention_mask'],
    max_new_tokens=150,  # Specify max new tokens to generate
    do_sample=True,      # Enable sampling for more diverse responses
    temperature=0.7,     # Control randomness of output
    top_p=0.9,           # Nucleus sampling to maintain quality
    num_return_sequences=1  # Number of generated sequences
)

# Decode the generated response
response = tokenizer.decode(outputs[0], skip_special_tokens=True)

print(response)
```

### Ethical Considerations
This model is developed with the intent to provide helpful, accurate, and responsible medical information. Users are encouraged to:
- Use the model responsibly
- Understand its limitations
- Seek professional medical advice for serious health concerns