MMedS-Llama-3-8B / README.md
Henrychur's picture
Update README.md
a2cab41 verified
metadata
license: llama3
datasets:
  - Henrychur/MMedC
  - Henrychur/MedS-Ins
language:
  - en
base_model: Henrychur/MMedS-Llama-3-8B
tags:
  - medical
library_name: transformers

MMedS-Llama3

💻Github Repo 🖨️arXiv Paper

The official codes for "Towards Evaluating and Building Versatile Large Language Models for Medicine"

Introduction

This repository hosts MMedS-Llama-3-8B. Its foundation model, MMed-Llama-3-8B, is a multilingual medical language model which has undergone additional continuous pretraining on MMedC. Furthermore, the model has been fine-tuned under supervision using MedS-Ins, a comprehensive dataset designed specifically for supervised fine-tuning (SFT), featuring 13.5 million samples across 122 tasks. For more details, please refer to our paper.

Usage

The model can be loaded as follows:

import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("Henrychur/MMed-Llama-3-8B-EnIns")
model = AutoModelForCausalLM.from_pretrained("Henrychur/MMed-Llama-3-8B-EnIns", torch_dtype=torch.float16)
  • Inference format is the same as Llama 3, you can check the inference code here.