YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

ANAH: Analytical Annotation of Hallucinations in Large Language Models

arXiv license

This page holds the InternLM2-7B model which is trained with the ANAH dataset. It is fine-tuned to annotate the hallucination in LLM's responses.

More information please refer to our project page.

πŸ€— How to use the model

You have to follow the prompt in our paper to annotate the hallucination.

The models follow the conversation format of InternLM2-chat, with the template protocol as:

dict(role='user', begin='<|im_start|>user\n', end='<|im_end|>\n'),
dict(role='assistant', begin='<|im_start|>assistant\n', end='<|im_end|>\n'),

πŸ–ŠοΈ Citation

If you find this project useful in your research, please consider citing:

@article{ji2024anah,
  title={ANAH: Analytical Annotation of Hallucinations in Large Language Models},
  author={Ji, Ziwei and Gu, Yuzhe and Zhang, Wenwei and Lyu, Chengqi and Lin, Dahua and Chen, Kai},
  journal={arXiv preprint arXiv:2405.20315},
  year={2024}
}
Downloads last month
19
Safetensors
Model size
7.74B params
Tensor type
BF16
Β·
Inference Examples
Inference API (serverless) does not yet support model repos that contain custom code.

Model tree for opencompass/anah-7b

Quantizations
1 model