README / README.md
Tijmen2's picture
Update README.md
3211d2d verified
# AstroMLab
AstroMLab is a diverse group of researchers dedicated to advancing the application of Large Language Models (LLMs) in astronomy. Our team includes:
- Leading astronomers, astrophysicists, and cosmologists.
- Natural language processing experts.
- Frontier arXivists from the NASA Astrophysics Data System
## Objectives
- Develop specialized LLMs for astronomy
- Create open-source models for advanced research
- Facilitate LLM-driven end-to-end agentic research in astronomy
## Current Work
Our ongoing projects include:
- Curation of an astronomy-based benchmarking dataset
- Development of specialized astronomy LLMs
- Performance evaluation of models on astronomical tasks
## Models and Performance
We have developed several models, including AstroSage-8B, AstroLLaMA-2-70B, and AstroLLaMA-3-8B. Our AstroSage-8B model has demonstrated strong performance in astronomy Q&A tasks ([Ting et al. 2024](https://arxiv.org/abs/2407.11194), Pan et al. 2024):
| Model | Score (%) |
|-------|-----------|
| **AstroSage-8B (AstroMLab)** | **80.9** |
| **AstroLLaMA-2-70B-Base (AstroMLab)** | **76.0** |
| LLaMA-3.1-8B | 73.7 |
| Phi-3.5-4B | 72.8 |
| Gemma-2-9B | 71.5 |
| LLaMA-2-70B | 70.7 |
| Qwen-2.5-7B | 70.4 |
| Yi-1.5-9B | 68.4 |
| InternLM-2.5-7B | 64.5 |
| Mistral-7B-v0.3 | 63.9 |
| ChatGLM3-6B | 50.4 |
| AstroLLaMA-2-7B (UniverseTBD) | 44.3 |
AstroSage-8B, our lightweight model, currently achieves the highest score among the ~7B parameter models in its astronomy knowledge recall ability.
![Cost and performance trade-off in astronomical Q&A](https://cdn-uploads.huggingface.co/production/uploads/643f1ddce2ea47d170103537/ip0Bk-LZRrCArimets4H7.png)
## Support and Resources
Our research benefits from:
- Access to the Frontier nodes at Oak Ridge Leadership Computing Facility
- Support from Microsoft's Accelerating Foundation Models Research (AFMR) program
## Contact
For inquiries or collaboration opportunities, please contact: [email protected]