AstroMLab

non-profit

AI & ML interests

AstroMLab is a collaborative initiative of astronomers and AI experts dedicated to advancing Large Language Models in astronomy. Our goal is to expedite scientific discovery through LLM-driven research.

AstroMLab

AstroMLab is a diverse group of researchers dedicated to advancing the application of Large Language Models (LLMs) in astronomy. Our team includes:

  • Leading astronomers, astrophysicists, and cosmologists.
  • Natural language processing experts.
  • Frontier arXivists from the NASA Astrophysics Data System

Objectives

  • Develop specialized LLMs for astronomy
  • Create open-source models for advanced research
  • Facilitate LLM-driven end-to-end agentic research in astronomy

Current Work

Our ongoing projects include:

  • Curation of an astronomy-based benchmarking dataset
  • Development of specialized astronomy LLMs
  • Performance evaluation of models on astronomical tasks

Models and Performance

We have developed several models, including AstroSage-8B, AstroLLaMA-2-70B, and AstroLLaMA-3-8B. Our AstroSage-8B model has demonstrated strong performance in astronomy Q&A tasks (Ting et al. 2024, Pan et al. 2024):

Model Score (%)
AstroSage-8B (AstroMLab) 79.1
AstroLLaMA-2-70B-Base (AstroMLab) 76.0
LLaMA-3.1-8B 73.7
Phi-3.5-4B 72.8
Gemma-2-9B 71.5
LLaMA-2-70B 70.7
Qwen-2.5-7B 70.4
Yi-1.5-9B 68.4
InternLM-2.5-7B 64.5
Mistral-7B-v0.3 63.9
ChatGLM3-6B 50.4
AstroLLaMA-2-7B (UniverseTBD) 44.3

AstroSage-8B, our lightweight model, currently achieves the highest score among the ~7B parameter models in its astronomy knowledge recall ability.

Cost and performance trade-off in astronomical Q&A

Support and Resources

Our research benefits from:

  • Access to the Frontier nodes at Oak Ridge Leadership Computing Facility
  • Support from Microsoft's Accelerating Foundation Models Research (AFMR) program

Contact

For inquiries or collaboration opportunities, please contact: astromachinelearninglab@gmail.com

datasets

None public yet