language:
- en
tags:
- physics
- astronomy
- astrophysics
- cosmology
license:
- llama3.1
base_model:
- meta-llama/Meta-Llama-3.1-8B
library_name: transformers
AstroSage-Llama-3.1-8B
AstroSage-Llama-3.1-8B is a domain-specialized natural-language AI assistant tailored for research in astronomy, astrophysics, and cosmology. Trained on the complete collection of astronomy-related arXiv papers from 2007-2024 along with millions of synthetically-generated question-answer pairs and other astronomical literature, AstroSage-Llama-3.1-8B demonstrates excellent proficiency on a wide range of questions. AstroSage-Llama-3.1-8B scores 80.9% on the AstroMLab-1 benchmark, greatly outperforming all models---proprietary and open-weight---in the 8-billion parameter class, and performing on par with GPT-4o. This achievement demonstrates the potential of domain specialization in AI, suggesting that focused training can yield capabilities exceeding those of much larger, general-purpose models. AstroSage-Llama-3.1-8B is freely available, enabling widespread access to advanced AI capabilities for astronomical education and research.
Model Details
- Model Type: Domain-specialized LLM
- Base Model: Meta-Llama-3.1-8B
- Parameters: 8 billion
- Training Focus: Astronomy, Astrophysics, Cosmology, and Astronomical Instrumentation
- License: Llama 3.1 Community License
- Development Process:
- Continued Pre-training (CPT) on astronomical literature
- Supervised Fine-tuning (SFT) on QA pairs and instruction sets
- Model merging with Meta-Llama-3.1-8B-Instruct (75% CPT+SFT / 25% Meta-Instruct)
Performance
- AstroMLab-1 Benchmark: 80.9% accuracy
- Outperforms all 8B parameter models
- Comparable to GPT-4o (80.4%)
- ~1000x more cost-effective than proprietary models
- 8 percentage-point improvement over base Llama-3.1-8b model on Astronomy Q&A benchmark
- General Capabilities: Maintains strong performance on standard benchmarks
- IF-EVAL: 41.4%
- BBH: 52.9%
- MATH: 8.4%
- GPQA: 31.2%
- MUSR: 38.9%
- MMLU-PRO: 34.6%
Training Data
- Continued Pre-training:
- ~250,000 arXiv preprints (2007-2024) from astro-ph and gr-qc
- Astronomy-related Wikipedia articles
- Selected astronomy textbooks
- Total: 3.3 billion tokens, 19.9 GB plaintext
- Supervised Fine-tuning:
- 8.8 million curated QA pairs
- Filtered Infinity-Instruct-7M dataset
- Paper summaries and metadata
- Total: 2.0 billion tokens, 9.8 GB plaintext
Intended Use
- Curiosity-driven question answering
- Brainstorming new ideas
- Astronomical research assistance
- Educational support in astronomy
- Literature review and summarization
- Scientific explanation of concepts
Limitations
- Training data cutoff: January 2024
- As with all LLMs, hallucinations are possible
- Limited by 8B parameter size for complex reasoning
- Paper metadata not perfectly memorized
- Performance primarily validated on multiple-choice questions
- Primarily trained for use in English
Ethical Considerations
- Should not be used as sole source for critical research decisions
- Output should be verified against primary sources
- May reflect biases present in astronomical literature
Technical Specifications
- Architecture: Based on Meta-Llama 3.1
- Training Infrastructure: ORNL OLCF Frontier
- Hosting: Hugging Face Hub (AstroMLab/AstroSage-8B)
Citation and Contact
- Contract: Corresponding author Tijmen de Haan, email: tijmen dot dehaan at gmail dot com and AstroMLab astromachinelearninglab at gmail dot com
- Please cite the AstroMLab 3 paper when referencing to this model.