Model Card for LLaMat-2-Chat

LLaMat-2-Chat is a specialized large language model designed to serve as a copilot for materials research. Finetuned from LLaMat-2, this model is adapted for tasks such as information extraction from material science text and tabular data.


Overview

  • Model Type: Large Language Model (LLM)
  • Base Model: LLaMat-2 (continued pretraining of LLaMA-2 on material science data)
  • Language: English
  • License: LLaMA-3 License
  • Tags: Material Science, Domain Adaptation, Table Understanding, Scientific Data Parsing, Materials Copilot

Model Details

Key Features

  • Instruction Following Abilities: Optimized for understanding and processing instructions in the material science domain.
  • Domain-Specific Expertise: Pretrained on material science tokens, enabling high performance in scientific applications.
  • Applications: information extraction, table understanding, and parsing data for research tasks.

Development and Support

  • Developed by: M3RG, IIT Delhi & DAIR, IIT Delhi
  • Compute Support:
    • Edinburgh International Data Facility (EIDF): Provided access to Cerebras CS2 clusters for pretraining.
    • IIT Delhi High-Performance Computing Cluster: Supported fine-tuning and inference stages.

Technical Specifications

Hardware Infrastructure

  • Pretraining: 2 Cerebras CS-2 Wafer-Scale Engines (WSE-2)
  • Finetuning: 8 NVIDIA A100 80GB GPUs
  • Inferencing: 1 NVIDIA A100 80GB GPU

Software Stack

  • Frameworks: PyTorch, Hugging Face Transformers

Model Sources

Downloads last month
8
Safetensors
Model size
6.74B params
Tensor type
F32
·
Inference API
Unable to determine this model's library. Check the docs .

Model tree for m3rg-iitd/llamat-2-chat

Finetuned
m3rg-iitd/llamat-2
Finetuned
(2)
this model

Collection including m3rg-iitd/llamat-2-chat