LiteMOE-3x460m
LiteMOE-3x460m is a merge of the following models using mergekit:
🧩 Configuration
base_model: ahxt/LiteLlama-460M-1T
gate_mode: random
dtype: bfloat16
experts:
- source_model: ahxt/LiteLlama-460M-1T
positive_prompts: [""]
- source_model: ahxt/LiteLlama-460M-1T
positive_prompts: [""]
- source_model: ahxt/LiteLlama-460M-1T
positive_prompts: [""]
Needs finetuning
- Downloads last month
- 12
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.