File size: 422 Bytes
c246a29
 
 
 
 
 
 
 
 
960565b
c246a29
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
{
  "model": "0x9/pho-7b-llama2",
  "base_model": "",
  "revision": "main",
  "private": false,
  "precision": "bfloat16",
  "params": 7.242,
  "architectures": "LlamaForCausalLM",
  "weight_type": "Original",
  "status": "FAILED",
  "submitted_time": "2024-03-11T14:52:31Z",
  "model_type": "\ud83d\udd36 : fine-tuned on domain-specific datasets",
  "job_id": "2443908",
  "job_start_time": "2024-03-11T15:06:27.833222"
}