Spaces:
Running
on
Zero
Running
on
Zero
File size: 1,495 Bytes
3eb01b6 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 |
from transformers import (
GlmForCausalLM,
GPT2TokenizerFast,
GraniteForCausalLM,
LlamaForCausalLM,
LlamaTokenizerFast,
PreTrainedTokenizerFast,
Qwen2ForCausalLM,
Qwen2TokenizerFast,
StableLmForCausalLM,
)
CONFIG = {
"01-ai/Yi-Coder-1.5B-Chat": {
"model": LlamaForCausalLM,
"tokenizer": LlamaTokenizerFast,
},
"HuggingFaceTB/SmolLM2-135M-Instruct": {
"model": LlamaForCausalLM,
"tokenizer": GPT2TokenizerFast,
},
"HuggingFaceTB/SmolLM2-360M-Instruct": {
"model": LlamaForCausalLM,
"tokenizer": GPT2TokenizerFast,
},
"HuggingFaceTB/SmolLM2-1.7B-Instruct": {
"model": LlamaForCausalLM,
"tokenizer": GPT2TokenizerFast,
},
"ibm-granite/granite-3.0-2b-instruct": {
"model": GraniteForCausalLM,
"tokenizer": GPT2TokenizerFast,
},
"Qwen/Qwen2.5-0.5B-Instruct": {
"model": Qwen2ForCausalLM,
"tokenizer": Qwen2TokenizerFast,
},
"Qwen/Qwen2.5-1.5B-Instruct": {
"model": Qwen2ForCausalLM,
"tokenizer": Qwen2TokenizerFast,
},
"Qwen/Qwen2.5-Coder-1.5B-Instruct": {
"model": Qwen2ForCausalLM,
"tokenizer": Qwen2TokenizerFast,
},
"stabilityai/stablelm-2-zephyr-1_6b": {
"model": StableLmForCausalLM,
"tokenizer": GPT2TokenizerFast,
},
"THUDM/glm-edge-1.5b-chat": {
"model": GlmForCausalLM,
"tokenizer": PreTrainedTokenizerFast,
},
}
|