Instructions to use farahbs/llama3.2_3b_instruct_simple_code_db with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use farahbs/llama3.2_3b_instruct_simple_code_db with Transformers:
# Load model directly from transformers import AutoModel model = AutoModel.from_pretrained("farahbs/llama3.2_3b_instruct_simple_code_db", dtype="auto") - Notebooks
- Google Colab
- Kaggle
| { | |
| "additional_special_tokens": [ | |
| { | |
| "content": "<|im_start|>", | |
| "lstrip": false, | |
| "normalized": false, | |
| "rstrip": false, | |
| "single_word": false | |
| }, | |
| { | |
| "content": "<|im_end|>", | |
| "lstrip": false, | |
| "normalized": false, | |
| "rstrip": false, | |
| "single_word": false | |
| } | |
| ], | |
| "bos_token": "<|im_start|>", | |
| "eos_token": "<|im_end|>", | |
| "pad_token": "<|im_end|>" | |
| } | |