--- license: apache-2.0 library_name: transformers tags: - code - granite - mlx base_model: ibm-granite/granite-20b-code-base datasets: - bigcode/commitpackft - TIGER-Lab/MathInstruct - meta-math/MetaMathQA - glaiveai/glaive-code-assistant-v3 - glaive-function-calling-v2 - bugdaryan/sql-create-context-instruction - garage-bAInd/Open-Platypus - nvidia/HelpSteer metrics: - code_eval pipeline_tag: text-generation inference: true model-index: - name: granite-20b-code-instruct results: - task: type: text-generation dataset: name: HumanEvalSynthesis(Python) type: bigcode/humanevalpack metrics: - type: pass@1 value: 60.4 name: pass@1 - type: pass@1 value: 53.7 name: pass@1 - type: pass@1 value: 58.5 name: pass@1 - type: pass@1 value: 42.1 name: pass@1 - type: pass@1 value: 45.7 name: pass@1 - type: pass@1 value: 42.7 name: pass@1 - type: pass@1 value: 44.5 name: pass@1 - type: pass@1 value: 42.7 name: pass@1 - type: pass@1 value: 49.4 name: pass@1 - type: pass@1 value: 32.3 name: pass@1 - type: pass@1 value: 42.1 name: pass@1 - type: pass@1 value: 18.3 name: pass@1 - type: pass@1 value: 43.9 name: pass@1 - type: pass@1 value: 43.9 name: pass@1 - type: pass@1 value: 45.7 name: pass@1 - type: pass@1 value: 41.5 name: pass@1 - type: pass@1 value: 41.5 name: pass@1 - type: pass@1 value: 29.9 name: pass@1 --- # mlx-community/granite-20b-code-instruct-8bit The Model [mlx-community/granite-20b-code-instruct-8bit](https://huggingface.co/mlx-community/granite-20b-code-instruct-8bit) was converted to MLX format from [ibm-granite/granite-20b-code-instruct](https://huggingface.co/ibm-granite/granite-20b-code-instruct) using mlx-lm version **0.13.0**. ## Use with mlx ```bash pip install mlx-lm ``` ```python from mlx_lm import load, generate model, tokenizer = load("mlx-community/granite-20b-code-instruct-8bit") response = generate(model, tokenizer, prompt="hello", verbose=True) ```