--- language: - en license: apache-2.0 tags: - mlx datasets: - cerebras/SlimPajama-627B - bigcode/starcoderdata - HuggingFaceH4/ultrachat_200k - HuggingFaceH4/ultrafeedback_binarized widget: - text: '<|system|> You are a chatbot who can help code! <|user|> Write me a function to calculate the first 10 digits of the fibonacci sequence in Python and print it out to the CLI. <|assistant|> ' --- # mlx-community/TinyLlama-1.1B-Chat-v1.0-mlx This model was converted to MLX format from [`TinyLlama/TinyLlama-1.1B-Chat-v1.0`](). Refer to the [original model card](https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v1.0) for more details on the model. ## Use with mlx ```bash pip install mlx-lm ``` ```python from mlx_lm import load, generate model, tokenizer = load("mlx-community/TinyLlama-1.1B-Chat-v1.0-mlx") response = generate(model, tokenizer, prompt="hello", verbose=True) ```