Instructions to use AlexGostroot/malayalam-llama3-manglish with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- llama-cpp-python
How to use AlexGostroot/malayalam-llama3-manglish with llama-cpp-python:
# !pip install llama-cpp-python from llama_cpp import Llama llm = Llama.from_pretrained( repo_id="AlexGostroot/malayalam-llama3-manglish", filename="meta-llama-3.1-8b.Q4_K_M.gguf", )
output = llm( "Once upon a time,", max_tokens=512, echo=True ) print(output)
- Notebooks
- Google Colab
- Kaggle
- Local Apps
- llama.cpp
How to use AlexGostroot/malayalam-llama3-manglish with llama.cpp:
Install from brew
brew install llama.cpp # Start a local OpenAI-compatible server with a web UI: llama-server -hf AlexGostroot/malayalam-llama3-manglish:Q4_K_M # Run inference directly in the terminal: llama-cli -hf AlexGostroot/malayalam-llama3-manglish:Q4_K_M
Install from WinGet (Windows)
winget install llama.cpp # Start a local OpenAI-compatible server with a web UI: llama-server -hf AlexGostroot/malayalam-llama3-manglish:Q4_K_M # Run inference directly in the terminal: llama-cli -hf AlexGostroot/malayalam-llama3-manglish:Q4_K_M
Use pre-built binary
# Download pre-built binary from: # https://github.com/ggerganov/llama.cpp/releases # Start a local OpenAI-compatible server with a web UI: ./llama-server -hf AlexGostroot/malayalam-llama3-manglish:Q4_K_M # Run inference directly in the terminal: ./llama-cli -hf AlexGostroot/malayalam-llama3-manglish:Q4_K_M
Build from source code
git clone https://github.com/ggerganov/llama.cpp.git cd llama.cpp cmake -B build cmake --build build -j --target llama-server llama-cli # Start a local OpenAI-compatible server with a web UI: ./build/bin/llama-server -hf AlexGostroot/malayalam-llama3-manglish:Q4_K_M # Run inference directly in the terminal: ./build/bin/llama-cli -hf AlexGostroot/malayalam-llama3-manglish:Q4_K_M
Use Docker
docker model run hf.co/AlexGostroot/malayalam-llama3-manglish:Q4_K_M
- LM Studio
- Jan
- Ollama
How to use AlexGostroot/malayalam-llama3-manglish with Ollama:
ollama run hf.co/AlexGostroot/malayalam-llama3-manglish:Q4_K_M
- Unsloth Studio new
How to use AlexGostroot/malayalam-llama3-manglish with Unsloth Studio:
Install Unsloth Studio (macOS, Linux, WSL)
curl -fsSL https://unsloth.ai/install.sh | sh # Run unsloth studio unsloth studio -H 0.0.0.0 -p 8888 # Then open http://localhost:8888 in your browser # Search for AlexGostroot/malayalam-llama3-manglish to start chatting
Install Unsloth Studio (Windows)
irm https://unsloth.ai/install.ps1 | iex # Run unsloth studio unsloth studio -H 0.0.0.0 -p 8888 # Then open http://localhost:8888 in your browser # Search for AlexGostroot/malayalam-llama3-manglish to start chatting
Using HuggingFace Spaces for Unsloth
# No setup required # Open https://huggingface.co/spaces/unsloth/studio in your browser # Search for AlexGostroot/malayalam-llama3-manglish to start chatting
- Docker Model Runner
How to use AlexGostroot/malayalam-llama3-manglish with Docker Model Runner:
docker model run hf.co/AlexGostroot/malayalam-llama3-manglish:Q4_K_M
- Lemonade
How to use AlexGostroot/malayalam-llama3-manglish with Lemonade:
Pull the model
# Download Lemonade from https://lemonade-server.ai/ lemonade pull AlexGostroot/malayalam-llama3-manglish:Q4_K_M
Run and chat with the model
lemonade run user.malayalam-llama3-manglish-Q4_K_M
List all available models
lemonade list
malayalam-llama3-manglish : GGUF
This model is a fine-tuned version of Meta LLaMA 3.1 8B, optimized for Malayalam and Manglish (Malayalam written in English script) conversations.
It was fine-tuned using LoRA with Unsloth and converted to GGUF format for efficient local inference with llama.cpp.
๐ค Author
Ash
๐ง Model Overview
This model is designed to:
Understand Manglish inputs Generate natural Malayalam / Manglish responses Handle casual conversational dialogue Work efficiently on low-resource systems using GGUF ๐ Dataset Details
Dataset used: https://github.com/mhdashikofficial/Manglish-LLM-dataset
Dataset Characteristics: ~20,000 conversation samples Chat-style message format (system, user, assistant) Mix of Malayalam and Manglish Human-like conversational tone Example Format: { "messages": [ {"role": "system", "content": "..."}, {"role": "user", "content": "..."}, {"role": "assistant", "content": "..."} ] } โ๏ธ Training Details Base Model: Meta LLaMA 3.1 8B Fine-tuning Method: LoRA (Unsloth) Training Steps: 300 Sequence Length: 1024 Hardware: Google Colab (T4 GPU) โ๏ธ Quantization Format: GGUF Method: Q4_K_M Optimized for fast inference and low memory usage ๐ก Usage
Run with llama.cpp:
llama-cli -hf AlexGostroot/malayalam-llama3-manglish --jinja ๐ฆ Available Files meta-llama-3.1-8b.Q4_K_M.gguf โ ๏ธ Limitations Limited reasoning depth (low training steps) May mix Malayalam and English inconsistently Not suitable for complex or critical tasks ๐ Future Improvements Increase training steps (1000+) Add more native Malayalam data Improve response consistency Expand dataset diversity ๐งพ Notes
This model is intended for:
Experimentation Local AI applications Malayalam conversational systems
Trained and converted using Unsloth: https://github.com/unslothai/unsloth
- Downloads last month
- 87
4-bit