A newer version of the Gradio SDK is available:
5.44.1
metadata
title: OpenLLM Inference Space
emoji: ๐
colorFrom: blue
colorTo: purple
sdk: gradio
sdk_version: 4.44.1
app_file: app.py
pinned: false
license: gpl-3.0
๐ OpenLLM Inference Space
Welcome to the OpenLLM Inference Space! This is a comprehensive interface for running inference on our trained OpenLLM models with customizable parameters.
๐ฏ Available Models
We provide 7 different models trained for varying numbers of steps:
Model | Training Steps | Description | Best Loss |
---|---|---|---|
4k Model | 4,000 | Early training stage, basic language patterns | ~6.2 |
6k Model | 6,000 | Improved coherence, better vocabulary usage | ~5.8 |
7k Model | 7,000 | Enhanced text generation quality | ~5.5 |
8k Model | 8,000 | More sophisticated language understanding | ~5.3 |
9k Model | 9,000 | Best performing model (latest training) | ~5.2 |
10k Model | 10,000 | Latest extended training, maximum performance | ~5.22 |
10k Improved | 10,000 | Improved training process, proper checkpoint format | ~5.1774 |
๐ฎ How to Use
- Select a Model from the dropdown menu
- Load the Model to see its information
- Enter Your Prompt in the text box
- Adjust Parameters (temperature, max length, etc.)
- Generate Text and see the results!
โ๏ธ Parameters
- Temperature: Controls randomness (0.1-2.0)
- Max Length: Number of tokens to generate (10-500)
- Top-K: Limits to top-k most likely tokens (1-100)
- Top-P: Nucleus sampling threshold (0.1-1.0)
๐ง Model Architecture
- Model Size: Small (35.8M parameters)
- Layers: 6 transformer layers
- Embedding: 512 dimensions
- Vocabulary: 32,000 tokens (SentencePiece)
- Context Length: 1,024 tokens
OpenLLM Inference Space - Experience the power of open-source language models! ๐