llm / README.md
lemms's picture
Upload README.md with huggingface_hub
5ae897b verified

A newer version of the Gradio SDK is available: 5.44.1

Upgrade
metadata
title: OpenLLM Inference Space
emoji: ๐Ÿš€
colorFrom: blue
colorTo: purple
sdk: gradio
sdk_version: 4.44.1
app_file: app.py
pinned: false
license: gpl-3.0

๐Ÿš€ OpenLLM Inference Space

Welcome to the OpenLLM Inference Space! This is a comprehensive interface for running inference on our trained OpenLLM models with customizable parameters.

๐ŸŽฏ Available Models

We provide 7 different models trained for varying numbers of steps:

Model Training Steps Description Best Loss
4k Model 4,000 Early training stage, basic language patterns ~6.2
6k Model 6,000 Improved coherence, better vocabulary usage ~5.8
7k Model 7,000 Enhanced text generation quality ~5.5
8k Model 8,000 More sophisticated language understanding ~5.3
9k Model 9,000 Best performing model (latest training) ~5.2
10k Model 10,000 Latest extended training, maximum performance ~5.22
10k Improved 10,000 Improved training process, proper checkpoint format ~5.1774

๐ŸŽฎ How to Use

  1. Select a Model from the dropdown menu
  2. Load the Model to see its information
  3. Enter Your Prompt in the text box
  4. Adjust Parameters (temperature, max length, etc.)
  5. Generate Text and see the results!

โš™๏ธ Parameters

  • Temperature: Controls randomness (0.1-2.0)
  • Max Length: Number of tokens to generate (10-500)
  • Top-K: Limits to top-k most likely tokens (1-100)
  • Top-P: Nucleus sampling threshold (0.1-1.0)

๐Ÿง  Model Architecture

  • Model Size: Small (35.8M parameters)
  • Layers: 6 transformer layers
  • Embedding: 512 dimensions
  • Vocabulary: 32,000 tokens (SentencePiece)
  • Context Length: 1,024 tokens

OpenLLM Inference Space - Experience the power of open-source language models! ๐Ÿš€