Spaces:
Runtime error
Runtime error
title: Chat Template Generation | |
emoji: 🔥 | |
colorFrom: red | |
colorTo: gray | |
sdk: streamlit | |
sdk_version: 1.28.2 | |
app_file: app.py | |
pinned: false | |
license: apache-2.0 | |
# Chat Template Generation: Make Chat Easier with Huggingface Tokenizer | |
![interface](./img/Interface.png) | |
This streamlit app is to serve as an easier way to check and push the chat template to your/exisiting huggingface repo | |
Current predefined templates: | |
- open-orca.jinja2 | |
- mistral-7b-openorca.jinja2 | |
- orca2.jinja2 | |
- airoboros_v2.jinja2 | |
- airoboros_v1.jinja2 | |
- openhermes-2.5-mistral.jinja2 | |
- zephyr.jinja2 | |
- falcon-chat.jinja2 | |
- chat-ml.jinja2 | |
- llama-2.jinja2 | |
- vicuna_v1.1.jinja2 | |
More templates will be predefined for easier setup of chat template. | |
With this interface you can easily | |
- update your tokenizer_config.json with chat-template attributes with predefined prompt template or custom template so that you can use `tokenizer.apply_chat_template(chat, tokenize=False, add_generation_prompt=True)` to generate chat prompt. | |
- contribute to existing model repository to support chat-template attributes | |
- review if your prompt template matches with your expected prompt | |
## To run locally | |
1. `python -m pip install -r requirements.txt` | |
2. `streamlit run app.py` | |
Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference | |