Spaces:
Running
Running
# Extensions | |
Extensions are defined by files named `script.py` inside subfolders of `text-generation-webui/extensions`. They are loaded at startup if the folder name is specified after the `--extensions` flag. | |
For instance, `extensions/silero_tts/script.py` gets loaded with `python server.py --extensions silero_tts`. | |
## [text-generation-webui-extensions](https://github.com/oobabooga/text-generation-webui-extensions) | |
The repository above contains a directory of user extensions. | |
If you create an extension, you are welcome to host it in a GitHub repository and submit a PR adding it to the list. | |
## Built-in extensions | |
|Extension|Description| | |
|---------|-----------| | |
|[openai](https://github.com/oobabooga/text-generation-webui/tree/main/extensions/openai)| Creates an API that mimics the OpenAI API and can be used as a drop-in replacement. | | |
|[multimodal](https://github.com/oobabooga/text-generation-webui/tree/main/extensions/multimodal) | Adds multimodality support (text+images). For a detailed description see [README.md](https://github.com/oobabooga/text-generation-webui/tree/main/extensions/multimodal/README.md) in the extension directory. | | |
|[google_translate](https://github.com/oobabooga/text-generation-webui/tree/main/extensions/google_translate)| Automatically translates inputs and outputs using Google Translate.| | |
|[silero_tts](https://github.com/oobabooga/text-generation-webui/tree/main/extensions/silero_tts)| Text-to-speech extension using [Silero](https://github.com/snakers4/silero-models). When used in chat mode, responses are replaced with an audio widget. | | |
|[whisper_stt](https://github.com/oobabooga/text-generation-webui/tree/main/extensions/whisper_stt)| Allows you to enter your inputs in chat mode using your microphone. | | |
|[sd_api_pictures](https://github.com/oobabooga/text-generation-webui/tree/main/extensions/sd_api_pictures)| Allows you to request pictures from the bot in chat mode, which will be generated using the AUTOMATIC1111 Stable Diffusion API. See examples [here](https://github.com/oobabooga/text-generation-webui/pull/309). | | |
|[character_bias](https://github.com/oobabooga/text-generation-webui/tree/main/extensions/character_bias)| Just a very simple example that adds a hidden string at the beginning of the bot's reply in chat mode. | | |
|[send_pictures](https://github.com/oobabooga/text-generation-webui/blob/main/extensions/send_pictures/)| Creates an image upload field that can be used to send images to the bot in chat mode. Captions are automatically generated using BLIP. | | |
|[gallery](https://github.com/oobabooga/text-generation-webui/blob/main/extensions/gallery/)| Creates a gallery with the chat characters and their pictures. | | |
|[superbooga](https://github.com/oobabooga/text-generation-webui/tree/main/extensions/superbooga)| An extension that uses ChromaDB to create an arbitrarily large pseudocontext, taking as input text files, URLs, or pasted text. Based on https://github.com/kaiokendev/superbig. | | |
|[ngrok](https://github.com/oobabooga/text-generation-webui/tree/main/extensions/ngrok)| Allows you to access the web UI remotely using the ngrok reverse tunnel service (free). It's an alternative to the built-in Gradio `--share` feature. | | |
|[perplexity_colors](https://github.com/oobabooga/text-generation-webui/tree/main/extensions/perplexity_colors)| Colors each token in the output text by its associated probability, as derived from the model logits. | | |
## How to write an extension | |
The extensions framework is based on special functions and variables that you can define in `script.py`. The functions are the following: | |
| Function | Description | | |
|-------------|-------------| | |
| `def setup()` | Is executed when the extension gets imported. | | |
| `def ui()` | Creates custom gradio elements when the UI is launched. | | |
| `def custom_css()` | Returns custom CSS as a string. It is applied whenever the web UI is loaded. | | |
| `def custom_js()` | Same as above but for javascript. | | |
| `def input_modifier(string, state, is_chat=False)` | Modifies the input string before it enters the model. In chat mode, it is applied to the user message. Otherwise, it is applied to the entire prompt. | | |
| `def output_modifier(string, state, is_chat=False)` | Modifies the output string before it is presented in the UI. In chat mode, it is applied to the bot's reply. Otherwise, it is applied to the entire output. | | |
| `def chat_input_modifier(text, visible_text, state)` | Modifies both the visible and internal inputs in chat mode. Can be used to hijack the chat input with custom content. | | |
| `def bot_prefix_modifier(string, state)` | Applied in chat mode to the prefix for the bot's reply. | | |
| `def state_modifier(state)` | Modifies the dictionary containing the UI input parameters before it is used by the text generation functions. | | |
| `def history_modifier(history)` | Modifies the chat history before the text generation in chat mode begins. | | |
| `def custom_generate_reply(...)` | Overrides the main text generation function. | | |
| `def custom_generate_chat_prompt(...)` | Overrides the prompt generator in chat mode. | | |
| `def tokenizer_modifier(state, prompt, input_ids, input_embeds)` | Modifies the `input_ids`/`input_embeds` fed to the model. Should return `prompt`, `input_ids`, `input_embeds`. See the `multimodal` extension for an example. | | |
| `def custom_tokenized_length(prompt)` | Used in conjunction with `tokenizer_modifier`, returns the length in tokens of `prompt`. See the `multimodal` extension for an example. | | |
Additionally, you can define a special `params` dictionary. In it, the `display_name` key is used to define the displayed name of the extension in the UI, and the `is_tab` key is used to define whether the extension should appear in a new tab. By default, extensions appear at the bottom of the "Text generation" tab. | |
Example: | |
```python | |
params = { | |
"display_name": "Google Translate", | |
"is_tab": True, | |
} | |
``` | |
The `params` dict may also contain variables that you want to be customizable through a `settings.yaml` file. For instance, assuming the extension is in `extensions/google_translate`, the variable `language string` in | |
```python | |
params = { | |
"display_name": "Google Translate", | |
"is_tab": True, | |
"language string": "jp" | |
} | |
``` | |
can be customized by adding a key called `google_translate-language string` to `settings.yaml`: | |
```python | |
google_translate-language string: 'fr' | |
``` | |
That is, the syntax for the key is `extension_name-variable_name`. | |
## Using multiple extensions at the same time | |
You can activate more than one extension at a time by providing their names separated by spaces after `--extensions`. The input, output, and bot prefix modifiers will be applied in the specified order. | |
Example: | |
``` | |
python server.py --extensions enthusiasm translate # First apply enthusiasm, then translate | |
python server.py --extensions translate enthusiasm # First apply translate, then enthusiasm | |
``` | |
Do note, that for: | |
- `custom_generate_chat_prompt` | |
- `custom_generate_reply` | |
- `custom_tokenized_length` | |
only the first declaration encountered will be used and the rest will be ignored. | |
## A full example | |
The source code below can be found at [extensions/example/script.py](https://github.com/oobabooga/text-generation-webui/tree/main/extensions/example/script.py). | |
```python | |
""" | |
An example of extension. It does nothing, but you can add transformations | |
before the return statements to customize the webui behavior. | |
Starting from history_modifier and ending in output_modifier, the | |
functions are declared in the same order that they are called at | |
generation time. | |
""" | |
import gradio as gr | |
import torch | |
from transformers import LogitsProcessor | |
from modules import chat, shared | |
from modules.text_generation import ( | |
decode, | |
encode, | |
generate_reply, | |
) | |
params = { | |
"display_name": "Example Extension", | |
"is_tab": False, | |
} | |
class MyLogits(LogitsProcessor): | |
""" | |
Manipulates the probabilities for the next token before it gets sampled. | |
Used in the logits_processor_modifier function below. | |
""" | |
def __init__(self): | |
pass | |
def __call__(self, input_ids, scores): | |
# probs = torch.softmax(scores, dim=-1, dtype=torch.float) | |
# probs[0] /= probs[0].sum() | |
# scores = torch.log(probs / (1 - probs)) | |
return scores | |
def history_modifier(history): | |
""" | |
Modifies the chat history. | |
Only used in chat mode. | |
""" | |
return history | |
def state_modifier(state): | |
""" | |
Modifies the state variable, which is a dictionary containing the input | |
values in the UI like sliders and checkboxes. | |
""" | |
return state | |
def chat_input_modifier(text, visible_text, state): | |
""" | |
Modifies the user input string in chat mode (visible_text). | |
You can also modify the internal representation of the user | |
input (text) to change how it will appear in the prompt. | |
""" | |
return text, visible_text | |
def input_modifier(string, state, is_chat=False): | |
""" | |
In default/notebook modes, modifies the whole prompt. | |
In chat mode, it is the same as chat_input_modifier but only applied | |
to "text", here called "string", and not to "visible_text". | |
""" | |
return string | |
def bot_prefix_modifier(string, state): | |
""" | |
Modifies the prefix for the next bot reply in chat mode. | |
By default, the prefix will be something like "Bot Name:". | |
""" | |
return string | |
def tokenizer_modifier(state, prompt, input_ids, input_embeds): | |
""" | |
Modifies the input ids and embeds. | |
Used by the multimodal extension to put image embeddings in the prompt. | |
Only used by loaders that use the transformers library for sampling. | |
""" | |
return prompt, input_ids, input_embeds | |
def logits_processor_modifier(processor_list, input_ids): | |
""" | |
Adds logits processors to the list, allowing you to access and modify | |
the next token probabilities. | |
Only used by loaders that use the transformers library for sampling. | |
""" | |
processor_list.append(MyLogits()) | |
return processor_list | |
def output_modifier(string, state, is_chat=False): | |
""" | |
Modifies the LLM output before it gets presented. | |
In chat mode, the modified version goes into history['visible'], | |
and the original version goes into history['internal']. | |
""" | |
return string | |
def custom_generate_chat_prompt(user_input, state, **kwargs): | |
""" | |
Replaces the function that generates the prompt from the chat history. | |
Only used in chat mode. | |
""" | |
result = chat.generate_chat_prompt(user_input, state, **kwargs) | |
return result | |
def custom_css(): | |
""" | |
Returns a CSS string that gets appended to the CSS for the webui. | |
""" | |
return '' | |
def custom_js(): | |
""" | |
Returns a javascript string that gets appended to the javascript | |
for the webui. | |
""" | |
return '' | |
def setup(): | |
""" | |
Gets executed only once, when the extension is imported. | |
""" | |
pass | |
def ui(): | |
""" | |
Gets executed when the UI is drawn. Custom gradio elements and | |
their corresponding event handlers should be defined here. | |
To learn about gradio components, check out the docs: | |
https://gradio.app/docs/ | |
""" | |
pass | |
``` | |