Update app.py
Browse files
app.py
CHANGED
@@ -94,6 +94,48 @@ with gr.Blocks(css=css) as demo:
|
|
94 |
gr.ChatInterface(
|
95 |
generate,
|
96 |
additional_inputs=additional_inputs,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
97 |
examples = [
|
98 |
["π° Welcome to the Kingdom of Elandria! You are Jim and Tim, two bumbling bros with a knack for mischief. π€΄π€΄ [Action: Introduce yourselves, Equipment: Scepters of Foolishness]"],
|
99 |
["π² You find yourselves in a forest filled with magical creatures and oddly specific 'Do Not Disturb' signs. π¦ [Action: Proceed cautiously, Equipment: Map of Social Etiquette]"],
|
|
|
94 |
gr.ChatInterface(
|
95 |
generate,
|
96 |
additional_inputs=additional_inputs,
|
97 |
+
markdown="""
|
98 |
+
| Feature | Description | Byline |
|
99 |
+
|---------|-------------|--------|
|
100 |
+
| πͺ Sliding Window Attention with 128K tokens span | Enables the model to have a larger context for each token. | Increases model's understanding of context, resulting in more coherent and contextually relevant outputs. |
|
101 |
+
| π GQA for faster inference | Graph Query Attention allows faster computation during inference. | Speeds up the model inference time without sacrificing too much on accuracy. |
|
102 |
+
| π Byte-fallback BPE tokenizer | Uses Byte Pair Encoding but can fall back to byte-level encoding. | Allows the tokenizer to handle a wider variety of input text while keeping token size manageable. |
|
103 |
+
| π License | Released under Apache 2.0 License | Gives you a permissive free software license, allowing you freedom to use, modify, and distribute the code. |
|
104 |
+
| π¦ Usage | | |
|
105 |
+
| π Available on Huggingface Hub | The model can be easily downloaded and set up from Huggingface. | Makes it easier to integrate the model into various projects. |
|
106 |
+
| π Python code snippets for easy setup | Provides Python code snippets for quick and easy model setup. | Facilitates rapid development and deployment, especially useful for prototyping. |
|
107 |
+
| π Expected speedups with Flash Attention 2 | Upcoming update expected to bring speed improvements. | Keep an eye out for this update to benefit from performance gains. |
|
108 |
+
# π Model Features and More π
|
109 |
+
|
110 |
+
## Features
|
111 |
+
|
112 |
+
- πͺ Sliding Window Attention with 128K tokens span
|
113 |
+
- **Byline**: Increases model's understanding of context, resulting in more coherent and contextually relevant outputs.
|
114 |
+
|
115 |
+
- π GQA for faster inference
|
116 |
+
- **Byline**: Speeds up the model inference time without sacrificing too much on accuracy.
|
117 |
+
|
118 |
+
- π Byte-fallback BPE tokenizer
|
119 |
+
- **Byline**: Allows the tokenizer to handle a wider variety of input text while keeping token size manageable.
|
120 |
+
|
121 |
+
- π License: Released under Apache 2.0 License
|
122 |
+
- **Byline**: Gives you a permissive free software license, allowing you freedom to use, modify, and distribute the code.
|
123 |
+
|
124 |
+
## Usage π¦
|
125 |
+
|
126 |
+
- π Available on Huggingface Hub
|
127 |
+
- **Byline**: Makes it easier to integrate the model into various projects.
|
128 |
+
|
129 |
+
- π Python code snippets for easy setup
|
130 |
+
- **Byline**: Facilitates rapid development and deployment, especially useful for prototyping.
|
131 |
+
|
132 |
+
- π Expected speedups with Flash Attention 2
|
133 |
+
- **Byline**: Keep an eye out for this update to benefit from performance gains.
|
134 |
+
|
135 |
+
"""
|
136 |
+
|
137 |
+
|
138 |
+
gr.Markdown(markdown)
|
139 |
examples = [
|
140 |
["π° Welcome to the Kingdom of Elandria! You are Jim and Tim, two bumbling bros with a knack for mischief. π€΄π€΄ [Action: Introduce yourselves, Equipment: Scepters of Foolishness]"],
|
141 |
["π² You find yourselves in a forest filled with magical creatures and oddly specific 'Do Not Disturb' signs. π¦ [Action: Proceed cautiously, Equipment: Map of Social Etiquette]"],
|