nsarrazin HF staff commited on
Commit
b46dc11
1 Parent(s): 6f7b315

add details about websearch to README

Browse files
Files changed (1) hide show
  1. README.md +7 -3
README.md CHANGED
@@ -92,6 +92,10 @@ PUBLIC_APP_DISCLAIMER=
92
  - `PUBLIC_APP_DATA_SHARING` Can be set to 1 to add a toggle in the user settings that lets your users opt-in to data sharing with models creator.
93
  - `PUBLIC_APP_DISCLAIMER` If set to 1, we show a disclaimer about generated outputs on login.
94
 
 
 
 
 
95
  ### Custom models
96
 
97
  You can customize the parameters passed to the model or even use a new model by updating the `MODELS` variable in your `.env.local`. The default one can be found in `.env` and looks like this :
@@ -135,7 +139,7 @@ MODELS=`[
135
 
136
  You can change things like the parameters, or customize the preprompt to better suit your needs. You can also add more models by adding more objects to the array, with different preprompts for example.
137
 
138
- ### Running your own models using a custom endpoint
139
 
140
  If you want to, you can even run your own models locally, by having a look at our endpoint project, [text-generation-inference](https://github.com/huggingface/text-generation-inference). You can then add your own endpoints to the `MODELS` variable in `.env.local`, by adding an `"endpoints"` key for each model in `MODELS`.
141
 
@@ -150,7 +154,7 @@ If you want to, you can even run your own models locally, by having a look at ou
150
 
151
  If `endpoints` is left unspecified, ChatUI will look for the model on the hosted Hugging Face inference API using the model name.
152
 
153
- ### Custom endpoint authorization
154
 
155
  Custom endpoints may require authorization, depending on how you configure them. Authentication will usually be set either with `Basic` or `Bearer`.
156
 
@@ -175,7 +179,7 @@ You can then add the generated information and the `authorization` parameter to
175
 
176
  ```
177
 
178
- ### Models hosted on multiple custom endpoints
179
 
180
  If the model being hosted will be available on multiple servers/instances add the `weight` parameter to your `.env.local`. The `weight` will be used to determine the probability of requesting a particular endpoint.
181
 
 
92
  - `PUBLIC_APP_DATA_SHARING` Can be set to 1 to add a toggle in the user settings that lets your users opt-in to data sharing with models creator.
93
  - `PUBLIC_APP_DISCLAIMER` If set to 1, we show a disclaimer about generated outputs on login.
94
 
95
+ ### Web Search
96
+
97
+ You can enable the web search by adding either `SERPER_API_KEY` ([serper.dev](https://serper.dev/)) or `SERPAPI_KEY` ([serpapi.com](https://serpapi.com/)) to your `.env.local`.
98
+
99
  ### Custom models
100
 
101
  You can customize the parameters passed to the model or even use a new model by updating the `MODELS` variable in your `.env.local`. The default one can be found in `.env` and looks like this :
 
139
 
140
  You can change things like the parameters, or customize the preprompt to better suit your needs. You can also add more models by adding more objects to the array, with different preprompts for example.
141
 
142
+ #### Running your own models using a custom endpoint
143
 
144
  If you want to, you can even run your own models locally, by having a look at our endpoint project, [text-generation-inference](https://github.com/huggingface/text-generation-inference). You can then add your own endpoints to the `MODELS` variable in `.env.local`, by adding an `"endpoints"` key for each model in `MODELS`.
145
 
 
154
 
155
  If `endpoints` is left unspecified, ChatUI will look for the model on the hosted Hugging Face inference API using the model name.
156
 
157
+ #### Custom endpoint authorization
158
 
159
  Custom endpoints may require authorization, depending on how you configure them. Authentication will usually be set either with `Basic` or `Bearer`.
160
 
 
179
 
180
  ```
181
 
182
+ #### Models hosted on multiple custom endpoints
183
 
184
  If the model being hosted will be available on multiple servers/instances add the `weight` parameter to your `.env.local`. The `weight` will be used to determine the probability of requesting a particular endpoint.
185