Spaces:
Running
on
Zero
Running
on
Zero
File size: 3,773 Bytes
0f43f8a |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 |
---
title: Examples
description: Get started by copying these code snippets into your terminal, a `.py` file, or a Jupyter notebook.
---
<CardGroup>
<Card
title="Interactive demo"
icon="gamepad-modern"
iconType="solid"
href="https://colab.research.google.com/drive/1WKmRXZgsErej2xUriKzxrEAXdxMSgWbb?usp=sharing"
>
Try Open Interpreter without installing anything on your computer
</Card>
<Card
title="Example voice interface"
icon="circle"
iconType="solid"
href="https://colab.research.google.com/drive/1NojYGHDgxH6Y1G1oxThEBBb2AtyODBIK"
>
An example implementation of Open Interpreter's streaming capabilities
</Card>
</CardGroup>
---
### Interactive Chat
To start an interactive chat in your terminal, either run `interpreter` from the command line:
```shell
interpreter
```
Or `interpreter.chat()` from a .py file:
```python
interpreter.chat()
```
---
### Programmatic Chat
For more precise control, you can pass messages directly to `.chat(message)` in Python:
```python
interpreter.chat("Add subtitles to all videos in /videos.")
# ... Displays output in your terminal, completes task ...
interpreter.chat("These look great but can you make the subtitles bigger?")
# ...
```
---
### Start a New Chat
In your terminal, Open Interpreter behaves like ChatGPT and will not remember previous conversations. Simply run `interpreter` to start a new chat:
```shell
interpreter
```
In Python, Open Interpreter remembers conversation history. If you want to start fresh, you can reset it:
```python
interpreter.messages = []
```
---
### Save and Restore Chats
In your terminal, Open Interpreter will save previous conversations to `<your application directory>/Open Interpreter/conversations/`.
You can resume any of them by running `--conversations`. Use your arrow keys to select one , then press `ENTER` to resume it.
```shell
interpreter --conversations
```
In Python, `interpreter.chat()` returns a List of messages, which can be used to resume a conversation with `interpreter.messages = messages`:
```python
# Save messages to 'messages'
messages = interpreter.chat("My name is Killian.")
# Reset interpreter ("Killian" will be forgotten)
interpreter.messages = []
# Resume chat from 'messages' ("Killian" will be remembered)
interpreter.messages = messages
```
---
### Configure Default Settings
We save default settings to a profile which can be edited by running the following command:
```shell
interpreter --profiles
```
You can use this to set your default language model, system message (custom instructions), max budget, etc.
<Info>
**Note:** The Python library will also inherit settings from the default
profile file. You can change it by running `interpreter --profiles` and
editing `default.yaml`.
</Info>
---
### Customize System Message
In your terminal, modify the system message by [editing your configuration file as described here](#configure-default-settings).
In Python, you can inspect and configure Open Interpreter's system message to extend its functionality, modify permissions, or give it more context.
```python
interpreter.system_message += """
Run shell commands with -y so the user doesn't have to confirm them.
"""
print(interpreter.system_message)
```
---
### Change your Language Model
Open Interpreter uses [LiteLLM](https://docs.litellm.ai/docs/providers/) to connect to language models.
You can change the model by setting the model parameter:
```shell
interpreter --model gpt-3.5-turbo
interpreter --model claude-2
interpreter --model command-nightly
```
In Python, set the model on the object:
```python
interpreter.llm.model = "gpt-3.5-turbo"
```
[Find the appropriate "model" string for your language model here.](https://docs.litellm.ai/docs/providers/)
|