|
--- |
|
language: |
|
- en |
|
license: mit |
|
library_name: transformers |
|
datasets: |
|
- julep-ai/samantha_finetune_dataset_03 |
|
pipeline_tag: text-generation |
|
--- |
|
|
|
# Samantha |
|
|
|
## Technical notes |
|
|
|
This model is trained on a specialized dataset and uses special sentinel tokens to demarcate conversations. |
|
|
|
### Usage |
|
For usage, you can refer to the [`chat.py`](https://huggingface.co/julep-ai/samantha-33b/blob/main/chat.py) file in this repo for an example. |
|
|
|
### Concepts |
|
- Each conversation consists of n "sections" |
|
- Each section can be one of: |
|
+ `me`: The model |
|
+ `person`: The speaker |
|
+ `situation`: relevant background information to set the context of the conversation |
|
+ `thought`: Thoughts generated by the model for parsing intermediate steps etc |
|
+ `information`: External information added into the context by the system running the model |
|
- The model and speaker sections can optionally include a name like `me (Samantha)` or `person (Dmitry)` |
|
|
|
### Sentinel Tokens |
|
- `<|im_start|>` token marks the start of a "section" |
|
- `<|im_end|>` token marks the end of a "section". |
|
|
|
## Example |
|
|
|
``` |
|
<|im_start|>situation |
|
I am talking to Diwank. I want to ask him about his food preferences.<|im_end|> |
|
<|im_start|>person (Diwank) |
|
Hey Samantha! What do you want to talk about?<|im_end|> |
|
<|im_start|>me (Samantha) |
|
``` |