xiaol's picture
Update README.md
7f8c5be
metadata
license: apache-2.0
datasets:
  - stingning/ultrachat
  - kaist-ai/CoT-Collection
  - mesolitica/google-translate-commitpackft
  - Wanfq/Explore_Instruct_Rewriting_32k
  - Wanfq/Explore_Instruct_Rewriting_10k
  - Wanfq/Explore_Instruct_Brainstorming_16k
  - xiyuez/red-dot-design-award-product-description

RWKV v4 7B world model

finetuned with ultrachat , COT and some novel instructions data, commitpackft and so on

use full ultrachat and cot data, about 3B tokens

if you wanna do Role play, use this model

Contributor

@JL-er @Remixa

Design of experiment

this model lose multi-turn chat ability,cause from using whole ultrachat datasets.

so i continue tuned multi-turn datasets with 2 aspects

1.role play

2.novel multiturn instruction

Training details

wandb.ai

CAses

image/jpeg

image/jpeg

Usage

adjust tempp and topp on different scenario. image/png

image/png

image/png

COT and lookback

image/png this model can do above task with 100% acc.

image/png

role play model

image/png

image/png

novel

image/png

image/png

demo site(temporary)

online showcase