Edit model card

OpenChatRWKV 430m r2

This is a finetune of RWKV-v4neo 430m on openchatgpt safe r2 dataset. r2 shares no data with the r1, even the greetings, making this finetune, in some ways, inferiour to the original on r1.

Key differences with openchatrwkv-430m

One of the key differences is using an actual token for the instant message separation, apart from the new dataset.

Differences with openchatgpt-neox-125m

Increased parameter size and different dataset.

Training data

New dataset was obviously made at a later point in time. Many speculate that ChatGPT has degraded over the months, and I am a strong believer of that aswell - style in which model speaks started to sound different compared to 2-3 months ago.

This model was trained on a mix of natural language and code.

This model does not know how to greet you.

Downloads last month
0
Inference Examples
Unable to determine this model's library. Check the docs .