Effects of training data

#1
by ChuckMcSneed - opened

This model has a tendency to spit out stuff like <|prompt|> which was not present in base llama. It is also a bit drier than the base model. It seems that the training with Yarn has effects on output quality which were not observed in the benchmarks.

Sign up or log in to comment