--- library_name: transformers tags: - text-generation datasets: Jellywibble/dalio-principles-cleaned-v3 widget: - text: 'This is a conversation where Ray Dalio is giving advice on being a manager and building a successful team. User: Hi Ray, thanks for talking with me today. I am excited to learn more about how to follow your principles and build a successful company. Ray: No problem, I am happy to help. What situation are you facing? User: It feels like I keep making decisions without thinking first - I do something without thinking and then I face the consequences afterwards. Ray:' example_title: Q&A - text: 'It’s easy to tell an open-minded person from a closed-minded person because they act very differently. Here are some cues to tell you whether you or others are being closed-minded: ' example_title: Principles --- ## Model Description Pre-training on cleaned version of Principles - removing numeric references to footnotes - removing numeric counts, i.e. 1) ... 2) ... 3) ... - correcting gramma, i.e. full stops must be followed by a space - finetuning OPT-30B model on the dataset above - Dataset location: Jellywibble/dalio-principles-cleaned-v3 ## Metrics - Checkpoint 8 served - Hellaswag Perplexity: 30.65 - 2.289 eval loss wandb link: https://wandb.ai/jellywibble/huggingface/runs/2jqc504o?workspace=user-jellywibble ## Model Parameters Trained on 4xA40, effective batchsize = 8 - base_model_name facebook/opt-30b - dataset_name Jellywibble/dalio-principles-cleaned-v3 - block_size 1024 - gradient_accumulation_steps 2 - per_device_train_batch_size 1 - seed 2 - num_train_epochs 1 - learning_rate 3e-6 ## Notes - It is important for the effective batch size to be at least 8 - Learning rate higher than 3e-6 will result in massive overfitting, i.e. much worse Hellaswag metrics