zetavg commited on
Commit
37f2c31
1 Parent(s): ede48a7

update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -1
README.md CHANGED
@@ -12,6 +12,7 @@ Making evaluating and fine-tuning LLaMA models with low-rank adaptation (LoRA) e
12
  * **[1-click up and running in Google Colab](#run-on-google-colab)** with a standard GPU runtime.
13
  * Loads and stores data in Google Drive.
14
  * Evaluate various LLaMA LoRA models stored in your folder or from Hugging Face.<br /><a href="https://youtu.be/IoEMgouZ5xU"><img width="640px" src="https://user-images.githubusercontent.com/3784687/231023326-f28c84e2-df74-4179-b0ac-c25c4e8ca001.gif" /></a>
 
15
  * Fine-tune LLaMA models with different prompt templates and training dataset format.<br /><a href="https://youtu.be/IoEMgouZ5xU?t=60"><img width="640px" src="https://user-images.githubusercontent.com/3784687/231026640-b5cf5c79-9fe9-430b-8d4e-7346eb9567ad.gif" /></a>
16
  * Load JSON and JSONL datasets from your folder, or even paste plain text directly into the UI.
17
  * Supports Stanford Alpaca [seed_tasks](https://github.com/tatsu-lab/stanford_alpaca/blob/main/seed_tasks.jsonl), [alpaca_data](https://github.com/tatsu-lab/stanford_alpaca/blob/main/alpaca_data.json) and [OpenAI "prompt"-"completion"](https://platform.openai.com/docs/guides/fine-tuning/data-formatting) format.
@@ -63,11 +64,13 @@ setup: |
63
  pip install wandb
64
  cd ..
65
  echo 'Dependencies installed.'
 
 
66
 
67
  # Start the app.
68
  run: |
69
  echo 'Starting...'
70
- python llama_lora_tuner/app.py --data_dir='/data' --wandb_api_key "$([ -f /data/secrets/wandb_api_key ] && cat /data/secrets/wandb_api_key | tr -d '\n')" --base_model='decapoda-research/llama-7b-hf' --share
71
  ```
72
 
73
  Then launch a cluster to run the task:
 
12
  * **[1-click up and running in Google Colab](#run-on-google-colab)** with a standard GPU runtime.
13
  * Loads and stores data in Google Drive.
14
  * Evaluate various LLaMA LoRA models stored in your folder or from Hugging Face.<br /><a href="https://youtu.be/IoEMgouZ5xU"><img width="640px" src="https://user-images.githubusercontent.com/3784687/231023326-f28c84e2-df74-4179-b0ac-c25c4e8ca001.gif" /></a>
15
+ * Switch between base models such as `decapoda-research/llama-7b-hf`, `nomic-ai/gpt4all-j`, `databricks/dolly-v2-7b`, `EleutherAI/gpt-j-6b`, or `EleutherAI/pythia-6.9b`.
16
  * Fine-tune LLaMA models with different prompt templates and training dataset format.<br /><a href="https://youtu.be/IoEMgouZ5xU?t=60"><img width="640px" src="https://user-images.githubusercontent.com/3784687/231026640-b5cf5c79-9fe9-430b-8d4e-7346eb9567ad.gif" /></a>
17
  * Load JSON and JSONL datasets from your folder, or even paste plain text directly into the UI.
18
  * Supports Stanford Alpaca [seed_tasks](https://github.com/tatsu-lab/stanford_alpaca/blob/main/seed_tasks.jsonl), [alpaca_data](https://github.com/tatsu-lab/stanford_alpaca/blob/main/alpaca_data.json) and [OpenAI "prompt"-"completion"](https://platform.openai.com/docs/guides/fine-tuning/data-formatting) format.
 
64
  pip install wandb
65
  cd ..
66
  echo 'Dependencies installed.'
67
+ echo 'Pre-downloading base models so that you won't have to wait for long once the app is ready...'
68
+ python llama_lora_tuner/download_base_model.py --base_model_names='decapoda-research/llama-7b-hf,nomic-ai/gpt4all-j,databricks/dolly-v2-7b'
69
 
70
  # Start the app.
71
  run: |
72
  echo 'Starting...'
73
+ python llama_lora_tuner/app.py --data_dir='/data' --wandb_api_key="$([ -f /data/secrets/wandb_api_key ] && cat /data/secrets/wandb_api_key | tr -d '\n')" --base_model=decapoda-research/llama-7b-hf --base_model_choices='decapoda-research/llama-7b-hf,nomic-ai/gpt4all-j,databricks/dolly-v2-7b --share
74
  ```
75
 
76
  Then launch a cluster to run the task: