Update README.md
Browse files
README.md
CHANGED
@@ -23,3 +23,5 @@ This is a finetune of Llama-2-13b, intended to extend the useful context window
|
|
23 |
All training was performed with 1x RTX 6000 Ada.
|
24 |
|
25 |
For full model card, including how to use PNTK, see any of the two merged models linked above.
|
|
|
|
|
|
23 |
All training was performed with 1x RTX 6000 Ada.
|
24 |
|
25 |
For full model card, including how to use PNTK, see any of the two merged models linked above.
|
26 |
+
|
27 |
+
**IMPORTANT** There are TWO sets of adapter weights. `adapter_model_base` is to be applied to llama-2-13b. The result can then be merged with `adapter_model`. These adapters correspond to the first and second training phase respectively. Applying the second phase adapter to base llama-2-13b will produce different results to what the merged full airophin model.
|