yhyu13
Upload
f021684
raw
history blame contribute delete
No virus
1.18 kB
from LM_Cocktail import mix_models_by_layers
import argparse
if __name__ == "__main__":
parser = argparse.ArgumentParser()
parser.add_argument("--model_type", type=str, default="decoder", help="Type of model to be mixed")
parser.add_argument("--output_path", type=str, default="./mixed_llm", help="Path to save the mixed model")
parser.add_argument("--max_length", type=int, default=100, help="Maximum length of the sequence to be generated")
parser.add_argument("--models", type=str, nargs='+', default=["meta-llama/Llama-2-7b-chat-hf", "Shitao/llama2-ag-news"], help="Path to the models to be mixed")
parser.add_argument("--weights", type=float, nargs='+', default=[0.7, 0.3], help="Weights of the models to be mixed")
parser.add_argument("--save_precision", type=str, default='float32', help="mixed model saved format")
args = parser.parse_args()
# Mix Large Language Models (LLMs) and save the combined model to the path: ./mixed_llm
model = mix_models_by_layers(
model_names_or_paths=args.models,
model_type=args.model_type,
weights=args.weights,
output_path=args.output_path)
print(model)