Edit model card

AgentLM-70B

🤗 [Dataset] • 💻 [Github Repo] • 📌 [Project Page] • 📃 [Paper]

AgentTuning represents the very first attempt to instruction-tune LLMs using interaction trajectories across multiple agent tasks. Evaluation results indicate that AgentTuning enables the agent capabilities of LLMs with robust generalization on unseen agent tasks while remaining good on general language abilities. We have open-sourced the AgentInstruct dataset and AgentLM.

Models

AgentLM models are produced by mixed training on AgentInstruct dataset and ShareGPT dataset from Llama-2-chat models.

The models follow the conversation format of Llama-2-chat, with system prompt fixed as

You are a helpful, respectful and honest assistant.

7B, 13B, and 70B models are available on Huggingface model hub.

Model Huggingface Repo
AgentLM-7B 🤗Huggingface Repo
AgentLM-13B 🤗Huggingface Repo
AgentLM-70B 🤗Huggingface Repo

Citation

If you find our work useful, please consider citing AgentTuning:

@misc{zeng2023agenttuning,
      title={AgentTuning: Enabling Generalized Agent Abilities for LLMs}, 
      author={Aohan Zeng and Mingdao Liu and Rui Lu and Bowen Wang and Xiao Liu and Yuxiao Dong and Jie Tang},
      year={2023},
      eprint={2310.12823},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}
Downloads last month
78
Safetensors
Model size
69B params
Tensor type
FP16
·
F32
·

Dataset used to train THUDM/agentlm-70b

Spaces using THUDM/agentlm-70b 8