--- license: mit library_name: peft language: - en pipeline_tag: text-generation tags: - facebook - meta - pytorch - llama - llama-2 --- **Website**: [FireAct Agent](https://fireact-agent.github.io) # **FireAct Llama-2/CodeLlama** FireAct Llama/CodeLlama is a collection of fine-tuned generative text models for performing ReAct with external search tools. Links to other models can be found in the Index section. ## Foundation Model Details *Note: As the foundation models, Llama-2 and CodeLlama, are developed by Meta, please also read the guidance and license on their website, [Llama-2](https://huggingface.co/meta-llama) and [CodeLlama](https://huggingface.co/codellama), before using FireAct models.* **Model Developers** System 2 Research, Cambridge LTL, Monash University, Princeton PLI. **Variations** FireAct models including Llama-2-7B full fine-tuned models, and Llama-2-[7B,13B], CodeLlama-[7B,13B,34B] LoRA fine-tuned models. All released models are fine-tuned on multi-task (HotpotQA/StrategyQA/MMLU) and multi-type (ReAct/CoT/Reflexion) data. **Input** Models input text only. **Output** Models generate text only. ## Index **Full Fine-tuned Model** FireAct Llama-2: - [fireact_llama_2_7b](https://huggingface.co/forestai/fireact_llama_2_7b) **LoRA Fine-tuned Model** FireAct Llama-2: - [fireact_llama_2_7b_lora](https://huggingface.co/forestai/fireact_llama_2_7b_lora) - [fireact_llama_2_13b_lora](https://huggingface.co/forestai/fireact_llama_2_13b_lora) FireAct CodeLlama: - [fireact_codellama_7b_lora](https://huggingface.co/forestai/fireact_codellama_7b_lora) - [fireact_codellama_13b_lora](https://huggingface.co/forestai/fireact_codellama_13b_lora) - [fireact_codellama_34b_lora](https://huggingface.co/forestai/fireact_codellama_34b_lora) ## LoRA Training procedure The following `bitsandbytes` quantization config was used during training: - load_in_8bit: True - load_in_4bit: False - llm_int8_threshold: 6.0 - llm_int8_skip_modules: None - llm_int8_enable_fp32_cpu_offload: False - llm_int8_has_fp16_weight: False - bnb_4bit_quant_type: fp4 - bnb_4bit_use_double_quant: False - bnb_4bit_compute_dtype: float32 ### Framework versions - PEFT 0.4.0