--- license: other language: - zh tags: - Text Generation - LLM ---

DevOps-Model-7B-Chat

🤗 Hugging Face • 🤖 ModelScope

DevOps-Model is a Chinese **DevOps large model**, mainly dedicated to exerting practical value in the field of DevOps. Currently, DevOps-Model can help engineers answer questions encountered in the all DevOps life cycle. Based on the Qwen series of models, we output the **Base** model after additional training with high-quality Chinese DevOps corpus, and then output the **Chat** model after alignment with DevOps QA data. Our Base model and Chat model can achieve the best results among models of the same scale based on evaluation data related to the DevOps fields.
At the same time, we are also building an evaluation benchmark [DevOpsEval](https://github.com/codefuse-ai/codefuse-devops-eval) exclusive to the DevOps field to better evaluate the effect of the DevOps field model.

# Evaluation We first selected a total of six exams related to DevOps in the two evaluation data sets of CMMLU and CEval. There are a total of 574 multiple-choice questions. The specific information is as follows: | Evaluation dataset | Exam subjects | Number of questions | |:-------:|:-------:|:-------:| | CMMLU | Computer science | 204 | | CMMLU | Computer security | 171 | | CMMLU | Machine learning | 122 | | CEval | College programming | 37 | | CEval | Computer architecture | 21 | | CEval | Computernetwork | 19 | We tested the results of Zero-shot and Five-shot respectively. Our 7B and 14B series models can achieve the best results among the tested models. More tests will be released later. |Model|Size|Zero-shot Score|Five-shot Score| |--|--|--|--| |**DevOps-Model-7B-Chat**|**7B**|**62.20**|**64.11**| |Qwen-7B-Chat|7B|46.00|52.44| |Baichuan2-7B-Chat|7B|52.26|54.46| |Internlm-7B-Chat|7B|52.61|55.75|
# Quickstart We provide simple examples to illustrate how to quickly use Devops-Model-Chat models with 🤗 Transformers. ## Requirement ```bash cd path_to_download_model pip install -r requirements.txt ``` ## Model Example ```python from transformers import AutoModelForCausalLM, AutoTokenizer from transformers.generation import GenerationConfig tokenizer = AutoTokenizer.from_pretrained("path_to_DevOps-Model", trust_remote_code=True) model = AutoModelForCausalLM.from_pretrained("path_to_DevOps-Model", device_map="auto", trust_remote_code=True, bf16=True).eval() model.generation_config = GenerationConfig.from_pretrained("path_to_DevOps-Model", trust_remote_code=True) resp, hist = model.chat(query='What is the difference between HashMap and Hashtable in Java', tokenizer=tokenizer, history=None) ``` # Disclaimer Due to the characteristics of language models, the content generated by the model may contain hallucinations or discriminatory remarks. Please use the content generated by the DevOps-Model family of models with caution. If you want to use this model service publicly or commercially, please note that the service provider needs to bear the responsibility for the adverse effects or harmful remarks caused by it. The developer of this project does not assume any responsibility for any consequences caused by the use of this project (including but not limited to data, models, codes, etc.) ) resulting in harm or loss. # Acknowledgments This project refers to the following open source projects, and I would like to express my gratitude to the relevant projects and research and development personnel. - [LLaMA-Efficient-Tuning](https://github.com/hiyouga/LLaMA-Efficient-Tuning) - [QwenLM](https://github.com/QwenLM)