Edit model card

DevOps-Model-7B-Chat

🤗 Hugging Face • 🤖 ModelScope

DevOps-Model is a Chinese DevOps large model, mainly dedicated to exerting practical value in the field of DevOps. Currently, DevOps-Model can help engineers answer questions encountered in the all DevOps life cycle.

Based on the Qwen series of models, we output the Base model after additional training with high-quality Chinese DevOps corpus, and then output the Chat model after alignment with DevOps QA data. Our Base model and Chat model can achieve the best results among models of the same scale based on evaluation data related to the DevOps fields.


At the same time, we are also building an evaluation benchmark [DevOpsEval](https://github.com/codefuse-ai/codefuse-devops-eval) exclusive to the DevOps field to better evaluate the effect of the DevOps field model.

Evaluation

We first selected a total of six exams related to DevOps in the two evaluation data sets of CMMLU and CEval. There are a total of 574 multiple-choice questions. The specific information is as follows:

Evaluation dataset Exam subjects Number of questions
CMMLU Computer science 204
CMMLU Computer security 171
CMMLU Machine learning 122
CEval College programming 37
CEval Computer architecture 21
CEval Computernetwork 19

We tested the results of Zero-shot and Five-shot respectively. Our 7B and 14B series models can achieve the best results among the tested models. More tests will be released later.

Model Size Zero-shot Score Five-shot Score
DevOps-Model-7B-Chat 7B 62.20 64.11
Qwen-7B-Chat 7B 46.00 52.44
Baichuan2-7B-Chat 7B 52.26 54.46
Internlm-7B-Chat 7B 52.61 55.75

Quickstart

We provide simple examples to illustrate how to quickly use Devops-Model-Chat models with 🤗 Transformers.

Requirement

cd path_to_download_model
pip install -r requirements.txt

Model Example

from transformers import AutoModelForCausalLM, AutoTokenizer
from transformers.generation import GenerationConfig

tokenizer = AutoTokenizer.from_pretrained("path_to_DevOps-Model", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("path_to_DevOps-Model", device_map="auto", trust_remote_code=True, bf16=True).eval()

model.generation_config = GenerationConfig.from_pretrained("path_to_DevOps-Model", trust_remote_code=True)

resp, hist = model.chat(query='What is the difference between HashMap and Hashtable in Java', tokenizer=tokenizer, history=None)

Disclaimer

Due to the characteristics of language models, the content generated by the model may contain hallucinations or discriminatory remarks. Please use the content generated by the DevOps-Model family of models with caution. If you want to use this model service publicly or commercially, please note that the service provider needs to bear the responsibility for the adverse effects or harmful remarks caused by it. The developer of this project does not assume any responsibility for any consequences caused by the use of this project (including but not limited to data, models, codes, etc.) ) resulting in harm or loss.

Acknowledgments

This project refers to the following open source projects, and I would like to express my gratitude to the relevant projects and research and development personnel.

Downloads last month
17
Inference API
Inference API (serverless) does not yet support model repos that contain custom code.