Model Card for rustral-2-2k-2024-1-6

This is an instruction finetune of Mistral-7B-v0.1 on a custom dataset of Rust code snippets. The current version was finetuned on 2k examples for 3 epochs, with instructions being extracted from the comments of the code itself. Future versions will be based on more data and generated instructions.

Model Details

Model Description

  • Developed by: Leon Luithlen
  • Model type: Mistral-7B-v0.1 variant
  • Language(s) (NLP): Rust, English
  • License: apache-2.0
  • Finetuned from model [optional]: Mistral-7B-v0.1

Uses

This model is intended as a code assistant that develops short snippets, such as a single Rust struct or function, from the immediately preceding code and an instruction.

Direct Use

After compiling to the appropriate quantization size for your hardware with llama.cpp, it can be used like other Mistral-7B variants with the following input format:

        Below is an instruction that describes a task, paired with an input that
        provides further context. Write a response that appropriately completes the
        request.

        ### Instruction:
        {instruction}

        ### Input:
        {up_to_two_previous_snippets}

        ### Response:

Out-of-Scope Use

Any programming language other than Rust, natural languages

Bias, Risks, and Limitations

Currently not evaluated systematically, as there is no agreed benchmark for code assistance tasks in Rust.

For uses outside its intended scope, the model is likely to behave similarly to Mistral-7B, but probably worse.

Recommendations

The same caveats that apply to Mistral-7B apply to this model as well.

Training Details

Training Data

Training data is sourced from open source and permissively licensed Rust projects, but is not public.

Downloads last month
2
Safetensors
Model size
7.24B params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.