GGUF
Inference Endpoints
File size: 1,191 Bytes
23ea8fe
a219a4e
23ea8fe
 
 
 
e9b9972
23ea8fe
 
 
 
 
9dc9906
b98d0cc
bd715cc
23ea8fe
 
 
 
02bb187
 
 
 
23ea8fe
 
 
 
02bb187
 
 
 
 
23ea8fe
 
e9b9972
23ea8fe
a219a4e
e9b9972
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
---
license: other
---

## Overview

Mistral 7B, a 7-billion-parameter Large Language Model by Mistral AI. Designed for efficiency and performance, it suits real-time applications requiring swift responses.

## Variants

| No | Variant | Cortex CLI command |
| --- | --- | --- |
| 1 | [7b-gguf](https://huggingface.co/cortexhub/mistral/tree/7b-gguf) | `cortex run mistral:7b-gguf` |
| 2 | [tensorrt-llm](https://huggingface.co/cortexso/mistral/tree/tensorrt-llm-windows-ada) | `cortex run mistral:tensorrt-llm` |
| 3 | [7b-onnx](https://huggingface.co/cortexso/mistral/tree/7b-onnx) | `cortex run mistral:7b-onnx` |

## Use it with Jan (UI)

1. Install **Jan** using [Quickstart](https://jan.ai/docs/quickstart)
2. Use in Jan model Hub:
    ```
    cortexhub/mistral
    ```

## Use it with Cortex (CLI)

1. Install **Cortex** using [Quickstart](https://cortex.jan.ai/docs/quickstart)
2. Run the model with command:
    ```
    cortex run mistral
    ```
    
## Credits

- **Author:** MistralAI
- **Converter:** [Homebrew](https://www.homebrew.ltd/)
- **Original License:** [Licence]((https://mistral.ai/licenses/MNPL-0.1.md))
- **Papers:** [Mistral paper](https://arxiv.org/abs/2310.06825)