Running in Multi-gpu's
5
#134 opened 3 months ago
by
kmukeshreddy
Update README.md
#133 opened 3 months ago
by
gmverbas
How to format custom dataset to finetune Mixtral with TRL SFT script?
#132 opened 3 months ago
by
icpro
How to use run the code on Colab Free Tier or Mac OS?
16
#131 opened 3 months ago
by
dounykim
Different answer after each request
2
#130 opened 3 months ago
by
amin2557
How to finetune the model?
2
#129 opened 3 months ago
by
akasranjan
How much Resource is needed to run the Mixtral ?
1
#128 opened 3 months ago
by
rkhapre
Update README.md
#126 opened 3 months ago
by
mariakatosvich
The inference API Endpoint gives wrongly formatted answer based on the context given but works well in example Spaces. How we can fix this?
9
#125 opened 3 months ago
by
rkhapre
Request: DOI
#124 opened 3 months ago
by
jsr2
Update README.md
#123 opened 3 months ago
by
Pawamami
what is max input token limit of this model?
1
#122 opened 3 months ago
by
vaidehirao
Request: SDFSDFSD
1
#120 opened 3 months ago
by
seedeera
Consistency check failed - model-00019-of-00019.safetensors
#118 opened 4 months ago
by
br1-pist
Difference in EOS token between Mistral/Mixtral and LLAMA.
1
#117 opened 4 months ago
by
xkszltl
Model Output is Changed
7
#116 opened 4 months ago
by
AnzaniAI
The chat template doesn't support a system prompt
6
#114 opened 4 months ago
by
sam-kap
How to get 'output_router_logits'
1
#113 opened 4 months ago
by
cts13
Run inference on 2 GPUs
1
#112 opened 4 months ago
by
bweinstein123
Running a 4-bit Quantized 7B Model on a PC: Feasibility and Insights
4
#109 opened 4 months ago
by
edw-hug-face
How to get result as good as Hugging Face Chat Mixtral-8x7b-Instruct
9
#107 opened 4 months ago
by
Panckackes
FR Discutions ouverte
3
#105 opened 4 months ago
by
Pablito2fois
Rapport d'étonnement
#104 opened 4 months ago
by
YannCHANET
How to use transfimer
#100 opened 4 months ago
by
sethdwumah
SFT is so BAD
#99 opened 4 months ago
by
GokhanAI
8bit quantization error
1
#98 opened 4 months ago
by
lovelyfrog
Key Error : Mixtral
8
#96 opened 4 months ago
by
jdjayakaran
Train the Model on Confluence
1
#95 opened 4 months ago
by
icemaro
Run Mistral model on Remote server
6
#94 opened 4 months ago
by
icemaro
Cuda Error
1
#93 opened 4 months ago
by
HuggySSO
Not supported with TGI
3
#92 opened 4 months ago
by
abhishek3jangid
deepspeed load mixtral-8x7B hang or oom
1
#91 opened 4 months ago
by
guowl
Add MOE (mixture of experts) tag
#90 opened 4 months ago
by
davanstrien
Update README.md
#89 opened 4 months ago
by
schuyler12
Failure in loading the model on AWS
8
#88 opened 4 months ago
by
bweinstein123
Hardware Requirements
4
#86 opened 4 months ago
by
ShivanshMathur007
Response content was truncated
19
#84 opened 4 months ago
by
ludomare
Best parameter setting for Mixtral model on the text-generation task
#83 opened 4 months ago
by
kmukeshreddy
Any hints on prompt to reduce / stop hallucinations
1
#82 opened 4 months ago
by
dnovak232
Still the best Mixtral based instruct model. We should change that
#81 opened 4 months ago
by
rombodawg
Could not convert to integer: 3221225477 error
#80 opened 4 months ago
by
KharabinDev42
Serving the model as API on vLLM and 2 x A6000
2
#78 opened 5 months ago
by
dnovak232
How much memory do I need for this model (on Windows)?
3
#77 opened 5 months ago
by
roboboot
Inconsistent prompt format. Which is correct the Model card or the tokenizer_config.json?
6
#75 opened 5 months ago
by
lemonflourorange
can not run sft full finetuning.
9
#74 opened 5 months ago
by
hegang126
[Chinese Version] Mixtral-8x7B model | 中文Mixtral-8x7B模型
#73 opened 5 months ago
by
wangrongsheng
Update the deprecated Flash Attention call parameter in from_pretrained() method
#72 opened 5 months ago
by
DeathReaper0965