Getting cut-off responses with Mixtral 8x7B-Instruct-v0.1 mostly in Date of Birth years
3
#182 opened 9 months ago
by
keskival
How can I run it on multiple GPUs?
11
#181 opened 9 months ago
by
barbery
Where is the mixtral-8x7b's tokenizer encoder? Is there a specific repository or node module?
1
#180 opened 9 months ago
by
RamanSB
What is the max token limit on this model?
2
#179 opened 9 months ago
by
RamanSB
Finetuning Mixtral 8x7B Instruct-v0.1 using Transformers
2
#178 opened 9 months ago
by
Ateeqq
Update chat template to resemble the prompt as stated in the model card.
7
#176 opened 9 months ago
by
nilsec
max_sequence_length
1
#175 opened 9 months ago
by
Ravnoor1
Awesome. I Got Very Good Responses, However...
#174 opened 9 months ago
by
deleted
How to run the full model ?
2
#171 opened 10 months ago
by
dounykim
Is there a working/quantized/exl2 (etc) version that will fit on a single 24GB video card (4090)
2
#170 opened 10 months ago
by
cleverest
403 error
1
#169 opened 10 months ago
by
minhphan-qbe
Adding Evaluation Results
#168 opened 10 months ago
by
leaderboard-pr-bot
Rename README.md to RegulusOne
#167 opened 10 months ago
by
Theguy666
Help: CUDA Out of Memory. Hardware requirements.
2
#147 opened 10 months ago
by
zebfreeman
Update README.md
#146 opened 10 months ago
by
frank76rm
Experimental use
#144 opened 10 months ago
by
yassineelkhadiri14
TemplateError: Conversation roles must alternate user/assistant/user/assistant/...
4
#143 opened 10 months ago
by
quamer23
Is instruction format necessary
2
#142 opened 10 months ago
by
supercharge19
[AUTOMATED] Model Memory Requirements
3
#141 opened 10 months ago
by
model-sizer-bot
Update README.md
#140 opened 10 months ago
by
woodyk
Cuda Out of memory issue when deploying mistralai/Mixtral-8x7B-Instruct-v0.1 on AWS "ml.g5.48xlarge"
1
#139 opened 10 months ago
by
sonalisbapte
slow response
1
#138 opened 10 months ago
by
bhavanam2809
Sparsity in mixtral
#137 opened 10 months ago
by
dpk17
Request: DOI
#136 opened 10 months ago
by
Sonny03
Running in Multi-gpu's
5
#134 opened 10 months ago
by
kmukeshreddy
Update README.md
#133 opened 10 months ago
by
gmverbas
How to format custom dataset to finetune Mixtral with TRL SFT script?
#132 opened 10 months ago
by
icpro
How to use run the code on Colab Free Tier or Mac OS?
16
#131 opened 10 months ago
by
dounykim
Different answer after each request
2
#130 opened 10 months ago
by
amin2557
How to finetune the model?
2
#129 opened 10 months ago
by
akasranjan
How much Resource is needed to run the Mixtral ?
1
#128 opened 10 months ago
by
rkhapre
Update README.md
#126 opened 11 months ago
by
mariakatosvich
The inference API Endpoint gives wrongly formatted answer based on the context given but works well in example Spaces. How we can fix this?
9
#125 opened 11 months ago
by
rkhapre
Request: DOI
#124 opened 11 months ago
by
jsr2
Update README.md
#123 opened 11 months ago
by
Pawamami
what is max input token limit of this model?
1
#122 opened 11 months ago
by
vaidehirao
Request: SDFSDFSD
1
#120 opened 11 months ago
by
seedeera
Consistency check failed - model-00019-of-00019.safetensors
#118 opened 11 months ago
by
br1-pist
Difference in EOS token between Mistral/Mixtral and LLAMA.
1
#117 opened 11 months ago
by
xkszltl
Model Output is Changed
9
#116 opened 11 months ago
by
AnzaniAI
The chat template doesn't support a system prompt
6
#114 opened 11 months ago
by
sam-kap
How to get 'output_router_logits'
1
#113 opened 11 months ago
by
cts13
Run inference on 2 GPUs
1
#112 opened 11 months ago
by
bweinstein123
Running a 4-bit Quantized 7B Model on a PC: Feasibility and Insights
4
#109 opened 11 months ago
by
edw-hug-face
How to get result as good as Hugging Face Chat Mixtral-8x7b-Instruct
9
#107 opened 11 months ago
by
Panckackes
FR Discutions ouverte
3
#105 opened 11 months ago
by
Pablito2fois
Rapport d'étonnement
#104 opened 11 months ago
by
YannCHANET