Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
Flagging models with incorrect tags
pinned
20
#540 opened 2 months ago
by
adamo1139
💬 Discussion thread: Model contamination techniques 💬
pinned
30
#472 opened 4 months ago
by
clefourrier
Future feature: system prompt and chat support
pinned
18
#459 opened 4 months ago
by
clefourrier
NEW! DROP was dropped discussion!
pinned
3
#420 opened 4 months ago
by
clefourrier
💬 Discussion thread: Model scores and model performances 💬
pinned
70
#265 opened 7 months ago
by
clefourrier
⁉️ FAQ - Start here before opening an issue ⁉️
pinned
7
#179 opened 8 months ago
by
clefourrier
💎 Resources and community initiatives around the Leaderboard! 💎
pinned#174 opened 8 months ago
by
clefourrier
MMLU blog post discussion
pinned
55
#82 opened 9 months ago
by
thomwolf
🚩 Report: Not working
2
#650 opened about 14 hours ago
by
Lyte
🚩 Report: Not working
1
#649 opened 1 day ago
by
nuralimedeunu
Leaderboard extremely slow to load
19
#648 opened 1 day ago
by
FenixInDarkSolo
Upload MANU9N4 OM Coursework 2024 Spring-4.docx
#647 opened 1 day ago
by
Patell
Evaluation disappeared!
4
#646 opened 2 days ago
by
david-infinimol
Leaderboard restarting
1
#645 opened 2 days ago
by
DenisTheDev
I submitted my model a while ago and it never got benchmarked
2
#644 opened 4 days ago
by
rombodawg
Covasna-0.1 failed
1
#643 opened 5 days ago
by
Mihaiii
Questionable results for the current top models
2
#642 opened 7 days ago
by
ammarali32
Seperate larger from smaller model benchmarking.
1
#640 opened 9 days ago
by
rombodawg
Submitting is dead?
2
#639 opened 9 days ago
by
Stopwolf
Feature Request: Option to show the average score of Arc, MMLU and WinoGrande.
3
#638 opened 12 days ago
by
Phil337
What tags need to be existing in a model card to be detected as merger?
2
#637 opened 12 days ago
by
ThiloteE
Feature Request: Add needle in haystack test for models, to rest for recall accuracy
2
#636 opened 12 days ago
by
CombinHorizon
Delete broken model from the HF leaderboard results and re-run Mega-Destroyer-8x7B.
5
#635 opened 15 days ago
by
Dampfinchen
Our model has disappeared from the leaderboard
3
#634 opened 15 days ago
by
bongchoi
Feature request: New evaluation
1
#633 opened 16 days ago
by
4Randomperson43
Request to delete a wrong precision model
5
#632 opened 16 days ago
by
4season
Please add CohereForAI/c4ai-command-r-v01 (trust_remote_code=True needed)
7
#631 opened 17 days ago
by
JosephusCheung
submit failed
1
#630 opened 17 days ago
by
jylee420
Title: Disconcerting Perception Surrounding Invisible "Merge" Due to Open LLM Leaderboard Defaults
8
#629 opened 18 days ago
by
gate369
bigstral and bigyi both failed
1
#628 opened 21 days ago
by
ehartford
Requesting to add This model
2
#627 opened 21 days ago
by
Abhaykoul
Mistrals are flooding the top of the leaderboard.
8
#626 opened 21 days ago
by
Phil337
Inquiry Regarding Accuracy Calculation for GSM8K Metric
2
#625 opened 22 days ago
by
qwerwxy
My submission has failed
10
#624 opened 23 days ago
by
MaziyarPanahi
🚩 Report: Not working
2
#623 opened 23 days ago
by
PierrunoYT
Model evaluation failed after 2 days
17
#622 opened 24 days ago
by
migtissera
Temporary removal of models for investigation
13
#621 opened 24 days ago
by
yam-peleg
Code for evaluating new models?
1
#620 opened 24 days ago
by
YannDubs
cognitivecomputations/dolphin-2.8-experiment26-7b
2
#619 opened 24 days ago
by
ehartford
Trying to submit a new model l3utterfly/phi-2-layla-v1, it says already submitted, but I don't see it anywhere
6
#618 opened 25 days ago
by
l3utterfly
Typo (?) regarding the number of tasks on the About page
3
#617 opened 26 days ago
by
TheRootOf3
Model resubmit after wrong dtype used
1
#616 opened 28 days ago
by
DavidGF
Model "{model_name}" was not found on hub!
#615 opened 28 days ago
by
NLUHOPOE
Pending Evaluation
5
#614 opened 29 days ago
by
Kukedlc
Evaluation time
4
#613 opened 30 days ago
by
yleo
Update src/display/about.py
1
#612 opened about 1 month ago
by
lewtun
Is lm-eval-harness stuck again or just taking a while?
1
#611 opened about 1 month ago
by
nisten
Evaluation failed.
3
#610 opened about 1 month ago
by
aigeek0x0
Model stuck in Open LLM eval for over 24hrs
5
#609 opened about 1 month ago
by
CorticalStack