Edit model card

Given-MPT-7B

This is a merge of the following MPT-7B models:

  • gorilla-llm/gorilla-mpt-7b-hf-v0
  • ibm/mpt-7b-instruct2
  • TehVenom/MPT-7b-WizardLM_Uncensored-Storywriter-Merge
  • emozilla/mpt-7b-storysummarizer
  • nomic-ai/gpt4all-mpt

Model License

Apache 2.0

Purpose

This model is for experimenting with merging and routing to expert layers.

Test eval on only 10% of eval set

hf-causal (pretrained=Multi-Domain-Expert-Layers/given-mpt-7b,dtype=bfloat16,trust_remote_code=True), limit: 0.1, provide_description: False, num_fewshot: 0, batch_size: None

Task Version Metric Value Stderr
arc_challenge 0 acc 0.4274 ± 0.0459
acc_norm 0.3846 ± 0.0452
arc_easy 0 acc 0.7863 ± 0.0381
acc_norm 0.7350 ± 0.0410
hellaswag 0 acc 0.5556 ± 0.0461
acc_norm 0.8120 ± 0.0363
hendrycksTest-college_chemistry 0 acc 0.3600 ± 0.0482
acc_norm 0.3700 ± 0.0485
hendrycksTest-college_computer_science 0 acc 0.3400 ± 0.0476
acc_norm 0.3600 ± 0.0482
hendrycksTest-college_mathematics 0 acc 0.2500 ± 0.0435
acc_norm 0.2900 ± 0.0456
hendrycksTest-college_medicine 0 acc 0.3675 ± 0.0448
acc_norm 0.3162 ± 0.0432
hendrycksTest-college_physics 0 acc 0.2451 ± 0.0428
acc_norm 0.2941 ± 0.0453
hendrycksTest-computer_security 0 acc 0.4800 ± 0.0502
acc_norm 0.4400 ± 0.0499
hendrycksTest-conceptual_physics 0 acc 0.2051 ± 0.0375
acc_norm 0.1709 ± 0.0350
hendrycksTest-econometrics 0 acc 0.2982 ± 0.0430
acc_norm 0.2368 ± 0.0400
hendrycksTest-electrical_engineering 0 acc 0.3248 ± 0.0435
acc_norm 0.3590 ± 0.0445
hendrycksTest-elementary_mathematics 0 acc 0.3333 ± 0.0438
acc_norm 0.3162 ± 0.0432
hendrycksTest-formal_logic 0 acc 0.3077 ± 0.0429
acc_norm 0.3248 ± 0.0435
hendrycksTest-global_facts 0 acc 0.3000 ± 0.0461
acc_norm 0.2700 ± 0.0446
hendrycksTest-high_school_biology 0 acc 0.3675 ± 0.0448
acc_norm 0.3077 ± 0.0429
hendrycksTest-high_school_chemistry 0 acc 0.2564 ± 0.0405
acc_norm 0.2906 ± 0.0422
hendrycksTest-high_school_computer_science 0 acc 0.4100 ± 0.0494
acc_norm 0.4400 ± 0.0499
hendrycksTest-high_school_european_history 0 acc 0.4359 ± 0.0460
acc_norm 0.3590 ± 0.0445
hendrycksTest-high_school_geography 0 acc 0.3248 ± 0.0435
acc_norm 0.3675 ± 0.0448
hendrycksTest-high_school_government_and_politics 0 acc 0.3932 ± 0.0454
acc_norm 0.3932 ± 0.0454
hendrycksTest-high_school_macroeconomics 0 acc 0.3333 ± 0.0438
acc_norm 0.3248 ± 0.0435
hendrycksTest-high_school_mathematics 0 acc 0.2051 ± 0.0375
acc_norm 0.2564 ± 0.0405
hendrycksTest-high_school_microeconomics 0 acc 0.3504 ± 0.0443
acc_norm 0.4188 ± 0.0458
hendrycksTest-high_school_physics 0 acc 0.2650 ± 0.0410
acc_norm 0.2906 ± 0.0422
hendrycksTest-high_school_psychology 0 acc 0.3761 ± 0.0450
acc_norm 0.3419 ± 0.0440
hendrycksTest-high_school_statistics 0 acc 0.3077 ± 0.0429
acc_norm 0.3504 ± 0.0443
hendrycksTest-high_school_us_history 0 acc 0.3333 ± 0.0438
acc_norm 0.3333 ± 0.0438
hendrycksTest-high_school_world_history 0 acc 0.3333 ± 0.0438
acc_norm 0.3419 ± 0.0440
hendrycksTest-human_aging 0 acc 0.3761 ± 0.0450
acc_norm 0.3162 ± 0.0432
hendrycksTest-human_sexuality 0 acc 0.4274 ± 0.0459
acc_norm 0.3761 ± 0.0450
hendrycksTest-international_law 0 acc 0.4188 ± 0.0458
acc_norm 0.4957 ± 0.0464
hendrycksTest-jurisprudence 0 acc 0.3148 ± 0.0449
acc_norm 0.4815 ± 0.0483
hendrycksTest-logical_fallacies 0 acc 0.3504 ± 0.0443
acc_norm 0.3675 ± 0.0448
hendrycksTest-machine_learning 0 acc 0.3214 ± 0.0443
acc_norm 0.2946 ± 0.0433
hendrycksTest-management 0 acc 0.3786 ± 0.0480
acc_norm 0.3495 ± 0.0472
hendrycksTest-marketing 0 acc 0.5043 ± 0.0464
acc_norm 0.4188 ± 0.0458
hendrycksTest-medical_genetics 0 acc 0.3200 ± 0.0469
acc_norm 0.4100 ± 0.0494
hendrycksTest-miscellaneous 0 acc 0.5299 ± 0.0463
acc_norm 0.4872 ± 0.0464
hendrycksTest-moral_disputes 0 acc 0.3248 ± 0.0435
acc_norm 0.3162 ± 0.0432
hendrycksTest-moral_scenarios 0 acc 0.3248 ± 0.0435
acc_norm 0.2479 ± 0.0401
hendrycksTest-nutrition 0 acc 0.3675 ± 0.0448
acc_norm 0.3932 ± 0.0454
hendrycksTest-philosophy 0 acc 0.2991 ± 0.0425
acc_norm 0.3504 ± 0.0443
hendrycksTest-prehistory 0 acc 0.2821 ± 0.0418
acc_norm 0.3248 ± 0.0435
hendrycksTest-professional_accounting 0 acc 0.2137 ± 0.0381
acc_norm 0.2222 ± 0.0386
hendrycksTest-professional_law 0 acc 0.3077 ± 0.0429
acc_norm 0.2735 ± 0.0414
hendrycksTest-professional_medicine 0 acc 0.2991 ± 0.0425
acc_norm 0.2650 ± 0.0410
hendrycksTest-professional_psychology 0 acc 0.3248 ± 0.0435
acc_norm 0.3419 ± 0.0440
hendrycksTest-public_relations 0 acc 0.3909 ± 0.0467
acc_norm 0.3545 ± 0.0458
hendrycksTest-security_studies 0 acc 0.3419 ± 0.0440
acc_norm 0.2906 ± 0.0422
hendrycksTest-sociology 0 acc 0.3761 ± 0.0450
acc_norm 0.3162 ± 0.0432
hendrycksTest-us_foreign_policy 0 acc 0.5000 ± 0.0503
acc_norm 0.4100 ± 0.0494
hendrycksTest-virology 0 acc 0.3932 ± 0.0454
acc_norm 0.3248 ± 0.0435
hendrycksTest-world_religions 0 acc 0.5299 ± 0.0463
acc_norm 0.5128 ± 0.0464
truthfulqa_gen 1 bleurt_max -0.8551 ± 0.0501
bleurt_acc 0.3590 ± 0.0445
bleurt_diff -0.1292 ± 0.0483
bleu_max 19.3738 ± 1.8461
bleu_acc 0.3932 ± 0.0454
bleu_diff -4.3883 ± 2.1748
rouge1_max 41.8428 ± 2.6156
rouge1_acc 0.3162 ± 0.0432
rouge1_diff -8.8583 ± 2.7745
rouge2_max 26.3956 ± 2.8311
rouge2_acc 0.2137 ± 0.0381
rouge2_diff -9.5287 ± 3.3258
rougeL_max 39.5215 ± 2.5620
rougeL_acc 0.3162 ± 0.0432
rougeL_diff -8.5753 ± 2.8259
Downloads last month
7
Inference Examples
Inference API (serverless) has been turned off for this model.