--- license: apache-2.0 tags: - llama-factory - lora - generated_from_trainer model-index: - name: UNAversal-2x7B-v1 results: [] --- # UNAversal-2x7B-v1 Merely Phase 1 UNA, only MLP's and its kinda of a beta. The goal was to produce a small but powerful MoE. This is a 2 MoE model, of 7B each expert. Based on intel-neural series v3. | Tasks |Version|Filter|n-shot| Metric |Value | |Stderr| |--------------|-------|------|-----:|----------|-----:|---|-----:| |arc_challenge |Yaml |none | 25|acc |0.7133|± |0.0132| | | |none | 25|acc_norm |0.7235|± |0.0131| |arc_easy |Yaml |none | 0|acc |0.8674|± |0.0070| | | |none | 0|acc_norm |0.8291|± |0.0077| |boolq |Yaml |none | 0|acc |0.8768|± |0.0057| |lambada_openai|Yaml |none | 0|perplexity|3.6656|± |0.0841| | | |none | 0|acc |0.7017|± |0.0064| |mathqa |Yaml |none | 0|acc |0.3474|± |0.0087| | | |none | 0|acc_norm |0.3585|± |0.0088| |piqa |Yaml |none | 0|acc |0.8411|± |0.0085| | | |none | 0|acc_norm |0.8526|± |0.0083| |sciq |Yaml |none | 0|acc |0.9600|± |0.0062| | | |none | 0|acc_norm |0.9370|± |0.0077|