--- license: apache-2.0 datasets: - HuggingFaceTB/cosmopedia pipeline_tag: text-classification --- An untrained precursor MoE created from Cosmo using mergekit. Gate routing initialized using prompt hidden state method. Five are based on the visualized topic clusters of Cosmopedia data, three are task-oriented. Degenerate layers are 0, 1, and 2 (I believe this means experts will be underutilized for the lowest-level features). Best I could do with test-and-try prompt-based routing. Further research might start from the reversed direction, if available in some interpretability tool (activating layer into prompts).