Genesis_FineTune_Core_25k (Master Scholar)
Developer / Brand: Within Us AI
A Genesis-style, end-to-end advanced fine-tuning starter pack: five datasets (5k each) packaged together to train an assistant’s behavioral core.
What this pack trains
- Instruction-following (multi-turn) — constraints, formatting, spec compliance
- Tool / function calling — schema-accurate tool calls + observations
- Preference learning (DPO-style) — chosen vs rejected responses
- Long-context retrieval — cite-the-source answering from multiple docs
- Code patch + debugging — unified diffs and test intent (including defensive fixes)
Files (25,000 total)
- instruct_multiturn_5k.jsonl
- tool_use_functioncalling_5k.jsonl
- preference_dpo_5k.jsonl
- longcontext_retrieval_5k.jsonl
- codepatch_debug_5k.jsonl
- README.md
- dataset_card.md
Unified schema
All records share the same top-level schema: id, type, prompt, response, meta, and optional artifacts.
Composition
By type: { "instruction_multiturn": 5000, "tool_use": 5000, "preference_pair": 5000, "long_context_retrieval": 5000, "code_patch": 5000 }
By safety label: { "allowed": 24011, "refuse": 989 }
Attribution
Within Us AI — Genesis_FineTune_Core_25k
License
Apache-2.0