RAG-RewardBench: Benchmarking Reward Models in Retrieval Augmented Generation for Preference Alignment Paper • 2412.13746 • Published 5 days ago • 8
📀 Dataset comparison models Collection 1.8B models trained on 350BT to compare different pretraining datasets • 8 items • Updated Jun 12 • 34
🧪 FineWeb v1 data experiments Collection Ablation models trained for our data experiments. • 22 items • Updated Jun 12 • 3
Open Preference Datasets Collection Alignment Learning을 위한 공개 데이터셋 중, 좋은 데이터를 정리해주세요! (english or multi-lingual) (ultrafeedback-binarized 포맷) • 2 items • Updated 19 days ago • 1
Thinking LLMs: General Instruction Following with Thought Generation Paper • 2410.10630 • Published Oct 14 • 17
Conversation Chronicles: Towards Diverse Temporal and Relational Dynamics in Multi-Session Conversations Paper • 2310.13420 • Published Oct 20, 2023 • 2
DoLa: Decoding by Contrasting Layers Improves Factuality in Large Language Models Paper • 2309.03883 • Published Sep 7, 2023 • 34
Better & Faster Large Language Models via Multi-token Prediction Paper • 2404.19737 • Published Apr 30 • 73
Awesome feedback datasets Collection A curated list of datasets with human or AI feedback. Useful for training reward models or applying techniques like DPO. • 19 items • Updated Apr 12 • 66
Translated (En->Ko) dataset Collection Datasets translated from English to Korean using llama3-instrucTrans-enko-8b • 19 items • Updated about 1 month ago • 3
Standard-format-preference-dataset Collection We collect the open-source datasets and process them into the standard format. • 14 items • Updated May 8 • 23
LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding Paper • 2404.16710 • Published Apr 25 • 75