-
Instruction Pre-Training: Language Models are Supervised Multitask Learners
Paper • 2406.14491 • Published • 86 -
Transformers are SSMs: Generalized Models and Efficient Algorithms Through Structured State Space Duality
Paper • 2405.21060 • Published • 63 -
Perplexed by Perplexity: Perplexity-Based Data Pruning With Small Reference Models
Paper • 2405.20541 • Published • 21 -
MMLU-Pro: A More Robust and Challenging Multi-Task Language Understanding Benchmark
Paper • 2406.01574 • Published • 43
Daeseong Kim
dkimds
·
AI & ML interests
RL, LLM, RLHF and so on.
Recent Activity
updated
a model
about 1 month ago
dkimds/mt0-large-ia3
updated
a model
about 1 month ago
dkimds/peft-vit-base-patch16-224-in21k-lora
updated
a model
about 1 month ago
dkimds/bloomz-560-m-peft-method
Organizations
None yet
Collections
1
models
19
dkimds/mt0-large-ia3
Updated
dkimds/peft-vit-base-patch16-224-in21k-lora
Updated
dkimds/bloomz-560-m-peft-method
Updated
dkimds/rl_course_vizdoom_health_gathering_supreme
Reinforcement Learning
•
Updated
dkimds/ppo-LunarLander-v2
Reinforcement Learning
•
Updated
•
5
dkimds/a2c-PandaReachDense-v3
Reinforcement Learning
•
Updated
dkimds/ppo-SnowballTarget
Reinforcement Learning
•
Updated
•
5
dkimds/ppo-Pyramids-Training
Reinforcement Learning
•
Updated
dkimds/PixelCopter-PLE-v0
Updated
dkimds/Reinforce-CartPole-v1
Reinforcement Learning
•
Updated
datasets
None public yet