Archangel is a suite of human feedback-aligned LLMs, released as part of the Human-Aware Loss Functions (HALOs) project by Ethayarajh et al. (2024).
ContextualAI
company
Verified
AI & ML interests
Enterprise-grade retrieval augmented language models
models
81
ContextualAI/tiny-random-MistralForCausalLM
Text Generation
•
Updated
•
3.72k
ContextualAI/zephyr_sft_dpo
Text Generation
•
Updated
•
4
ContextualAI/zephyr_sft_kto_unary
Text Generation
•
Updated
•
4
ContextualAI/zephyr_sft_kto
Text Generation
•
Updated
•
9
•
1
ContextualAI/Contextual_KTO_Mistral_PairRM
Text Generation
•
Updated
•
131
•
30
ContextualAI/archangel_kto_pythia6-9b
Text Generation
•
Updated
•
11
ContextualAI/archangel_kto_pythia2-8b
Text Generation
•
Updated
•
70
ContextualAI/archangel_kto_pythia1-4b
Text Generation
•
Updated
•
7
ContextualAI/archangel_ppo_pythia2-8b
Text Generation
•
Updated
•
15
ContextualAI/archangel_ppo_pythia1-4b
Text Generation
•
Updated
•
7
datasets
14
ContextualAI/ultrafeedback_clair_32k
Viewer
•
Updated
•
29.1k
•
2
•
4
ContextualAI/ultrafeedback_rlaif-offpolicy_32k
Viewer
•
Updated
•
29.1k
•
2
•
1
ContextualAI/ultrafeedback_stronger-preferred_32k
Viewer
•
Updated
•
29.1k
•
15
•
1
ContextualAI/ultrafeedback_rlaif_32k
Viewer
•
Updated
•
31.8k
•
2
•
1
ContextualAI/ultrabin_clean_max_chosen_min_rejected_rationalized
Viewer
•
Updated
•
60.9k
•
2
ContextualAI/ultrabin_clean_max_chosen_rand_rejected_rationalized
Viewer
•
Updated
•
60.9k
•
2
ContextualAI/ultrabin_clean_max_chosen_min_rejected_rationalized_helpfulness
Viewer
•
Updated
•
60.9k
•
2
ContextualAI/ultrabin_clean_max_chosen_min_rejected_rationalized_truthfulness
Viewer
•
Updated
•
60.9k
•
2
ContextualAI/ultrabin_clean_max_chosen_min_rejected_rationalized_honesty
Viewer
•
Updated
•
60.9k
•
2
ContextualAI/ultrabin_clean_max_chosen_min_rejected_rationalized_instruction_following
Viewer
•
Updated
•
60.9k
•
2
•
3