Florence-2: Advancing a Unified Representation for a Variety of Vision Tasks Paper • 2311.06242 • Published Nov 10, 2023 • 70
Qwen2 Collection Qwen2 language models, including pretrained and instruction-tuned models of 5 sizes, including 0.5B, 1.5B, 7B, 57B-A14B, and 72B. • 29 items • Updated Jun 6 • 260
OpenRLHF: An Easy-to-use, Scalable and High-performance RLHF Framework Paper • 2405.11143 • Published May 20 • 33
Aya 23: Open Weight Releases to Further Multilingual Progress Paper • 2405.15032 • Published May 23 • 21
Meteor: Mamba-based Traversal of Rationale for Large Language and Vision Models Paper • 2405.15574 • Published May 24 • 52
NV-Embed: Improved Techniques for Training LLMs as Generalist Embedding Models Paper • 2405.17428 • Published May 27 • 15
view article Article Training and Finetuning Embedding Models with Sentence Transformers v3 May 28 • 123
Creación de corpus en comunidad Collection Colección de esfuerzos colaborativos para crear corpus en español de calidad. Toda persona hispanohablante puede contribuir :) • 7 items • Updated 5 days ago • 6
C4AI Aya 23 Collection Aya 23 is an open weights research release of an instruction fine-tuned model with highly advanced multilingual capabilities. • 3 items • Updated May 23 • 40
Critique Models (CM) on the 🤗 Hub Collection This collection contains some Critique Models (CM) for LLM evaluation available in the HuggingFace Hub • 5 items • Updated May 7 • 3
Prometheus 2: An Open Source Language Model Specialized in Evaluating Other Language Models Paper • 2405.01535 • Published May 2 • 109
SystemChat Preferences Collection This collection contains the results of the effort on extending `abacusai/SystemChat-1.1` to convert it into a preference dataset • 12 items • Updated Apr 30 • 1
Capybara Preferences Collection This collection contains the results of the effort on extending `LDJnr/Capybara` to convert it into a preference dataset, with 7B LLMs • 8 items • Updated Apr 17 • 1
Replacing Judges with Juries: Evaluating LLM Generations with a Panel of Diverse Models Paper • 2404.18796 • Published Apr 29 • 67
Preference Fine-Tuning of LLMs Should Leverage Suboptimal, On-Policy Data Paper • 2404.14367 • Published Apr 22 • 1
view article Article Jack of All Trades, Master of Some, a Multi-Purpose Transformer Agent Apr 22 • 76
Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone Paper • 2404.14219 • Published Apr 22 • 243
Is DPO Superior to PPO for LLM Alignment? A Comprehensive Study Paper • 2404.10719 • Published Apr 16 • 3
Best Practices and Lessons Learned on Synthetic Data for Language Models Paper • 2404.07503 • Published Apr 11 • 25
Megalodon: Efficient LLM Pretraining and Inference with Unlimited Context Length Paper • 2404.08801 • Published Apr 12 • 62
Meta Llama 3 Collection This collection hosts the transformers and original repos of the Meta Llama 3 and Llama Guard 2 releases • 5 items • Updated Apr 18 • 647
Zephyr ORPO Collection Models and datasets to align LLMs with Odds Ratio Preference Optimisation (ORPO). Recipes here: https://github.com/huggingface/alignment-handbook • 3 items • Updated Apr 12 • 15
LLM2Vec: Large Language Models Are Secretly Powerful Text Encoders Paper • 2404.05961 • Published Apr 9 • 63
Mixture-of-Depths: Dynamically allocating compute in transformer-based language models Paper • 2404.02258 • Published Apr 2 • 102
Qwen1.5 Collection Qwen1.5 is the improved version of Qwen, the large language model series developed by Alibaba Cloud. • 55 items • Updated Jun 6 • 202
Foundation AI Papers Collection Curated List of Must-Reads on LLM reasoning at Temus AI team • 135 items • Updated Jun 15 • 24
About ORPO Collection Contains some information and experiments fine-tuning LLMs using 🤗 `trl.ORPOTrainer` • 8 items • Updated May 7 • 5
ORPO Collection This is the official collection of "ORPO: Monolithic Preference Optimization without Reference Model". • 5 items • Updated Apr 12 • 10
ORPO: Monolithic Preference Optimization without Reference Model Paper • 2403.07691 • Published Mar 12 • 59
Zephyr 7B Gemma Collection Models, dataset, and Demo for Zephyr 7B Gemma. For code to train the models, see: https://github.com/huggingface/alignment-handbook • 5 items • Updated Apr 12 • 15
Sora: A Review on Background, Technology, Limitations, and Opportunities of Large Vision Models Paper • 2402.17177 • Published Feb 27 • 88
Gemma release Collection Groups the Gemma models released by the Google team. • 40 items • Updated 25 days ago • 320
OLMo Suite Collection Artifacts for the first set of OLMo models. • 14 items • Updated 27 days ago • 43
Understanding LLMs: A Comprehensive Overview from Training to Inference Paper • 2401.02038 • Published Jan 4 • 60
LLM in a flash: Efficient Large Language Model Inference with Limited Memory Paper • 2312.11514 • Published Dec 12, 2023 • 255
Improving Text Embeddings with Large Language Models Paper • 2401.00368 • Published Dec 31, 2023 • 78
Some things are more CRINGE than others: Preference Optimization with the Pairwise Cringe Loss Paper • 2312.16682 • Published Dec 27, 2023 • 5
Notux 8x7B v1 Collection Notux 8x7B v1 model (DPO fine-tune of Mixtral 8x7B Instruct v0.1) and datasets used. More information at https://github.com/argilla-io/notus • 6 items • Updated Dec 29, 2023 • 7
SOLAR 10.7B: Scaling Large Language Models with Simple yet Effective Depth Up-Scaling Paper • 2312.15166 • Published Dec 23, 2023 • 56