-
Attention Is All You Need
Paper • 1706.03762 • Published • 35 -
MetaGPT: Meta Programming for Multi-Agent Collaborative Framework
Paper • 2308.00352 • Published • 2 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 11 -
XLNet: Generalized Autoregressive Pretraining for Language Understanding
Paper • 1906.08237 • Published
Collections
Discover the best community collections!
Collections including paper arxiv:1706.03762
-
Attention Is All You Need
Paper • 1706.03762 • Published • 35 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 11 -
Improving Text Embeddings with Large Language Models
Paper • 2401.00368 • Published • 73 -
Gemini: A Family of Highly Capable Multimodal Models
Paper • 2312.11805 • Published • 44
-
SMOTE: Synthetic Minority Over-sampling Technique
Paper • 1106.1813 • Published • 1 -
Scikit-learn: Machine Learning in Python
Paper • 1201.0490 • Published • 1 -
Identity Mappings in Deep Residual Networks
Paper • 1603.05027 • Published • 2 -
Deep Residual Learning for Image Recognition
Paper • 1512.03385 • Published • 5
-
Attention Is All You Need
Paper • 1706.03762 • Published • 35 -
ImageNet Large Scale Visual Recognition Challenge
Paper • 1409.0575 • Published • 6 -
Sequence to Sequence Learning with Neural Networks
Paper • 1409.3215 • Published • 3 -
Language Models are Few-Shot Learners
Paper • 2005.14165 • Published • 9
-
Understanding LLMs: A Comprehensive Overview from Training to Inference
Paper • 2401.02038 • Published • 59 -
The Impact of Reasoning Step Length on Large Language Models
Paper • 2401.04925 • Published • 15 -
Lost in the Middle: How Language Models Use Long Contexts
Paper • 2307.03172 • Published • 31 -
Attention Is All You Need
Paper • 1706.03762 • Published • 35
-
Attention Is All You Need
Paper • 1706.03762 • Published • 35 -
You Only Look Once: Unified, Real-Time Object Detection
Paper • 1506.02640 • Published -
HEp-2 Cell Image Classification with Deep Convolutional Neural Networks
Paper • 1504.02531 • Published -
Sleeper Agents: Training Deceptive LLMs that Persist Through Safety Training
Paper • 2401.05566 • Published • 23
-
LoRA: Low-Rank Adaptation of Large Language Models
Paper • 2106.09685 • Published • 24 -
Attention Is All You Need
Paper • 1706.03762 • Published • 35 -
Direct Preference Optimization: Your Language Model is Secretly a Reward Model
Paper • 2305.18290 • Published • 37 -
Lost in the Middle: How Language Models Use Long Contexts
Paper • 2307.03172 • Published • 31