TransformerFAM: Feedback attention is working memory Paper • 2404.09173 • Published about 1 month ago • 42
view article Article Jack of All Trades, Master of Some, a Multi-Purpose Transformer Agent 22 days ago • 69
Can Mamba Learn How to Learn? A Comparative Study on In-Context Learning Tasks Paper • 2402.04248 • Published Feb 6 • 25
Large Language Models as Generalizable Policies for Embodied Tasks Paper • 2310.17722 • Published Oct 26, 2023 • 6
Retroformer: Retrospective Large Language Agents with Policy Gradient Optimization Paper • 2308.02151 • Published Aug 4, 2023 • 18