VidEgoThink: Assessing Egocentric Video Understanding Capabilities for Embodied AI Paper • 2410.11623 • Published Oct 15 • 46
StableToolBench: Towards Stable Large-Scale Benchmarking on Tool Learning of Large Language Models Paper • 2403.07714 • Published Mar 12 • 1
ConvLLaVA: Hierarchical Backbones as Visual Encoder for Large Multimodal Models Paper • 2405.15738 • Published May 24 • 43
EgoThink Collection An evaluation benchmark for VLMs from the first-person perspective. • 2 items • Updated Dec 6, 2023 • 1
EgoThink Collection An evaluation benchmark for VLMs from the first-person perspective. • 1 item • Updated Dec 6, 2023 • 1
Can Vision-Language Models Think from a First-Person Perspective? Paper • 2311.15596 • Published Nov 27, 2023 • 2
OpenChat Collection OpenChat: Advancing Open-source Language Models with Mixed-Quality Data • 7 items • Updated Jul 31 • 33
How FaR Are Large Language Models From Agents with Theory-of-Mind? Paper • 2310.03051 • Published Oct 4, 2023 • 34
OpenChat: Advancing Open-source Language Models with Mixed-Quality Data Paper • 2309.11235 • Published Sep 20, 2023 • 16