DeepSeek-R1 Thoughtology: Let's <think> about LLM Reasoning Paper • 2504.07128 • Published 11 days ago • 53
DeepSeek-R1 Thoughtology: Let's <think> about LLM Reasoning Paper • 2504.07128 • Published 11 days ago • 53
Can Large Language Models Infer and Disagree Like Humans? Paper • 2305.13788 • Published May 23, 2023
Stable Language Model Pre-training by Reducing Embedding Variability Paper • 2409.07787 • Published Sep 12, 2024
Diffusion Models Through a Global Lens: Are They Culturally Inclusive? Paper • 2502.08914 • Published Feb 13
When Tom Eats Kimchi: Evaluating Cultural Bias of Multimodal Large Language Models in Cultural Mixture Contexts Paper • 2503.16826 • Published 23 days ago
Can LVLMs and Automatic Metrics Capture Underlying Preferences of Blind and Low-Vision Individuals for Navigational Aid? Paper • 2502.14883 • Published Feb 15
Sightation Counts: Leveraging Sighted User Feedback in Building a BLV-aligned Dataset of Diagram Descriptions Paper • 2503.13369 • Published 26 days ago • 7
Social Bias Probing: Fairness Benchmarking for Language Models Paper • 2311.09090 • Published Nov 15, 2023 • 2
Can Community Notes Replace Professional Fact-Checkers? Paper • 2502.14132 • Published Feb 19 • 6
Can Community Notes Replace Professional Fact-Checkers? Paper • 2502.14132 • Published Feb 19 • 6
Qorgau: Evaluating LLM Safety in Kazakh-Russian Bilingual Contexts Paper • 2502.13640 • Published Feb 19
Unstructured Evidence Attribution for Long Context Query Focused Summarization Paper • 2502.14409 • Published Feb 20 • 3