metadata
title: HF Papers Search
emoji: 📚
colorFrom: purple
colorTo: blue
sdk: docker
app_port: 7860
short_description: Fast-agent MCP server for Hugging Face Daily Papers search
Community Tool: Production Agents + Eval Workspace
This repo combines two things in one place:
- Production-facing agents/tools for:
- Hugging Face Hub Community API workflows
- Hugging Face Daily Papers search
- Evaluation assets:
- benchmark prompts
- scoring scripts
- generated reports/plots
- interpretation notes
If you're new, start with docs/SPACE.md for a guided map.
For generated metrics and reports, use docs/RESULTS.md.
Quick start
1) Run the production community agent
fast-agent go \
--env .fast-agent \
--agent-cards .fast-agent/tool-cards \
--agent hf_hub_community \
-m "show recent activity for huggingface org"
2) Run the production papers search agent
fast-agent go \
--env .fast-agent \
--agent-cards .fast-agent/tool-cards \
--agent hf_papers_search \
-m "find diffusion papers this month"
Repo layout
Production cards/tools
.fast-agent/tool-cards/hf_hub_community.md.fast-agent/tool-cards/hf_api_tool.py.fast-agent/tool-cards/hf_paper_search.md.fast-agent/tool-cards/hf_papers_tool.py
Eval prompts + expected labels
scripts/hf_hub_community_challenges.txtscripts/hf_hub_community_coverage_prompts.jsonscripts/tool_routing_challenges.txtscripts/tool_routing_expected.json
Eval runners/scorers
scripts/score_hf_hub_community_challenges.pyscripts/score_hf_hub_community_coverage.pyscripts/eval_hf_hub_prompt_ab.pyscripts/score_tool_routing_confusion.pyscripts/run_tool_routing_batch.pyscripts/eval_tool_description_ab.pyscripts/plot_tool_description_eval.py
Eval outputs + analysis docs
docs/hf_hub_community_challenge_report.{md,json}docs/tool_routing_eval/*docs/tool_description_eval/*docs/tool_description_eval_setup.mddocs/tool_description_eval/tool_description_interpretation.md
Notes
- Eval scripts resolve repo paths relative to the script location.
- You can run them from any clone path without editing hardcoded roots.
Run all evals
scripts/run_all_evals.sh