hfpaper: why every AI agent needs a research librarian
I was feeding my son banana yogurt when I saw Hugging Face announce a new Papers skill for AI agents. Someone replied: "This needs a CLI." So I told my agents to build it. 45 minutes later, hfpaper was live — a single Go binary that gives any agent a research librarian. What it unlocks is bigger than the tool.

I was feeding my son Atlas banana yogurt when I saw Clem Delangue announce that Hugging Face had just made it easier for agents to read research papers. A few tweets down, someone replied: "This needs to be a CLI or MCP."
So I told my agents to build it. 45 minutes later, hfpaper was live.
The problem nobody talks about
AI agents are increasingly asked to do research-grade work. "Read this paper." "What's the state of the art?" "Find me the model that implements this approach."
But they have no good way to access papers. They fetch raw HTML from arXiv, get mangled LaTeX, lose figures, miss metadata. They don't know which papers have model checkpoints or datasets. They can't cite what they read.
Hugging Face built the API. They published a skill that teaches agents how to use it. But the skill is a prompt — it teaches one agent in one session how to make curl calls. It's not composable. It's not installable. It doesn't work outside the context window it lives in.
What hfpaper does
A single Go binary wrapping the entire HF Papers API. 7.8MB. Zero dependencies. --json for agents, --table for humans.
Install however you want:
npx hfpaper search "transformers"
brew install zakelfassi/tap/hfpaper
go install github.com/zakelfassi/hfpaper@latest
Or grab a binary from Releases.
The MCP angle
The real feature: hfpaper mcp starts a Model Context Protocol server. Seven tools, stdio transport. Drop this in your Claude or Cursor config:
{
"mcpServers": {
"hfpaper": {
"command": "hfpaper",
"args": ["mcp"]
}
}
}
Your agent can now search and read papers natively. Not through a prompt that teaches it curl. Through a compiled tool that speaks JSON-RPC. The difference between giving your agent a textbook and giving it a librarian.
The 45-minute build
| Time | What happened |
|---|---|
| 3:31 PM | Saw the tweet. Checked feasibility. |
| 3:33 PM | Spawned a coding agent. |
| 3:37 PM | Binary compiled. All commands working. |
| 3:38 PM | GitHub repo + landing page agent spawned. |
| 3:42 PM | Landing page live. |
| 3:48 PM | Two more agents in parallel: MCP + CLI polish. |
| 3:53 PM | MCP working. v0.2.0 released. |
| 4:08 PM | npm + Homebrew agents spawned. |
| 4:14 PM | All 4 install paths live. |
I was feeding Atlas the entire time.
What this actually unlocks
A CLI that reads papers sounds small. It isn't. The moment an agent can search, read, and cite research natively, the workflows that open up are the ones nobody's built yet.
Daily research pulse. hfpaper daily --trending piped into a cron that filters by your interests — multimodal reasoning, RL from human feedback, whatever your stack touches. Every morning, your agent tells you what moved in the fields you care about. Not a newsletter someone else curated. A digest tuned to your context, built from the primary sources.
Paper-to-experiment pipeline. Your agent reads a paper, identifies the architecture, checks if the model checkpoint exists on HF (hfpaper models <id>), and scaffolds a reproduction. Projects like Paper2Code and AutoReproduce have shown this pipeline works — but they require custom infrastructure. hfpaper is the first step: give the agent the paper. The rest is prompt engineering.
Research-informed building. "What's the current state of the art for X?" is a question every builder asks and every agent fumbles. With MCP wired in, the question becomes a tool call. Your agent searches, reads the abstract, pulls the related models, and answers with citations — not hallucinated references to papers that don't exist, but real links to real work.
Competitive intelligence. What papers are trending in your space today? Which ones have model checkpoints already? Which ones have zero implementations — meaning the gap between paper and product is still open? That's alpha, and it refreshes daily.
The common thread: we're moving from building to discovery. The build step is collapsing toward zero. The hard part is knowing what to build next — and that answer lives in the research frontier. The agent that can read papers is the agent that can find the next thing worth building.
We're at the "Google yet to be born" era of intelligence infrastructure. The papers exist. The models exist. The gap is the connective tissue between knowing and doing. hfpaper is one strand of that tissue.
The meta-question
At roughly the same time I was building hfpaper, Muratcan Koylan — the person who tweeted "this needs a CLI" — was building his own internal skill called context-research that does overlapping work. His skill orchestrates parallel searches, triages by relevance, fetches full content as markdown. Sound familiar?
Two people. Same impulse. Same day. No awareness of each other until after both shipped.
When building takes 45 minutes, duplication becomes the default failure mode. Not because people are careless — because there's no surface where work-in-progress is visible before it's finished. GitHub shows you code after it's pushed. Twitter shows you launches after they ship. Nothing shows you the intent to build while it's still forming.
That's what I'm building with Gabl.us — a communication platform where the unit isn't the avatar, it's the project. Stories scoped to what you're working on. Private by default. Visible to the people who'd want to join. Not a feed of finished products. Not a social network. A place where you can see that someone already started the thing you're about to start, before you both waste 45 minutes arriving at the same binary.
The world doesn't need faster building. It needs fewer parallel builds. The infrastructure for that doesn't exist yet. But the pressure is building — 45 minutes at a time.
Subscribe to the systems briefings
Practical diagnostics for products, teams, and institutions navigating AI-driven change.
Subscribe to the systems briefings. Practical diagnostics for products, teams, and institutions navigating AI-driven change. — Occasional briefs that connect agentic AI deployments, organizational design, and geopolitical coordination. No filler - only the signal operators need.
About the Author
Builder · Founder · Systems engineer
What's next
A few handpicked reads to continue the thread.
- →4 min read
My Tmux Skills Were Replaced by a SKILL.md File
For two years I lived in tmux. Then I started orchestrating an AI swarm through Telegram. My terminal skills didn't atrophy — they got encoded into a markdown file. Now I'm bringing the swarm home.
agents - →8 min read
The Harness
OpenAI shipped Symphony — a daemon that monitors your issue tracker and deploys agents to close tickets. The README says it works best in codebases that have adopted harness engineering. So you click the link. Then you find the Ralph citation. Then it gets interesting.
forgeloop - →5 min read
SkDD: Skills-Driven Development
I accidentally named it SDD in October. Spec-Driven Development claimed that acronym. So: SkDD — the methodology I have been running in production since a git init at 1:54 AM.
skills-driven-development