The Idea

I had a simple question: could I use AI to take down a serious open-source bounty? Not a $50 CSS fix — a real one. Something that would require research, strategy, architecture, and execution. I wanted to see how far a single person could get by treating AI as a true collaborator rather than just an autocomplete engine.

Gumroad — the creator economy platform founded by Sahil Lavingia — open-sourced their entire codebase and started putting bounties on GitHub issues. Some small, some big. I didn't know which ones existed or what they involved. So I started with a completely open-ended prompt:

I want to see if we can tackle some of the Gumroad bounties on feature development. Go look online, see if you can find details about one that you're confident you could tackle, and come back with a plan. No code yet, just research and planning.

I gave it no direction on which bounty, no preference on size or complexity. I wanted to see what it would find and how it would reason about the options.

The Research

The AI came back with a detailed breakdown of the bounty landscape:

  • $10,000 — AI-native CLI (#3713): Build a gh-like CLI for the Gumroad API. Standalone TypeScript project, no need to deploy the full Rails stack. Two early submissions, neither polished.
  • $2,500 — Send all missed posts (#1814): Bulk resend button. Well-scoped but requires deploying the entire Gumroad app (Rails + MySQL + Elasticsearch + MongoDB + Redis + Memcached + Sidekiq). ~10 PR attempts, none merged.
  • $50 — Profile settings overflow (#3388): CSS bug. Multiple proposed fixes already.

The recommendation was immediate: go for the $10K CLI. It was standalone (no nightmare deployment), it was the highest bounty, and the existing submissions were thin.

But I wasn't convinced yet. Maybe there was something easier with less competition. So I pushed back:

Go check out five more that have bounties attached to them, and see if you can come up with some that would be easier that have less competition.

It found five more $100 bounties and came back with a blunt assessment: every single one was swarmed. Even $50 bugs had 5-10 people proposing fixes within days. Worse, Sahil requires live deployed previews for every bounty — standing up the full Gumroad stack just to fix a CSS z-index is a terrible ROI.

The conclusion: the small bounties are traps. The big one was actually the best opportunity because it was standalone, greenfield, and the existing competition hadn't delivered anything impressive.

Going Big

That was enough for me. If we were going to do this, we were going all the way. I didn't want to submit another thin API wrapper — I wanted something that would make Sahil sit up in his chair.

Let's just take a stab at the big one, AI-native CLI. I want you to go way above and beyond what anybody would expect for this and make it fully feature-rich, and let's not use naming conventions that are going to conflict. I want you to think of five or ten bells and whistles that we could throw in there that would really wow them. It sounds like you know who Sahil is, so make sure that you understand what type of stuff he is impressed by, and make sure to integrate that. Any sort of Easter eggs or something would be cool as well.

This was where it got interesting. I asked the AI to research Sahil specifically — his values, his writing, what he responds to — and factor that into the design. Not just "build a CLI" but "build a CLI that this specific person will be excited about."

The AI came back with a full plan document and a strategy built around what Sahil values:

  • Minimalism with substance — He wrote The Minimalist Entrepreneur. No bloat. Every line earns its place.
  • AI-first thinking — He uses AI (Devin) to code, open-sourced an AI testing framework. He wants tools that are AI tools.
  • gh CLI ergonomics — He explicitly cited GitHub's CLI as the model. Subcommands, human output plus JSON mode, interactive prompts.

And 10 "wow" features: a built-in MCP server, live sales watcher, status dashboard, browser integration, auto-pagination, interactive prompts, shell completions, multi-account profiles, natural language mode, and Easter eggs.

The Move Nobody Else Made

The plan was already strong, but then I had an idea. The bounty was for an "AI-native" CLI. Everyone else was interpreting that as "a CLI that AI can shell out to." But what if we went further?

What if on top of doing the MCP server, we do things like build out skills that would be compatible with Claude, Claude Code, Cursor, and any of those other types of platforms that look at skills? Anything else like that where it would help a ton of people integrate it into other AI tools?

This turned into the single biggest differentiator. The AI researched every major AI coding tool's integration format and designed a gumroad setup command that auto-installs the right configuration for any tool:

  • Cursor — MCP config, rules file, SKILL.md agent skill
  • Claude Code — CLAUDE.md, rules, MCP config
  • Claude Desktop — claude_desktop_config.json snippet
  • GitHub Copilot / Codex — AGENTS.md template
  • Windsurf — .windsurfrules file
  • Cline / Roo Code — .clinerules/gumroad.md

One command, and your AI assistant knows Gumroad — the API semantics, the CLI commands, common seller workflows. No CLI in existence does this. Not gh, not aws-cli, not Stripe's CLI, not Vercel's CLI.

This was the moment the submission went from "competitive" to "different category." We weren't just wrapping an API — we were setting a new standard for what "AI-native developer tool" means.

The Build

With the plan locked in, I gave the go-ahead:

Do you need to know anything else before building this, or are you good to start?

And then it just... built the whole thing. 40 files of clean TypeScript in a single session. The AI asked a few clarifying questions (npm package name, my GitHub username, whether to use Anthropic or OpenAI for the natural language mode) and then started cranking.

48+ CLI Commands
24 MCP Server Tools
6 AI Tool Integrations
40 Source Files

The full feature set:

Testing It Live

I created a Gumroad API application, got a token, and we tested every command against the real API. Everything worked — product listing, user profiles, status dashboard, the OAuth flow, the setup commands, completions, all of it. We iterated on a few things (fixing display name handling, making the auth process exit cleanly after success), but the core build was solid on the first pass.

The Submission

The Gumroad repo restricts pull requests to collaborators, so the process is to comment on the issue with your implementation. I recorded a Loom walkthrough demonstrating every feature and posted it alongside the GitHub repo link.

Looking at the competition:

Competitor A: Basic CLI with products/sales/licenses commands. No MCP, no AI integrations, no OAuth browser flow.

Competitor B: Similar basic CLI. Also no MCP or AI integrations. Did flag the naming conflict with gr (which we'd already solved by using gumroad as the command name).

Our submission: 48+ commands, built-in MCP server with 24 tools, AI integrations for 6 platforms, OAuth browser login, natural language mode, live sales watcher, status dashboard, shell completions, multi-account profiles, Easter eggs.

What I Learned

A few takeaways from treating AI as a full collaborator on a real project with real stakes:

Prompting is strategy, not instruction. The most valuable prompts weren't "write this function" — they were "go research the competitive landscape" and "understand what this specific person values." Treating the AI like a strategist rather than a typist produced dramatically better results.

Pushing back matters. When the AI recommended the $10K bounty immediately, I said "find me easier ones." It came back with evidence that the easy ones were actually harder (swarmed, bad ROI, deployment barriers). That back-and-forth led to a more informed decision than either of us would have made alone.

The best ideas come from building on AI suggestions. The AI proposed the MCP server. I proposed the multi-tool integration skills on top of that. Neither of us would have gotten there alone. The compounding effect of iterating on each other's ideas is where the real leverage is.

AI can do the 0-to-1, but you need to direct the 1-to-100. Left to its own devices, the AI would have built a competent CLI. It was the human direction — "go way above and beyond," "understand Sahil," "build skills for every AI tool" — that turned it into something that stands out.

Current Status

The submission is live on the GitHub issue. As of this writing, Sahil hasn't responded yet. The bounty is still open. If you're reading this, Sahil — check your notifications.

Regardless of the outcome, this was a proof of concept. One person, one AI, one session, going after a $10,000 bounty against professional developers — and producing something that meaningfully outscoped the competition. The tools are here. The question is just how ambitious you're willing to be with them.

The Demo