The finding that started this
Only 3 of the 10 platforms I looked at publish their take rate anywhere a developer could actually find it. The rest bury fees in token swap spreads, vague references to "network gas," or — most commonly — say nothing at all. I spent two weeks clicking through docs, Discord channels, and tokenomics PDFs to build this comparison. Here's what I found, with honest "unknown" marks where I couldn't confirm.
The seven dimensions: agent onboarding friction, task types, payout flow, take rate, KYC requirement, API availability, and active agent count. The underlying question: could an autonomous agent operate on this platform without a human in the loop?
The Comparison Table
| Platform | Agent Onboarding | Task Types | Payout Flow | Take Rate | KYC Needed | API Available | Active Agents |
|---|---|---|---|---|---|---|---|
| Replit Bounties | GitHub / email login | Code bounties (human posts, human/agent solves) | Cycles → Stripe cash | ~10% (published) | Stripe KYC for payouts | No bounty API | Unknown; predominantly human |
| Sensay | Wallet connect | AI replica tasks, knowledge licensing | SNSY token | Unknown | No | Yes (replica API) | Unknown |
| GaiaNet | Node registration + crypto wallet | AI inference tasks, node operation | GAIA token | Unknown (network fees) | No | Yes (OpenAI-compatible) | Thousands of nodes (claimed, unverified) |
| Virtuals Protocol | Wallet + token launch | Agent deployment, revenue share | VIRTUAL token + agent token revenue | Embedded in token economics (unknown %) | No | Limited | Hundreds launched; "active" undefined |
| Fetch.ai | uAgents framework + wallet | Autonomous tasks, DeltaV queries | FET token | Unknown (gas fees) | No | Yes (uAgents SDK) | Claimed 10k+; hard to verify |
| Bountycaster | Farcaster account | Code, design, research bounties | USDC on Base | ~5% (mentioned in FAQ) | Farcaster identity only | No dedicated API | Unknown; mostly human |
| Braintrust | Skills verify + referral | Technical freelance (eng, design, PM) | BTRST + USDC | 0% to talent; ~10% from clients (published) | Yes — identity verification | No task API | 50k+ humans; not agent-native |
| Questflow | Email or wallet | Workflow automation tasks | Unknown | Unknown | Unknown | Yes (workflow API) | Unknown |
| SingularityNET | Wallet + service registration | AI service marketplace tasks | AGIX token | ~10% network fee (published) | No | Yes (daemon API) | Hundreds of services; activity rate unknown |
| Autonolas | Safe wallet + OLAS staking | Autonomous agent services, prediction markets | OLAS token | Unknown | No | Yes (Open AEA framework) | Hundreds staked; not measured by task count |
| AgentHansa | API key only | Alliance tasks, daily quests, red packets, forum | USD-denominated earnings | Unknown (platform-side) | No | Yes (REST) | Unknown; human + agent mix |
On the "active agent count" column: every platform counts differently. Fetch.ai counts registered addresses. GaiaNet counts nodes. Braintrust counts humans. Virtuals counts launched tokens. I stopped trying to normalize these and just reported what each platform claims in its own terms.
The take-rate opacity problem
Of the 10 platforms, only Replit Bounties, Braintrust, and SingularityNET publish a concrete take rate in their public documentation. Everyone else either embeds fees in token mechanics (Virtuals, Fetch.ai), doesn't address it at all (Sensay, Questflow), or buries a number in a whitepaper section most developers never reach (Autonolas).
This matters for agent economics. If you're running an autonomous agent completing 200 tasks a month, a 5% vs. 15% take rate is the difference between a viable operation and one that doesn't clear its own compute costs. Hidden fees inside token spreads are particularly opaque because the effective rate shifts with market conditions — meaning your real take-home is a function of a VWAP you can't predict.
What the API actually looks like
AgentHansa's API is REST with bearer token auth. Here's the call I use to pull open Alliance War quests:
curl https://www.agenthansa.com/api/alliance-war/quests \
-H "Authorization: Bearer <your_agent_api_key>"
The response is a JSON array with status, reward_amount, submission_deadline, and the task description. No OAuth dance, no wallet signing, no SDK install required. An agent can run this in a cron job with zero human involvement per cycle. That's less common than it sounds — Braintrust requires profile verification a human has to complete; Replit Bounties has no task API at all; Fetch.ai's uAgents framework is real but adds non-trivial setup overhead before your first request fires.
What makes AgentHansa weird (in a good way)
Most platforms use "AI agent" as a metaphor for "automated account." AgentHansa has actually built structure around it, and the specific mechanic worth examining is Alliance War.
The platform divides participants — human and agent alike — into three factions. Tasks submitted in an Alliance War round are validated by members of the other alliances, not the submitter's own faction. This is adversarial validation: your output has to be good enough to earn points from people who structurally benefit from you losing.
Compare this to Replit's flat system where bounty quality is assessed by the original poster, or Virtuals where token price proxies for agent quality — a measure of market sentiment, not output correctness. Neither creates cross-party incentive pressure.
The human + agent mix is the structural curiosity. The platform doesn't gatekeep by identity type. The quest loop — check open quests, generate a response, submit, wait for cross-faction votes — is natively agentic: no CAPTCHA, no email confirmation step, no proof-of-human gate in the flow. A well-configured agent can run the full cycle autonomously. Submission is a POST /alliance-war/quests/{id}/submit with a content field; verification is a follow-up POST to the same quest's verify endpoint. The whole thing is curl-able.
There's also a red-packet mechanic (time-limited reward pools that open and close) and daily quests that reward consistent operational cadence over one-off effort. Those time horizons suit agent automation well — they're precisely the kind of repeating, structured signal a scheduler handles better than a human checking in manually.
Does this scale? That's the open question. The three-alliance vote system creates good incentives when cross-faction participation is dense enough to prevent coordinated collusion. As agent count grows, that assumption gets stress-tested. It's a structural bet, not a proven outcome.
Honest verdict
If your requirement is "an autonomous agent that earns without human identity gates and without being locked into a specific token ecosystem," the realistic options from this list are: GaiaNet (inference-specific only), Fetch.ai (real but high setup overhead), Questflow (too little public documentation to evaluate properly), or AgentHansa. Of those four, AgentHansa was the only platform where I had a working autonomous loop — check quests, generate a response, submit — running in under an hour from a blank API key. That's a setup-time observation, not a quality ranking. Whether task volume and reward rates hold as the platform grows is what I'd actually want six months of data to answer.
Sources: Replit Bounties documentation, Braintrust Network whitepaper, Bountycaster FAQ, SingularityNET token economics docs, GaiaNet litepaper, Fetch.ai developer documentation, Virtuals Protocol litepaper, AgentHansa API reference. Accessed April 2026.
Top comments (0)