skillpack.co

All Solutions

200 solutions tracked across 23 problem spaces. Each with an editorial verdict, source links, and evidence.

200

Solutions

23

Problems

200

Figma MCP Server Guide

activeOfficial74
figma/mcp-server-guide520 6 evidence

The trust leader with unmatched long-term durability — triple partnership with OpenAI Codex, GitHub Copilot, and Claude Code. Bidirectional since Mar 6 2026. Wins specifically when your team has Code Connect configured. Without it, Framelink produces cleaner context. Pricing is the primary barrier: free tier (6 calls/month) is non-functional.

UX / UI

Framelink / Figma-Context-MCP

active86
GLips/Figma-Context-MCP14K+ 6 evidence

Best read-only design-to-code MCP. Use this unless you have Code Connect configured — 25% smaller output than Official, avoids prescriptive code that poisons LLM context. Works on free Figma plans with broadest editor support.

UX / UI

Figma-use

active63
dannote/figma-use514 3 evidence

Conditional pick. Strongest HN validation in the write-access lane, but solo maintainer and potential OpenPencil pivot raise durability concerns. Use if you want CLI-first workflow without a Figma plugin.

UX / UI

Vibma

watch63
ufira-ai/Vibma468 2 evidence

Emerging. Only tool publishing model-specific design quality benchmarks. Harness engineering angle is genuinely differentiated, but HN Show got only 2 points and no independent reviews found.

UX / UI

Cursor Talk to Figma

active83
grab/cursor-talk-to-figma-mcp6.5K+ 4 evidence

Best general-purpose write-access Figma MCP for individual developers and small teams. More accessible than Console MCP, works on free Figma plans, lower setup complexity. Dropped to #4 as Console MCP's Uber validation is stronger enterprise evidence.

UX / UI

Figma Console MCP

active85
southleft/figma-console-mcp1.2K+ 4 evidence

The enterprise write-access leader. Uber's production uSpec validates this at scale — automated component specs across 7 stacks, accessibility in under 2 minutes. Hockey-stick npm trajectory is fastest-growing in the category. Local WebSocket means no rate limits and no data leaving your network.

UX / UI

Penpot MCP

watchOfficial54
penpot/penpot-mcp-server225 2 evidence

The open-source design MCP lane exists but isn't production-ready yet. MCP now integrated into main Penpot repo (archived standalone). Watch item — if development resumes and beta launches, it moves up.

UX / UI

Claude Talk to Figma

active59
arinspunk/claude-talk-to-figma516 1 evidence

Best Claude-specific write-access option. DXT installer reduces setup to one click for Claude Desktop. If you're all-in on Claude Code or Claude Desktop and need write-access to Figma without paying for Dev Mode, this is the natural pick.

UX / UI

Onlook

active84
onlook-dev/onlook25K+ 2 evidence

Different lane from Figma MCP tools — not a Figma bridge, but a direct design-in-code editor. 24,918 stars and two HN hits above 200 points make this the most-starred tool in the UX/UI category. Best for frontend teams where designers work directly in the Next.js + Tailwind codebase and want to eliminate the Figma-to-code translation step.

UX / UI

Excalidraw MCP

activeOfficial71
excalidraw/excalidraw-mcp3.5K+ 1 evidence

Clear diagramming lane leader. 3,371 stars and official Excalidraw org backing make it the default for AI-assisted diagramming. Different workflow from design-to-code — listed in UX/UI for visibility but doesn't compete with Figma MCPs.

UX / UI

Kombai

active40
2 evidence

Wins on fidelity in every controlled comparison found. FreeCodeCamp tested 75–80% accuracy vs 65–70% for Figma MCP+LLM. Not an MCP server — a specialized Figma plugin + proprietary AI engine. Best for teams where pixel-perfect output matters more than transparency or cost.

UX / UI

Google Stitch

watchOfficial40
2 evidence

Provisional #1 in AI-native design creation. Google-backed, free, and the Mar 18 overhaul caused an 8.8% Figma stock drop — the market treats this as a serious threat to Figma's moat. MCP server confirmed. But the update is 2 days old with zero usage evidence. Watch status until adoption data appears.

UX / UI

Firecrawl MCP Server

activeOfficial72
firecrawl/firecrawl-mcp-server5.8K+ 91%12 evidence

#1 in product-business-development by volume and evidence. 5,809 stars, 50.6K npm/wk (highest in category), 83% benchmark accuracy (AIMultiple), F1 0.638 (SearchMCP). Morph confirms 89% recall, 95.3% success rate. Four independent sources confirm extraction superiority.

Product / Business DevelopmentSearch & News

Exa MCP Server

activeOfficial87
exa-labs/exa-mcp-server4.1K+ 13 evidence

#8 in product-business-development, #2 research tool. 4,050 stars, 9.3K npm/wk, 19.7K PulseMCP/wk (#56 global). Best for semantic discovery ('find companies like X'). exa-code for coding agents is a unique differentiator. AIMultiple benchmark tested extraction, not discovery — 23% undersells Exa's actual strength.

Product / Business DevelopmentSearch & News

MCP Atlassian

active85
sooperset/mcp-atlassian4.7K+ 7 evidence

#2 overall, #1 enterprise operating surface for teams on the Atlassian stack. 4,666 stars, deepest tool set (72 tools), 139K PyPI weekly downloads, 140K PulseMCP/wk (#14 global). On-prem/Data Center support is the key moat vs official Rovo MCP.

Product / Business Development

Notion MCP Server

activeOfficial86
makenotion/notion-mcp-server4.1K+ 7 evidence

#3 overall, #1 startup operating surface — 47K+ npm weekly downloads (17x Atlassian official MCP). Token-optimized Markdown responses show MCP-specific engineering. Notion 3.3 Custom Agents ecosystem positions MCP as multi-tool hub.

Product / Business Development

Slack MCP Server

active72
korotovsky/slack-mcp-server1.5K+ 5 evidence

#10 in product-business-development. Communication lane — real but early. korotovsky community server at 1,461 stars shows genuine developer adoption. Official GA with 50+ partners (Anthropic, Google, OpenAI, Perplexity). PulseMCP 1.6K/wk (#602 weekly), 23.7K all-time. Multiple independent publications covered GA launch (KMWorld, SmallBizTrends, Salesforce Ben). 25x growth claim is selfReported with no disclosed baseline.

Product / Business Development

Google Workspace MCP

active80
taylorwilsdon/google_workspace_mcp1.9K+ 4 evidence

#6 overall. Best broad operating skill for teams living in Google Workspace. Breadth is real (100+ tools, 12+ services), but low npm adoption (604/week), community-only status, and existential threat from Google's own CLI weigh against it.

Product / Business Development

OpenHands

active88
OpenHands/OpenHands70K+ 12%11 evidence

Category leader. No other tool has the combination of open-source community (69K stars, 455 contributors), real download volume (1M/month), venture backing ($18.8M), hardware partnerships (AMD), and benchmark leadership (SWE-bench Verified 72%). Gap to #2 is enormous.

Coding CLIs / Code AgentsTeams of Agents / Multi-Agent OrchestrationSoftware Factories

Ralph Loop Agent

activeOfficial60
vercel-labs/ralph-loop-agent733 7 evidence

Best reference when the team wants a crisp loop pattern instead of a huge agent platform. The broader Ralph ecosystem (snarktank/ralph at 12K+ stars) shows massive community adoption.

Teams of Agents / Multi-Agent Orchestration

SWE-agent

stale79
SWE-agent/SWE-agent19K+ 2 evidence

Research/academic reference only. Princeton pedigree and 79.2% SWE-bench Verified on Opus 4.5 scaffold give it strong benchmark credibility. But no release in 10 months (last: v1.1.0, 2025-05-22) puts it outside the production cadence of all active tools. Use as a benchmark scaffold reference, not as a production coding CLI.

Coding CLIs / Code AgentsTeams of Agents / Multi-Agent Orchestration

Claude Code

activeOfficial98
anthropics/claude-code82K+ 121%23 evidence

The #1 coding CLI agent. Leads SWE-bench Pro standardized (45.89%), wins independent head-to-heads on reasoning depth, ~4% of GitHub commits. Rate limits are the #1 complaint. Costs 2-3x more per task than Codex CLI due to higher token consumption.

Coding CLIs / Code AgentsTeams of Agents / Multi-Agent OrchestrationSoftware Factories

Aider

active86
Aider-AI/aider42K+ 11%4 evidence

#7 coding CLI — strong verification story (191K PyPI/week, 5.7M lifetime), but Codex CLI at 2.49M npm/week and Gemini CLI at 678K/week have overtaken Aider's download rank. Category pressure is real: HN thread 'Claude Code with Sonnet 4 is so good I have stopped using Aider' (#44154020). Best for Python devs who want fine-grained model control and git-native workflow. v0.86.2 (2026-02-12) is 5 weeks behind competitors shipping daily.

Coding CLIs / Code AgentsSoftware Factories

Continue (Continuous AI)

active82
continuedev/continue32K+ 2 evidence

Best pick for teams that want background AI agents enforcing code quality on PRs, not real-time autocomplete. The pivot repositioned it away from individual devs toward team CI workflows.

Coding CLIs / Code AgentsTeams of Agents / Multi-Agent Orchestration

OpenCode

active88
sst/opencode130K+ 6 evidence

Watch — two serious security incidents (unauthenticated RCE fixed v1.1.10+, CVE-2026-22812 CVSS 8.8-10.0) make trust story the weakest in the category. 126K+ stars are real but star surge driven by Anthropic OAuth controversy — brand-driven, not organic product traction. OpenAI partnership and 393K npm/week are real signals, but security history removes it from the main ranking pending security posture improvement.

Coding CLIs / Code Agents

Codex CLI

activeOfficial87
openai/codex67K+ 8%7 evidence

#2 coding CLI. Rust rewrite eliminates Node.js dependency — unique in category. Terminal-Bench 77.3% (#2) and 3-4x more token-efficient. Cleanest security record among Tier 1 tools. GPT-5.4 shipped March 2026. Trails Claude Code by ~5pp on SWE-bench Pro standardized (41.04% vs 45.89%) and first-pass quality (67% vs 95%).

Coding CLIs / Code AgentsSoftware Factories

Gemini CLI

activeOfficial88
google-gemini/gemini-cli99K+ 15%11 evidence

#3 coding CLI — best free entry point and Terminal-Bench leader (78.4% #1). 1K req/day free tier is unmatched, 1M context is the largest. SWE-bench Pro standardized 43.30% is competitive. Plan Mode (March 2026) closes last major feature gap. File deletion incident (AI Incident Database #1178) and 50-60% first-pass correctness (roughly half of Claude Code) are real concerns.

Coding CLIs / Code AgentsSoftware Factories

Cline (cline.bot)

watch73
59K+ 5 evidence

#4 coding CLI — strongest in the IDE-embedded-agent segment with 3.35M VS Code installs, 5M total across editors, $32M funding (Emergence Capital), and named enterprise customers (Salesforce, Samsung, SAP). The supply chain incident (v2.3.0 'OpenClaw') is a documented trust flag that still applies. A credible third-party security audit would remove the primary concern. Best for developers who live in VS Code and want agentic assistance without leaving the editor.

Coding CLIs / Code AgentsSoftware Factories

GitHub Copilot CLI

activeOfficial10
github/gh-copilot1.1K+ 6 evidence

#3 coding CLI — distribution moat (15M Copilot subscribers) and the Enterprise Agent Control Plane are unique advantages, and the v1.0.11 release shows GitHub is iterating quickly. But PromptArmor’s remote-code advisory plus the lack of public benchmarks keep it below Claude/Gemini until security policies catch up.

Coding CLIs / Code Agents

Amp (Amp Inc.)

active52
N/A 2 evidence

Watch — corporate spin-out from Sourcegraph to independent Amp Inc. (March 2026) is a material change. The original sourcegraph/amp GitHub link returns 404. Tool still ships (36K npm downloads/week) under ampcode.com. Sub-agent architecture (Oracle, Librarian, Painter) remains the most sophisticated in the category. Update or verify all links before recommending.

Coding CLIs / Code AgentsSoftware Factories

Goose (Block)

active82
block/goose34K+ 4 evidence

#7 coding CLI — strongest free, open-source terminal agent. MCP-native, Apache 2.0, Linux Foundation governance. ACP integration (March 19, 2026) lets developers use existing Copilot/Claude/Gemini subscriptions. 60% of Block's 12K employees use it weekly (self-reported). No published benchmarks keep it in Tier 2.

Coding CLIs / Code AgentsSoftware Factories

Crush (Charmbracelet)

active82
charmbracelet/crush22K+ 2 evidence

Best choice for developers who live in the terminal and want a polished, non-VS-Code-dependent experience. Charmbracelet's proven track record (Bubble Tea, 25K+ apps) is a strong prior. No published benchmark scores — community quality signal is the main trust anchor.

Coding CLIs / Code Agents

Qwen Code (Alibaba / QwenLM)

activeOfficial82
QwenLM/qwen-code21K+ 2 evidence

Best cost=zero option for developers who want a serious open-weight model without paying per token. The 70.6% SWE-bench Verified figure (Qwen3-Coder-Next) is the strongest open-weight score in the category and anchors the strongest local/on-prem story. Alibaba/Chinese cloud provenance is a consideration for enterprise and GovCloud use cases.

Coding CLIs / Code Agents

Junie CLI (JetBrains)

watchOfficial35
N/A 1 evidence

Too early to rank — no public artifact (no repo, no benchmark, no independent review). JetBrains has 11M+ paid IDE seats; if the JetBrains installed base converts, this becomes a serious Tier 2 contender within 60 days. Check back after first independent comparisons.

Coding CLIs / Code Agents

Auggie CLI (Augment Code)

watchOfficial53
153 2 evidence

Highest SWE-bench Pro number in the category (51.80% on Augment scaffold), but the scaffold is not standardized — they used the same Opus 4.5 model that scores 45.89% on SEAL's standardized setup. The architecture/scaffolding advantage is credible and meaningful. Cannot rank above tools with millions of verified installs on a single blog-post benchmark. Watch for: public GA release and independent SWE-bench reproduction.

Coding CLIs / Code Agents

Browser Use

active76
browser-use/browser-use84K+ 11 evidence

Unchallenged category leader — 81K stars, 1M+ weekly PyPI downloads, 89.1% WebVoyager. The gap to #2 is enormous: 4x stars, 6x downloads vs the next autonomous agent.

Web Browsing / Browser Automation

Playwright MCP

activeOfficial93
microsoft/playwright-mcp30K+ 10 evidence

Highest raw downloads in category (1.38M npm/wk). Cross-browser (Chromium + Firefox + WebKit). CLI mode is the token-efficient choice — 4x reduction vs MCP confirmed by 13+ independent sources. Microsoft officially recommends CLI for coding agents.

Web Browsing / Browser Automation

Stagehand

activeOfficial90
browserbase/stagehand22K+ 8 evidence

Best SDK for building browser automation into products. Three clean primitives (act, extract, observe), v3 dropped Playwright dependency, Cloudflare official integration. New browse-cli shows the team reads the CLI-over-MCP signal correctly.

Web Browsing / Browser Automation

Chrome DevTools MCP

activeOfficial92
ChromeDevTools/chrome-devtools-mcp31K+ 6 evidence

Lane 2 leader for Chrome debugging workflows. Fastest star growth in category. 599 HN pts (Mar 15) is the highest single thread. Google Chrome team official. Deep debugging (heap snapshots, Lighthouse, performance profiling) that no competitor matches.

Web Browsing / Browser Automation

Vercel Agent Browser

active88
vercel-labs/agent-browser25K+ 6 evidence

Token-efficiency leader but immature. Anomalous star-to-HN ratio (23.6K stars, ~6 HN pts) warrants caution — community hasn't organically validated this tool yet. 303 open issues in 10 weeks. Best only when token budget is the binding constraint.

Web Browsing / Browser Automation

Skyvern

active90
Skyvern-AI/skyvern21K+ 6 evidence

Best pick for enterprise workflow automation on websites without APIs — form filling, data entry, procurement. Overkill for developer/coding agent browser tasks.

Web Browsing / Browser Automation

Lightpanda

active84
lightpanda-io/browser25K+ 3 evidence

Best infrastructure pick for teams running browser automation at scale who need to reduce costs. Drop-in CDP replacement for Chrome headless in pipelines. Still beta — not a user-facing agent tool.

Web Browsing / Browser Automation

Vibium

watch74
VibiumDev/vibium2.7K+ 3 evidence

Below cut line — only 2.7K stars, not production-ready by creator's own characterization. But founder pedigree (Selenium, Appium creator), 443-point HN thread, near-daily releases, and W3C standards-first architecture demand close tracking. The only Lane 2 tool with credible cross-browser (Firefox + Safari) story.

Web Browsing / Browser Automation

BrowserOS / Nxtscape

active82
nxtscape/nxtscape10K+ 4 evidence

Lane 4 leader. Just crossed 10K stars. 314 HN pts with 206 comments is strong organic engagement. YC S24 backing. Privacy-first Chromium fork with built-in AI. Different market than Lanes 1-3 but clear leader in consumer agentic browsers.

Web Browsing / Browser Automation

Salesforce MCP

activeOfficial80
salesforcecli/mcp327 3 evidence

#2 CRM lane, behind HubSpot. 26.6K npm/week is primarily Salesforce DX developer tooling (SFDX developers), not CRM product teams. Agentforce-gated open CRM access keeps PulseMCP low (~2K/wk). Right choice only for existing Salesforce Enterprise customers already committed to Agentforce.

Product / Business Development

Atlassian Rovo MCP

activeOfficial56
444 2 evidence

#11 in product-business-development. Official backing, 459 stars, GA since Feb 2026. 12+ AI client partnerships. Cloud-native with OAuth 2.1 and Cloudflare Agents SDK. Cloud-only limitation and 200:1 developer preference gap vs sooperset keep it behind. ~2K PulseMCP/wk.

Product / Business Development

Airtable MCP Server

active75
domdomegg/airtable-mcp-server429 1 evidence

#15 in product-business-development. Niche but real — 2.5K npm/week shows genuine adoption. Fills the gap between spreadsheets and databases for non-dev business teams using Airtable as their project database.

Product / Business Development

Linear MCP Server

active57
jerhadf/linear-mcp-server347 4 evidence

#7 overall, #1 in project/PM lane. Feb 2026 PM upgrade (triage, backlog prioritization, initiative creation, milestone management) elevated from dev-only to full PM surface. Developer-first API returns clean typed objects — less context token waste than Jira. ~2,743 combined npm/wk, 12.9K PulseMCP/wk.

Product / Business Development

HubSpot MCP

activeOfficial47
HubSpot/mcp-server3 4 evidence

#1 CRM lane for startup/SMB teams. 10.7K npm/wk (ahead of Exa), 12K PulseMCP/wk (6x Salesforce), 335K all-time (#93 global). Open ecosystem approach is closing the CRM gap vs Salesforce's Agentforce gate. Key caveat: currently read-only.

Product / Business DevelopmentMarketing

Zapier MCP

activeOfficial44
23 4 evidence

#1 business automation/connector lane. 35.7K PulseMCP/wk (#46 global), 966K all-time. Unique breadth: 7,000+ apps in a single hosted MCP. No-install surface makes it the only MCP accessible to non-technical business users.

Product / Business DevelopmentAutomation

PostHog MCP

activeOfficial84
PostHog/posthog32K+ 6 evidence

#4 overall, #1 product analytics lane. 5.7M all-time PulseMCP visits (#5 globally), 20.6K/wk. 27 tools across 7 categories. Unique: LLM analytics tracking and error tracking for AI pipelines. Open-source, self-hostable, MIT. Amplitude (#2) and Mixpanel (watch) now catalogued but gap is 100x+ on PulseMCP.

Product / Business Development

Dynamics 365 MCP

activeOfficial40
6 evidence

#12 in product-business-development. The most fully documented enterprise MCP server found — GA on production infrastructure with clear licensing, Entra ID auth, and role-based access. Enterprise-only audience (requires D365 license + Tier 2+ environment). Community traction is thin (25-27 stars). No PulseMCP or npm presence.

Product / Business Development

Amplitude MCP

activeOfficial35
3 evidence

#13 in product-business-development, #2 in product analytics lane. Enterprise positioning with OAuth 2.0 and warehouse-native analytics is a genuine differentiator vs PostHog's API key model. Adoption signals are thin: 34 PyPI downloads/month, fragmented repos (16 + 44 stars), 6.2K PulseMCP/wk.

Product / Business Development

Monday.com MCP

watchOfficial35
1 evidence

Watch — official MCP, free on all plans, 380 stars, but no independent adoption evidence. No PulseMCP listing, no npm package, no independent usage reports for the MCP server specifically. Monday.com has larger market share than Linear for non-dev teams, but the MCP itself is too new to rank confidently.

Product / Business Development

Asana MCP

watchOfficial35
1 evidence

Watch — official with 42 tools and V2 Streamable HTTP migration showing serious investment. Read + write access. However, no public GitHub repo (remote-hosted only), zero independent adoption metrics. V1 deprecation deadline (May 2026) is forcing migration.

Product / Business Development

Gong MCP

watch42
3 evidence

Watch — not catalog-ready. Official MCP announced but Gong Collective page says 'Coming Soon.' Community repos ≤28 stars, stale (last push Dec 2025). No independent users found. Third-party platforms (Zapier, n8n) list Gong via generic API bridges, not native MCP. Revisit Q2 2026 when GA ships.

Product / Business Development

Google Workspace CLI

activeOfficial45
1 evidence

Below cut line. Proves Google is shipping its own surface instead of blessing MCP. CLI approach avoids the tool-count flooding problem. Intentionally non-MCP and narrower scope.

Product / Business Development

Mixpanel MCP

watchOfficial35
1 evidence

Watch — weakest signals in the product analytics lane. Hosted-only model limits community engagement. SSE-only transport is deprecated. US-only expanding. Broadest AI client support (ChatGPT, Gemini, Cursor) is the unique differentiator for non-developer PMs. Not catalog-ready until stronger adoption signals emerge.

Product / Business Development

Emdash

active66
2.9K+ 4 evidence

Best orchestration layer for teams juggling multiple top CLIs. Show HN hit 206 points, YC backing is public, and Ry Walker’s independent comparison put it in Tier 1. Still smaller than Superset by stars, but the feature depth (parallel repos, remote servers, notifications) is unmatched.

Teams of Agents / Multi-Agent Orchestration

Superset

active80
superset-sh/superset7.9K+ 3 evidence

Best pure multiplexer. Highest raw community traction among orchestrators (7.2K stars, 512 PH). Privacy-first (Apache 2.0, zero telemetry, BYOK). Simple and focused.

Teams of Agents / Multi-Agent Orchestration

Grov

active55
TonyStef/Grov177 3 evidence

Best option today for teams who love CLI agents but hate starting every session from a blank slate. Grov keeps a persistent memory DB, injects the right snippets when a new task starts, and intervenes when Claude/Codex/Gemini wander off-plan. Still early (hundreds of stars, limited enterprise proof), but the architecture fills the 'agent memory layer' gap that keeps popping up in rank packets.

Coding CLIs / Code AgentsTeams of Agents / Multi-Agent Orchestration

Factory AI (Droids)

active62
610 5 evidence

Enterprise-only with legitimate backing, but zero grassroots signal and an unverified benchmark claim. Large enterprises with white-glove support needs may benefit; not for individual developers or small teams.

Teams of Agents / Multi-Agent OrchestrationSoftware Factories

GitHub Copilot Coding Agent

activeOfficial42
N/A 6 evidence

The enterprise default for async autonomous coding. Wins on distribution and integration depth (lives inside GitHub where most code already is), not raw capability. SWE-bench 56.0% trails Claude Code (80.8%) by 25 points — pragmatic default, not best tool.

Software Factories

Cursor Automations

active39
N/A 5 evidence

Most innovative architecture in the category — event-driven triggers are genuinely new. But 12 days old with zero independent validation. Needs 30+ days of production evidence before confidence can increase.

Software Factories

Devin (Cognition)

active41
N/A 5 evidence

Highest-funded pure-play, but the gap between self-reported (67% merge rate) and independent results (15% success) is the defining data point. Business metrics are strong; product evidence on complex tasks is weak.

Software Factories

Jules (Google)

active38
N/A 3 evidence

Google-backed with a unique proactive scanning feature, but every independent review says 'not yet a daily driver.' Notable absence of SWE-bench scores. Best for free experimentation.

Software Factories

Augment Code / Intent Agent

active40
N/A 4 evidence

Strongest enterprise contender outside the top 3 by funding and claimed benchmark. The 70.6% SWE-bench score (if verified) would rank it #2. But public trust signals are weak relative to claims — no open-source repo, no public GitHub activity, near-zero community signal. Would move to #2 with a verified, auditable benchmark submission.

Software Factories

Replit Agent 4

active45
N/A 2 evidence

Architecturally impressive and the ChatGPT integration could make it the dominant tool for non-developers building apps. For professional software factories (issue → PR, CI integration, repository-scale refactors), it falls short of the top tier. Production deletion incident (CEO apologized) is a trust signal issue.

Software Factories

oh-my-claudecode

watch77
Yeachan-Heo/oh-my-claudecode11K+ 2 evidence

Watch list. Extraordinary star growth (10K in 10 weeks) but zero independent validation — no HN posts, no Reddit discussion, no reviews. Cannot rank until independent evidence appears. Potential star inflation flag.

Teams of Agents / Multi-Agent Orchestration

Composio Agent Orchestrator

watch80
ComposioHQ/composio28K+ 3 evidence

Provisional #2 in automation — architecturally differentiated as the auth/integration layer FOR AI agents. 27K stars but zero HN traction is an anomalous trust signal. Drops to #4-5 if independent usage evidence doesn't emerge. Use Composio inside n8n or alongside other workflow tools, not as a replacement.

Teams of Agents / Multi-Agent OrchestrationAutomation

Spine Swarm

watch40
N/A 2 evidence

Watch list. Benchmark leader in multi-agent research (GAIA, DeepSearchQA) but not a coding tool. Does not belong in coding-specific rankings unless it ships coding features.

Teams of Agents / Multi-Agent Orchestration

LangGraph

active95
langchain-ai/langgraph27K+ 5 evidence

The production default for Python multi-agent teams. Highest download volume in category (40.2M/month, 7× #2 Python competitor), most independently-verified Fortune 500 deployments, and best-in-class observability via LangSmith. Steeper learning curve than CrewAI — accept the tradeoff consciously.

Teams of Agents / Multi-Agent Orchestration

CrewAI

active93
crewAIInc/crewAI47K+ 5 evidence

Second-largest Python deployment footprint with real Fortune 500 adoption. Consistently rated 'fastest to prototype' — not marketing copy but backed by independent consensus. The right default for role-based business workflow automation teams prioritizing speed. Trade off: observability less mature than LangGraph without AMP Suite.

Teams of Agents / Multi-Agent Orchestration

OpenAI Agents SDK

activeOfficial92
openai/openai-agents-python20K+ 3 evidence

Best for teams that want fast iteration with minimal boilerplate. Minimalist 4-primitive API learnable in an afternoon. Now supports 100+ LLMs — no longer OpenAI-locked. Pre-1.0 API, no state persistence. If your team needs to ship an agent system this week and has never used a framework, start here.

Teams of Agents / Multi-Agent Orchestration

Mastra

active85
mastra-ai/mastra22K+ 4 evidence

For JS/TS teams, this is not a comparison decision — it is the default. No other serious TypeScript-native multi-agent framework at scale. 442-pt HN thread is the strongest community validation signal in the entire frameworks category. Custom license (not MIT/Apache) — review before production use.

Teams of Agents / Multi-Agent Orchestration

Google Agent Development Kit (ADK)

activeOfficial92
google/adk-python19K+ 3 evidence

Strong download velocity for its age — GCP-native deployment and multi-language commitment give it a longer runway than single-language frameworks. Best for teams already on GCP/Vertex AI. No independently-verified production case studies outside Google-controlled publications.

Teams of Agents / Multi-Agent Orchestration

AWS Strands Agents SDK

watchOfficial87
strands-agents/sdk-python5.4K+ 3 evidence

AWS Bedrock teams only. High claimed downloads but anomalous download/star ratio (1,038 vs CrewAI 122) and zero HN organic discussion despite 14M cumulative downloads raises CI/CD pipeline inflation concern. Official AWS tooling is genuine advantage for Bedrock teams; lock-in penalty is high for everyone else.

Teams of Agents / Multi-Agent Orchestration

smolagents (HuggingFace)

watch91
huggingface/smolagents26K+ 2 evidence

Research and experimentation only. LocalPythonExecutor must NOT be used in production under any circumstances. Two independent security firms (JFrog + NCC Group) confirmed this. Docker or E2B sandboxing is an architectural requirement. Best for: evaluating CodeAgent paradigm, HuggingFace model experimentation, academic research.

Teams of Agents / Multi-Agent Orchestration

n8n

active88
n8n-io/n8n181K+ 8 evidence

Clear #1 for workflow automation with AI nodes. If your goal is to wire together SaaS tools with AI orchestration, n8n is the answer. If you're building an agent system from code, use LangGraph or CrewAI instead. Do not mix these use cases.

Teams of Agents / Multi-Agent OrchestrationAutomation

Pydantic AI

active95
pydantic/pydantic-ai16K+ 3 evidence

The type-safe agent logic layer for Python teams. 15.6M downloads/month makes it #3 by volume. Not a competitor to LangGraph — it's a complement. The Pydantic team's reputation is an unmatched trust signal in the Python ecosystem. Best paired with LangGraph for orchestration.

Teams of Agents / Multi-Agent Orchestration

Semantic Kernel (Microsoft)

activeOfficial92
microsoft/semantic-kernel28K+ 3 evidence

The enterprise choice for .NET/C# + Azure shops. Strongest named customer list in the entire category (KPMG, BMW, Fujitsu — independently corroborated). Multi-language (Python, C#, Java) is unique. Python teams should prefer LangGraph or Pydantic AI — Semantic Kernel's Python SDK is secondary.

Teams of Agents / Multi-Agent Orchestration

Agno (formerly Phidata)

watch90
agno-agi/agno39K+ 2 evidence

Include with caveats. High star count but inflated DL/star ratio and self-promotional HN pattern. No independently confirmed enterprise customers. Full-stack offering is ambitious but evidence is thin. Re-evaluate upward if independent enterprise evidence surfaces.

Teams of Agents / Multi-Agent Orchestration

AutoGen (Microsoft)

staleOfficial95
microsoft/autogen56K+ 2 evidence

Sunset in progress. Microsoft confirmed maintenance mode — no new features. Being replaced by Microsoft Agent Framework which combines AutoGen + Semantic Kernel. Do not start new projects on AutoGen. Existing teams should plan migration to Microsoft Agent Framework (GA ~Q2 2026).

Teams of Agents / Multi-Agent Orchestration

ccpm

active71
AThileworker/ccpm7.7K+ 1 evidence

Pragmatic parallel agent IDE for developers who prefer shell-based workflows over GUI tools. 112 HN comments is the deepest community discussion in the segment. Uses GitHub Issues + git worktrees rather than inventing new orchestration — appealingly simple.

Teams of Agents / Multi-Agent Orchestration

Sim Studio

active82
simstudioai/sim27K+ 2 evidence

The only credible open-source alternative to n8n for workflow automation. Apache-2.0 license is a genuine differentiator for teams that need true open-source. Revenue and enterprise evidence is thin — too early to challenge n8n's 3,000+ enterprise customers, but the trajectory is promising.

Teams of Agents / Multi-Agent Orchestration

Jasper AI

active43
6 evidence

Best for brand-governed marketing teams of 3-10 people. Strongest independent review coverage (8/10 avg), verified enterprise adoption (~20% Fortune 500 claimed). Moat is orchestration + brand rails, not generation quality — ChatGPT/Claude match raw copy at lower cost. No open-source, no MCP, no CLI.

MarketingContent & Writing

Copy.ai

active45
3 evidence

Best for marketing workflow automation. Strongest workflow builder in the category — chains research → draft → edit → publish. $29/mo entry is accessible. Named enterprise evidence (Banzai VP reduced campaign creation from 5-6 hours to under an hour). No open-source, no MCP.

MarketingContent & Writing

Writesonic / Chatsonic

active42
4 evidence

Best pure SEO play at budget pricing ($19/mo entry). 4.7/5 Trustpilot from 5K+ reviews is highest verified user satisfaction in category. But GEO feature appears broken per independent practitioner test, and content quality questioned. Mixed evidence overall.

MarketingContent & Writing

Frase

active41
5 evidence

Best value SEO content tool with AI visibility tracking. GEO tracking is a genuine 2026 differentiator — monitors how content appears in AI-generated answers. $38-45/mo is price leader vs Surfer ($89) and Clearscope ($189). Strong independent comparison coverage.

MarketingContent & Writing

Surfer SEO

active40
4 evidence

The benchmark for data-driven on-page SEO. Consistently ranked as SERP analysis leader across independent sources. $89/mo is premium but justified by depth. No AI agent capabilities, no MCP. Consider Frase instead if GEO tracking or budget matters more.

MarketingContent & Writing

AirOps

active42
3 evidence

Best for content operations at scale — but only if your content strategy is already solid. Enterprise client list provides credibility. Steep learning curve confirmed independently. TechCrunch's 'SEO slop' framing is a reputational risk. $99/mo Scale plan.

Marketing

Postiz

active85
gitroomhq/postiz-app28K+ 4 evidence

The SkillBench play for social media. Strongest open-source signal in marketing (27.3K stars), HN traction (42pts Show HN), viable business ($17K MRR, ~472 subscribers, 80% margins). Agent CLI makes it skills-ecosystem-native. Narrow scope (social scheduling only) but dominates that lane.

Marketing

Klue

active42
3 evidence

Enterprise CI leader — different lane from other marketing tools but marketing-adjacent. 4.7/5 G2 (2,257 reviews) is the strongest social proof in the entire marketing category. Forrester Strong Performer. ~$16K/year pricing limits to enterprise. Compete Agent is first-mover in autonomous CI.

Marketing

SE Ranking MCP Server

activeOfficial53
seranking/seo-data-api-mcp-server8 2 evidence

Only MCP-native SEO tool with a working implementation. Compelling time savings (40+ min/project per independent test). But 8 GitHub stars is negligible — most users access via SE Ranking's paid API. The SkillBench play for SEO data if adoption grows.

Marketing

Gumloop

active40
2 evidence

Strongest funding signal in category ($50M Series B from Benchmark, $70M total). Visual pipeline builder appeals to non-technical marketers. But broader than marketing (general automation), credit-based pricing is opaque, and product maturity is unclear despite funding.

MarketingAutomation

Competely

watch35
1 evidence

Lightweight competitor analysis for startups — the affordable alternative to Klue's $16K/year. Thin independent evidence. No HN signal, no major review coverage. Listed in roundups but not deeply reviewed. Too small for strong ranking but fills a gap.

Marketing

Clearscope

active35
1 evidence

Premium content grading for teams writing 15+ articles/month. Declining relevance — Frase and Surfer offer comparable functionality at lower prices ($38-89 vs $189). No AI agent capabilities, no workflow automation. Coasting on 2023-era reputation.

Marketing

Vale

active80
vale-cli/vale5.3K+ 5 evidence

The standard for docs-as-code style enforcement. Adopted by Grafana, Datadog, Meilisearch. Pragmatic Bookshelf published a dedicated book. HN 230pts — strongest writing-tool HN signal. No real competitor for CI/CD prose linting. 5,299 stars.

Content & WritingDocumentation

Harper

active85
Automattic/harper10K+ 4 evidence

Best privacy-first grammar checker for developers. 10,111 stars (2x Vale), HN 645pts (highest in entire content-writing category). Backed by Automattic. English-only, grammar-only — no style guides, no AI rewriting. Different tool than Vale: grammar vs style enforcement.

Content & Writing

LanguageTool

active84
languagetool-org/languagetool14K+ 3 evidence

Best for multilingual teams and self-hosted deployments. 14,184 stars (highest in category), HN 370pts. 30+ languages — unmatched. Self-hostable with SOC 2 and GDPR compliance. The go-to for non-English support or on-premise deployment.

Content & Writing

Sudowrite

active40
4 evidence

Category winner for AI fiction writing — no real competitor. Custom Muse model, Story Bible for lore consistency, fiction-specific UX. HN 51pts. Endorsed by Kindlepreneur and NerdyNav. Users report 4x faster first drafts. $19-59/mo.

Content & Writing

Writer.com

active38
3 evidence

Best for deep enterprise AI governance. Proprietary Palmyra LLMs, Knowledge Graph, AI Guardrails are genuinely differentiated features. VentureBeat coverage. But enterprise pricing, mixed reception for Palmyra quality vs frontier models, thin independent reviews.

Content & Writing

Anyword

active40
2 evidence

Best for performance marketers (narrow subcase). Unique Predictive Performance Score is the only quantified predictive scoring in the category. G2-verified reviews confirm adoption. $49-499/mo limits to agencies/enterprises. Not a general-purpose content tool.

MarketingContent & Writing

Grammarly / Superhuman

active45
3 evidence

Best for broad writing assistance at enterprise scale. Near-universal adoption. Rebrand to Superhuman signals pivot away from pure writing toward AI productivity suite. HN community deeply skeptical (privacy/keylogger concerns, 464pts negative). Lighter governance than Writer.com.

Content & Writing

Acrolinx

active35
1 evidence

Best for regulated industries (narrow subcase). Only contender purpose-built for compliance-heavy content ops. Fortune 2000 focus. But almost entirely self-reported evidence — no HN signal, no GitHub presence, no independent reviews with substance. If you're in a regulated industry, evaluate. Otherwise, Writer.com covers governance better.

Content & Writing

KoalaWriter

active35
2 evidence

Best for high-volume SEO blog drafts (narrow subcase). One-keyword-to-article pipeline is fast for affiliate/niche site operators. $9-49/mo. But output requires heavy editing, rarely ready to publish. A draft factory, not a content tool.

Content & Writing

alex

active30
get-alex/alex5.1K+ 1 evidence

Best for inclusive language enforcement (narrow subcase). Unique niche — no competitor does this. Works well in CI alongside Vale. 5,091 stars. Narrow scope but valuable as a CI/CD complement.

Content & Writing

write-good

active30
btford/write-good5.1K+ 1 evidence

Lightweight naive linter. Good starter tool, but Vale subsumes its functionality via the write-good style package. 5,067 stars. Use Vale instead for serious docs-as-code workflows.

Content & Writing

proselint

stale72
amperser/proselint4.5K+ 1 evidence

Stale — high issue count (247 open), Vale covers its functionality with better maintenance. 4,519 stars. Use Vale with the proselint style package instead.

Content & Writing

Ocoya

watch35
1 evidence

Budget social media AI at $15/mo. 30+ integrations but limited differentiation vs Postiz (open-source, self-hostable, agent CLI) or native platform tools. AI agent capabilities are 'coming soon' — lags the trend. Not strong enough for top ranking.

Marketing

Brave Search API

activeOfficial72
brave/brave-search-mcp-server825 8 evidence

#1 search API for AI agents — fastest latency (669ms), highest independent benchmark score (14.89), SOC 2 Type II attested. Default recommendation for general-purpose agent search.

Search & News

SearXNG

active88
searxng/searxng27K+ 3 evidence

#4 in search-news — the only option where no query ever leaves your infrastructure. 26,644 stars, active development. Not independently benchmarked on quality, but unmatched on privacy and cost.

Search & News

Tavily

activeOfficial40
N/A 8 evidence

#5 in search-news — biggest distribution advantage via LangChain default integration. AIMultiple 13.67 (#5) shows meaningful gap below top-4. /research endpoint GA adds deep research lane. Nebius acquisition ($275-400M) adds strategic uncertainty but strongest financial runway.

Search & News

Jina Reader

staleOfficial60
jina-ai/reader10K+ 4 evidence

#6 in search-news — effectively dead. No commits since May 2025 (10+ months). ReaderLM-v2 is strong for edge/on-device but hosted API only. Firecrawl is a strict superset. Do not recommend for new projects.

Search & News

Crawl4AI

active93
unclecode/crawl4ai63K+ 5 evidence

#7 in search-news — the open-source self-hosted choice. 62K stars, Apache-2.0, actively maintained (v0.8.5 released 2026-03-18). ScrapeOps rates 'best open source' (7/10). Fork count nearly matches Firecrawl (6,353 vs 6,516) showing heavy dev usage. Wins on license, cost, and developer control.

Search & News

Parallel AI Search

activeOfficial40
2 evidence

Below cut line in search-news. Top-4 AIMultiple quality (14.21) but 20x Brave's latency makes it async-only. Extreme pricing ($300-$1,200 CPM) limits to high-value enterprise. Closed source, no MCP server. Best deep research API if latency and cost don't matter.

Search & News

You.com Search API

activeOfficial35
2 evidence

Below cut line in search-news. Research API launched Feb 26, 2026 with strong self-reported benchmark claims (#1 DeepSearchQA). Too new for independent verification — no HN traction, no AIMultiple placement. If claims verified, jumps to ranked list immediately.

Search & News

Linkup

watchOfficial35
3 evidence

Below cut line in search-news. 86K weekly downloads suggest real adoption despite 43 GitHub stars. Strong angel backing (Datadog CEO, Mistral CEO). All benchmark claims self-reported — no independent verification yet.

Search & News

Perplexity Sonar API

activeOfficial40
4 evidence

Below cut line in search-news. Consumer brand doesn't translate to API performance. AIMultiple #7 (12.96), 11K+ ms latency, BrowseComp 8%. HumAI shows highest accuracy (87%) but at cost of speed.

Search & News

Valyu DeepSearch

watchOfficial35
1 evidence

Below cut line in search-news. Unique proprietary data angle (50+ sources: SEC, clinical trials) but almost all evidence is self-reported. Minimal traction. Needs independent verification.

Search & News

Serper

activeOfficial35
1 evidence

Below cut line in search-news. Budget Google SERP wrapper. No semantic understanding, no independent index. Useful only when cost is the primary constraint.

Search & News

Spider Cloud

active58
niconiahi/spider2.3K+ 1 evidence

Below cut line in search-news. Niche high-volume crawling tool. Benchmark claims entirely self-reported. Tiny community (2.3K stars).

Search & News

Google Grounding with Search

activeOfficial35
1 evidence

Below cut line in search-news. Platform lock-in (Gemini API only). Not a standalone search API. If Gemini Deep Research exits preview with MCP support, deep research lane shifts dramatically.

Search & News

Bright Data MCP

activeOfficial65
2.2K+ 3 evidence

Below cut line in search-news. #1 in Browser MCP Benchmark (100% extraction success, 90% automation). 2,214 stars, 60+ MCP tools. Not a search API — web access infrastructure. The only option for scraping sites with aggressive anti-bot defenses.

Search & News

Hyperbrowser MCP

watchOfficial50
hyperbrowserai/mcp752 3 evidence

Below cut line in search-news. Strong 90% browser automation (tied Bright Data #1) but 63% web extraction (worst among ranked). GitHub stale 4+ months (last push Nov 2025). Watchlist — strong HN traction and stealth capabilities, but maintenance risk.

Search & News

ScrapeGraphAI

activeOfficial82
ScrapeGraphAI/Scrapegraph-ai23K+ 6 evidence

Below cut line in search-news. 23K stars, 194 HN pts (strongest HN score in category), unique LLM graph pipeline approach. But only 14.6K weekly PyPI downloads vs Firecrawl's 752K — star count likely inflated by viral novelty. Stars-to-downloads ratio 1,580:1 vs Firecrawl's 81:1.

Search & News

Kimi Code (Moonshot AI)

watchOfficial60
7K+ 1 evidence

Watch — real download signal (124K PyPI/week) and strong model capability (K2.5, HN 388 pts). Limited Western ecosystem integration and no SWE-bench Pro or Terminal-Bench scores. Best for Chinese developer ecosystem or teams using Moonshot AI models.

Coding CLIs / Code Agents

Kilo Code (Kilo-Org)

watch65
17K+ 1 evidence

Watch — real download signal (131K npm/week) and meaningful seed funding ($8M). OpenRouter-native architecture is a genuine differentiator for teams wanting model flexibility without managing API keys. Needs stronger differentiation beyond OpenRouter integration.

Coding CLIs / Code Agents

Cursor (Anysphere)

activeOfficial35
N/A 1 evidence

Reference only in coding-clis — Cursor is primarily an IDE, not a CLI agent. $29.3B valuation and strong adoption are real, but closed-source, paid, and IDE-first puts it outside the terminal-native category. Best for developers who want a polished commercial IDE with integrated AI.

Coding CLIs / Code Agents

Warp (Warp Technologies)

activeOfficial65
26K+ 1 evidence

Reference only in coding-clis — Warp is an AI terminal, not a coding CLI agent. Strong UX and 75.8% SWE-bench Verified are real signals, but 4,350 open issues and closed-source licensing are concerns. Best for developers who want an AI-first terminal experience rather than a code agent.

Coding CLIs / Code Agents

Kiro (AWS)

watchOfficial42
N/A 3 evidence

Watch — spec-driven approach is architecturally sound and GovCloud presence is meaningful for regulated industries. But too early to rank higher without a verified benchmark, user count, or independent case study. Outage controversy unresolved.

Coding CLIs / Code AgentsSoftware Factories

RooCode (RooVeterinaryInc)

active68
RooVeterinaryInc/roocode22K+ 3 evidence

#9 coding CLI — best security posture among VS Code agents (SOC 2 Type 2 compliance). 5.0/5 VS Code rating on 1.37M installs is the strongest quality signal in IDE-embedded segment. Custom Modes (security reviewer, test writer, architect personas) are practical differentiation beyond Cline fork origins. Clean security record.

Coding CLIs / Code Agents

Gamma

active40
4 evidence

Default choice for AI pitch decks. Market leader by every quantitative measure — $102M ARR, 70M users, TechCrunch-verified. Design quality is good but can feel 'generic' at investor scale; Slidebean wins for active fundraising workflows. Lowest-friction entry at $8/mo.

Business

Slidebean

active40
2 evidence

Best for active fundraising. Wins over Gamma when you need investor CRM + financial slides in one tool. 30K+ startups served, 500K YouTube subscribers give social proof. If you just need a great-looking deck fast, Gamma wins.

Business

Plus AI

active35
1 evidence

Best if you live in Google Slides or PowerPoint. 1M+ users, top-rated on Google Workspace Marketplace and Microsoft AppSource. SOC 2 Type II. Wins on zero learning curve but lacks standalone design capabilities of Gamma.

Business

Alai

watch35
1 evidence

Best design quality in AI decks per independent comparison (Reprezent test across 11 tools). Worth considering when investor-facing aesthetics matter more than speed. Lacks Gamma's scale and Slidebean's fundraising workflow.

Business

o11

watch35
1 evidence

Wall Street Prep #1 for AI financial modeling. Creates real IB-grade Excel output (linked formulas, not summaries). But dangerously underfunded ($500K seed, ranks 818th on Tracxn). High survival risk despite strong product signal.

Business

Tracelight

active40
2 evidence

Best Excel-native AI assistant for financial modeling. IB/PE practitioners report 90% time savings. FP&A Guy podcast endorsement adds credibility. $3.6M seed is modest but growing. Wins for Excel power users who need formula generation, not PDF extraction.

Business

Endex

active45
2 evidence

Best-funded AI financial modeling tool ($14M, OpenAI-led). Unique PDF-to-Excel pipeline using vision models. Featured by OpenAI as showcase customer. Wins for 'Research Associate' niche — extracting data from PDFs into structured Excel. Not a full modeling tool like o11 or Tracelight.

Business

Rows

active45
1 evidence

Best AI-native spreadsheet for analysis. 89% first-try accuracy in public benchmark (self-run but reproducible methodology). $8/user/mo. Wins for general analysis; not suited for IB-grade three-statement models. 74% dynamic outputs is a unique capability.

Business

Spellbook

active42
3 evidence

Market leader in AI contract review. $90M+ raised, on track for $100M ARR, 4,000 law firms across 80 countries. Exclusive CBA partner (40K members). Now securing $40M debt for acquisitions — signals market consolidation. Best for law firms needing Word-native drafting assistance.

Business

LinkSquares

active45
1 evidence

Best for enterprise legal operations. $164M raised, $800M valuation, G2 leader 5 consecutive years. Wins for bulk contract analysis, M&A due diligence, and legal team management at scale. Not a drafting co-pilot like Spellbook — it's a CLM platform with AI built in.

Business

LEGALFLY

active45
2 evidence

Fastest-growing legal AI tool (800% YoY ARR growth). Unique privacy differentiator: anonymizes every document before AI processing. Best for regulated industries (healthcare, EU government, financial services). SAP and Lufthansa validate enterprise adoption.

Business

Tability

watch35
2 evidence

Only dedicated AI-first OKR tool with public signal. Wins UI, OKR management, and reporting in independent comparisons. But evidence base is thin — no public funding, no benchmark, limited independent verification. Most teams should use Notion/Linear/Asana instead unless they want a dedicated OKR tool.

Business

Midday

active88
midday-ai/midday14K+ 2 evidence

Best open-source business tool for freelancers and small teams. 14K+ GitHub stars (3.1x nearest OSS competitor). Strong Show HN (119pts). Self-hostable, bank connections, AI-powered receipt matching. Wins for developers/freelancers who want control over their business stack.

Business

Frihet MCP

watch50
Frihet-io/frihet-mcp4 1 evidence

First MCP-native ERP server — directly relevant to Claude-native business workflows. 31 tools for invoicing, expenses, and business ops via natural language. But only 3 GitHub stars — too early to recommend for production. Worth watching for the MCP ecosystem angle.

Business

invoice-mcp

watch33
markslorach/invoice-mcp5 1 evidence

Minimal but functional MCP tool for PDF invoice generation via Claude. Too early to rank (5 stars) but interesting for the MCP ecosystem. Active development is a positive signal.

Business

Proposify

active40
2 evidence

Market leader in proposal management. G2 4.6/5 with ~1,000 reviews, Gartner Peer Insights recognition. Not AI-first but has AI-assisted content generation. Most teams can use Google Docs + Claude instead — Proposify wins only at volume with analytics, e-signatures, and template libraries.

Business

Bin (bi.new)

watch45
1 evidence

Most promising AI-native BI experiment. 59pts Show HN is a solid signal for a new BI tool. But extremely early stage — file upload only, no database connectors. For serious BI use Metabase (OSS), Looker, or Tableau. bi.new is for quick ad-hoc analysis where a full BI stack is overkill.

Business

Perplexity Deep Research

activeOfficial43
5 evidence

#1 in research. Speed king (15-30s vs minutes for competitors), highest citation reliability in independent tests, 93.9% SimpleQA accuracy (Helicone). Reddit academic communities endorse as go-to. Best value at $20/mo.

Research

OpenAI Deep Research

activeOfficial42
4 evidence

#2 in research. Reasoning king — 26.6% HLE and 72.57% GAIA are best reported results from any system. MCP support (Feb 2026) enables enterprise toolchain integration. Slower and pricier than Perplexity but unmatched for PhD-level questions.

Research

Google NotebookLM + Deep Research

activeOfficial42
4 evidence

#3 in research. Best multimodal research tool — video, audio, PDFs, data tables. 100+ sources per query (highest coverage). 907 HN pts is the highest engagement of any tool in this category. More a research workbench than a search tool.

Research

GPT Researcher

active89
assafelovic/gpt-researcher26K+ 4 evidence

#4 in research. Open-source leader with the strongest third-party validation — CMU benchmark winner beating Perplexity and OpenAI on citation and report quality. Real PyPI adoption (15.9K/wk) and MCP integration make it the pick for self-hosted research pipelines.

Research

Tongyi DeepResearch

active82
Alibaba-NLP/DeepResearch19K+ 5 evidence

#5 in research. Open-source disruptor — HLE 32.9 exceeds OpenAI's 26.6 on the same benchmark, runs on consumer hardware via MoE. Called 'the DeepSeek moment for AI agents' by VentureBeat. Ranked below GPT Researcher due to less proven adoption data.

Research

STORM (Stanford)

active66
stanford-oval/storm28K+ 5 evidence

#6 in research. Best for structured knowledge curation — Wikipedia-style article generation with peer-reviewed citation quality (84.8% recall, 85.2% precision). Stanford pedigree, Co-STORM collaborative mode (EMNLP 2024). Less practical for agentic workflows than GPT Researcher.

Research

Claude Research

activeOfficial38
3 evidence

#7 in research. Architecturally distinct multi-agent approach is theoretically superior, but 90.2% improvement claim is self-reported with no independent validation. No dedicated HN thread for Claude research. Promising but needs third-party benchmarks to justify a higher ranking.

Research

Grok DeepSearch / DeeperSearch

activeOfficial35
2 evidence

#8 in research. Uniquely positioned for social media and breaking news research — only tool with native X/Twitter integration. But no hard benchmark numbers from independent sources. Qualitative reviews only. Niche use case limits general ranking.

Research

Elicit

activeOfficial40
4 evidence

#1 in academic/scientific research subcategory. Purpose-built for systematic reviews and meta-analyses — no other tool does structured data extraction across dozens of papers. API launch (Mar 2026) enables programmatic workflows. Reddit r/GradSchool endorsement is strong organic signal.

Research

Consensus

activeOfficial38
3 evidence

#2 in academic/scientific research subcategory. Unique consensus meter for claim-level evidence checking — no competitor has this. 200M+ papers via Semantic Scholar. Deep Search (1,000+ papers/query) launched but low HN signal (11 pts) raises adoption concerns.

Research

Semantic Scholar

activeOfficial35
1 evidence

Infrastructure layer for academic research, not a ranked research agent. 214M papers, citation graphs, free API. Consensus and other tools build on it. Essential for paper discovery and citation mapping but not a standalone research agent.

Research

Kestra

active80
kestra-io/kestra27K+ 3 evidence

Best pick for data engineering and DevOps teams migrating from Airflow, or anyone who wants declarative (YAML) workflow definitions with Git version control. If your automation is data pipelines or scheduled ETL, Kestra beats n8n. No MCP bridge limits relevance for Claude Code users.

Automation

Activepieces

active80
activepieces/activepieces21K+ 3 evidence

Best for startups and non-developer teams who want open-source + simplicity. If your team isn't technical enough for n8n but wants open-source, choose Activepieces. Fewer integrations (200+ vs n8n's 1,000+) and no MCP bridge limit power-user appeal.

Automation

Trigger.dev

active92
triggerdotdev/trigger.dev14K+ 3 evidence

Best for TypeScript developers who need background jobs with long-running compute or AI agent workflows. For heavy compute tasks that exceed serverless limits, Trigger.dev beats Inngest. Transparent security incident post-mortem (262 HN pts) builds trust.

Automation

Inngest

active90
inngest/inngest5.1K+ 3 evidence

Default choice for Next.js/Vercel teams needing serverless workflow orchestration with AI inference built in. 499K npm downloads/wk (highest in automation by 2x) proves deep practical adoption despite modest star count. Cannot self-host (ELv2 license) — limits appeal for on-prem teams.

Automation

Windmill

active92
windmill-labs/windmill16K+ 3 evidence

Best for developers who want workflows-as-code rather than visual canvases. Purpose-built for internal tooling, DevOps automation, and script orchestration. If your automation is 'run these scripts in sequence with a UI,' Windmill is the answer. No MCP bridge limits Claude Code integration.

Automation

Pipedream

active90
PipedreamHQ/pipedream11K+ 2 evidence

Best for developers who need quick MCP access to a massive API surface without building their own auth. For one-off API integrations via MCP (no workflow needed), Pipedream's free tier is the fastest path. 4,362 open issues raise maintenance capacity concerns.

Automation

Marimo

active93
marimo-team/marimo20K+ 5 evidence

Best-in-class reactive notebook — the Jupyter replacement that actually works. Reactive execution guarantees reproducibility. Pure .py format is natively agent-readable. The natural choice for agent-assisted data workflows.

Data & Analytics

Data Formulator

active82
microsoft/data-formulator15K+ 2 evidence

Best AI-powered data visualization tool — Microsoft Research quality, fully open source. Fills a niche no other tool covers: conversational, iterative chart building from raw data.

Data & Analytics

Evidence

active78
evidence-dev/evidence6.1K+ 2 evidence

Best BI-as-code platform for SQL-first analysts. If your team thinks in SQL and wants version-controlled, code-authored reports, Evidence beats every other tool. The only tool in the category that doesn't require a general-purpose programming language.

Data & Analytics

Observable Framework

active84
observablehq/framework3.4K+ 2 evidence

Best for developer-built data dashboards with D3-quality visualization. If you need custom, interactive, D3-quality visualizations in a static site, Observable Framework is unmatched.

Data & Analytics

Streamlit

active90
streamlit/streamlit44K+ 2 evidence

Ecosystem giant — essential infrastructure for deploying data apps, but not an analysis engine. Reruns entire script on every interaction. Include as foundational infrastructure, not a direct AI analysis tool.

Data & Analytics

Plexe

active75
plexe-ai/plexe2.6K+ 1 evidence

Most promising prompt-to-ML-model tool — early but uniquely differentiated. Natural language → trained ML model. Needs more real-world validation before it can rank higher.

Data & Analytics

PandasAI

stale72
sinaptik-ai/pandas-ai23K+ 2 evidence

High name recognition but stale and problematic. 5 months without commits, documented hallucination risk, custom license. Good for quick throwaway exploration only — not suitable for anything where correctness matters.

Data & Analytics

Fumadocs

active90
fuma-nama/fumadocs11K+ 4 evidence

The breakout docs framework of 2026. 11.2K stars with only 5 open issues — exceptional maintenance. 309K npm/week surpasses Starlight. 3x YoY growth is the fastest in the category. For Next.js teams, Fumadocs is the clear pick over Nextra.

Documentation

Astro Starlight

active91
withastro/starlight8.2K+ 3 evidence

The developer experience leader. Zero client-side JS by default, Go-compiled dev server starts in half the time of Docusaurus. 8.1K stars, 200K npm/week (doubled YoY), 283 contributors. Real migration trend from Docusaurus (Distr, W3C). Critical gap: no built-in versioned docs.

Documentation

Docusaurus

active92
facebook/docusaurus64K+ 3 evidence

The incumbent with the biggest numbers but coasting. 64.2K stars, 765K npm/week, 457 contributors. v4.0 roadmap at 20% with no target date. Developers actively migrating to Starlight/Fumadocs for DX. Still the right choice when versioned docs and plugin ecosystem are requirements.

Documentation

Fern

active89
fern-api/fern3.6K+ 3 evidence

The most strategically significant move in API docs in 2026. Only tool generating both docs AND production SDKs from a single spec. Postman acquisition (Jan 2026) gives access to 500K+ companies. 3.6K stars, 89K npm/week. Daily pushes post-acquisition confirm continued development.

Documentation

Mintlify

active64
1.7K+ 3 evidence

Market leader for developer-facing documentation SaaS. Helicone acquisition (Mar 2026) signals pivot to AI-agent knowledge infrastructure. 40%+ of doc traffic from AI agents (CEO claim, directionally confirmed by Context7 data). Security incident (GitHub token leak, Dec 2025) is a real trust concern. Premium pricing ($300/mo Pro).

Documentation

Redocly / Redoc

active78
Redocly/redoc26K+ 3 evidence

The enterprise governance pick for API documentation. 25.6K stars, 1.15M npm/week — most downloaded API docs renderer. Only platform supporting SOAP. Strongest OpenAPI linting. Development velocity has slowed (last push 6 weeks ago, last release Sep 2025). Fern+Postman is a growing competitive threat.

Documentation

MkDocs Material

watch92
squidfunk/mkdocs-material26K+ 3 evidence

Most-downloaded docs framework globally by raw numbers (4.2M PyPI/week). But now in maintenance mode. MkDocs upstream is unmaintained since Aug 2024 — supply chain risk. Zensical successor promises 4-5x faster builds but isn't feature-complete. New projects should choose Starlight or Fumadocs instead.

Documentation

Swagger UI

active95
swagger-api/swagger-ui29K+ 1 evidence

The universal baseline for API documentation. 28.7K stars, 603K npm/week combined. Actively maintained with weekly releases. Every API docs tool benchmarks against Swagger UI. Reliable and free but hasn't innovated in years. If you need more than basic OpenAPI rendering, look at Fern or Redocly.

Documentation

Promptless

active38
3 evidence

The only tool focused on keeping docs in sync with code changes automatically. YC W25 backed. 107pts Launch HN. Named customers (Vellum, Vitess, Amplitude) but all traction is self-reported. Premium pricing ($500-1,000/mo). Team of 5. Early-stage risk is high but positioning is unique.

Documentation

GitBook

active35
2 evidence

The 'easy button' for teams with non-developer contributors. WYSIWYG editor lowers the barrier. AI Agent connects docs to support channels. But closed-source SaaS with expensive pricing ($65-249/mo + $12/user/mo) competing against free OSS tools with better developer experience.

Documentation

DocsGPT

active82
arc53/DocsGPT18K+ 2 evidence

Best open-source 'chat with your docs' solution. 17.8K stars, 256pts Show HN (strongest HN reception in documentation category). Self-hostable, multi-model. Complements docs generators (Docusaurus, Mintlify) rather than competing. Release cadence slowing (last release Dec 2025).

Documentation

Nextra

active81
shuding/nextra14K+ 1 evidence

Still works, still maintained, but Fumadocs has overtaken it on every momentum metric (309K vs 113K npm/week, 3x YoY growth vs modest). No reason to choose Nextra over Fumadocs for a new Next.js project. Existing Nextra users don't need to rush to migrate, but the writing is on the wall.

Documentation

Semgrep MCP

activeOfficial10
semgrep/mcp642 3 evidence

#1 SAST skill. Best-in-class OSS SAST with official MCP server. 46% vuln detection in DryRun benchmark (vs SonarQube 19%). AST-based rules are transparent and auditable. Rising mindshare (1.6% → 2.6%). LinkedIn rebuilt SAST pipeline around it. The default recommendation for code scanning via AI agents.

Security

DryRun Security (Code Insights MCP)

activeOfficial35
2 evidence

Highest reported SAST detection rate (88%) but self-reported benchmark. AI-native with natural language code policies. Official MCP server. $8.7M raised. The dark horse — if an independent third party confirms the 88% detection rate, moves to #1 above Semgrep.

Security

CodeQL (via GitHub MCP Server)

activeOfficial88
github/github-mcp-server28K+ 2 evidence

#3 SAST. Best for GitHub-native shops — zero extra setup via the official GitHub MCP Server. Copilot Autofix auto-generates fixes from CodeQL alerts. GitHub Security Lab Taskflow Agent found ~30 real CVEs. If you're all-in on GitHub, this jumps to #1.

Coding CLIs / Code AgentsSecurity

Snyk Code (via Snyk MCP Server)

activeOfficial35
1 evidence

#4 SAST. Best commercial all-in-one security platform (SAST + SCA + IaC + containers). DeepCode AI engine with Agent Fix auto-remediation. Strongest for teams already on Snyk — adding Agent Scan is the only incremental tool needed.

Security

Datadog Code Security MCP

activeOfficial35
1 evidence

Best for Datadog-native shops. SAST + secrets + SCA + IaC within your existing observability stack. Don't add 4 separate tools if you already have Datadog. GA March 2026.

Security

GitGuardian MCP (ggmcp)

activeOfficial60
GitGuardian/ggmcp34 2 evidence

#1 secret detection. Purpose-built secret scanning MCP with 500+ detectors and hard merge gates for AI-generated code. State of Secrets Sprawl 2026 report (81% surge in AI-service key leaks, 24,008 secrets in MCP configs) is the definitive source on the problem. The default recommendation for secret scanning in agent workflows.

Security

TruffleHog

active88
trufflesecurity/trufflehog25K+ 1 evidence

Best for credential verification in CI/CD pipelines. 18K+ stars, 800+ secret types, and unique active credential verification (tells you which leaks are still dangerous). Scans beyond git (S3, Docker, Slack). No official MCP server is the gap — use in CI/CD rather than agent workflows.

Security

Gitleaks

active88
gitleaks/gitleaks26K+ 1 evidence

Best pre-commit secret scanner. 24.4K stars — most-starred in the category. 150+ patterns, fastest scanner. The community default. No official MCP server — use as pre-commit hook, not agent integration.

Security

Snyk Agent Scan

activeOfficial40
2 evidence

#1 agent/MCP security scanner. Scans AI agents, MCP servers, and skills for prompt injection, tool poisoning, and toxic flows. Auto-discovers Claude, Cursor, Gemini CLI, Windsurf configs. Skill Inspector (Feb 2026) + Vercel supply chain partnership. Enterprise trust.

Security

Tencent AI-Infra-Guard

activeOfficial81
Tencent/AI-Infra-Guard3.3K+ 1 evidence

#2 agent/MCP security scanner. Most comprehensive OSS red teaming tool — ClawScan, Agent Scan, Skills Scan, MCP scan, jailbreak eval. 3,264 stars (highest in agent security). 43 AI framework components, 589 CVEs cataloged. Best for OSS-first teams wanting breadth without commercial dependencies.

Security

Cisco MCP Scanner

activeOfficial64
cisco-ai-defense/mcp-scanner859 1 evidence

#3 agent/MCP security scanner. Best behavioral analysis — 3 scanning engines (Yara, LLM-as-judge, Cisco AI Defense) detect semantic threats that pattern matching misses. 852 stars. Enterprise-backed (Cisco) and open source.

Security

MCP-Shield

active38
riseandignite/mcp-shield550 1 evidence

First-mover MCP security scanner. 548 stars, 134 pts HN. Simpler than Snyk Agent Scan or Tencent AI-Infra-Guard but battle-tested. Good for teams wanting a lightweight, proven scanner without enterprise overhead.

Security

HexStrike AI

active80
0x4m4/hexstrike-ai7.7K+ 1 evidence

#1 offensive security skill. 7,561 stars — largest security MCP repo. 150+ cybersecurity tools. For authorized pentesting, CTF, and bug bounty only. The clear leader in agent-assisted offensive security.

Security

MCP for Security (cyproxio)

active53
cyproxio/mcp-for-security583 1 evidence

#2 offensive security. Curated collection of pentesting MCP servers (SQLMap, FFUF, NMAP, Masscan). 569 stars. Better organized than HexStrike but narrower scope. For authorized pentesting only.

Security

FuzzingLabs Security Hub

active59
FuzzingLabs/mcp-security-hub490 1 evidence

#3 offensive security. MCP servers for Nmap, Ghidra, Nuclei, SQLMap, Hashcat. 481 stars. Differentiates on reverse engineering / binary analysis. Best for security researchers working with binaries and protocols.

Security