skillpack.co

All Skills

91 skills tracked across 15 categories. Each with an editorial verdict, source links, and evidence.

91

Skills

15

Categories

91

Figma MCP Server Guide

activeOfficial62
figma/mcp-server-guide431 5 evidence

The trust leader with unmatched long-term durability — triple partnership with OpenAI Codex, GitHub Copilot, and Claude Code. Bidirectional since Mar 6 2026. Wins specifically when your team has Code Connect configured. Without it, Framelink produces cleaner context. Pricing is the primary barrier: free tier (6 calls/month) is non-functional.

UX / UI

Framelink / Figma-Context-MCP

active70
GLips/Figma-Context-MCP14K+ 5 evidence

Best read-only design-to-code MCP. Use this unless you have Code Connect configured — 25% smaller output than Official, avoids prescriptive code that poisons LLM context. Works on free Figma plans with broadest editor support.

UX / UI

Figma-use

active48
dannote/figma-use503 3 evidence

Conditional pick. Strongest HN validation in the write-access lane, but solo maintainer and potential OpenPencil pivot raise durability concerns. Use if you want CLI-first workflow without a Figma plugin.

UX / UI

Vibma

watch48
ufira-ai/Vibma426 2 evidence

Emerging. Only tool publishing model-specific design quality benchmarks. Harness engineering angle is genuinely differentiated, but HN Show got only 2 points and no independent reviews found.

UX / UI

Cursor Talk to Figma

active57
grab/cursor-talk-to-figma-mcp6.5K+ 4 evidence

Best general-purpose write-access Figma MCP for individual developers and small teams. More accessible than Console MCP, works on free Figma plans, lower setup complexity. Dropped to #4 as Console MCP's Uber validation is stronger enterprise evidence.

UX / UI

Figma Console MCP

active59
southleft/figma-console-mcp1.1K+ 3 evidence

The enterprise write-access leader. Uber's production uSpec validates this at scale — automated component specs across 7 stacks, accessibility in under 2 minutes. Hockey-stick npm trajectory is fastest-growing in the category. Local WebSocket means no rate limits and no data leaving your network.

UX / UI

Penpot MCP

watchOfficial43
penpot/penpot-mcp-server225 2 evidence

The open-source design MCP lane exists but isn't production-ready yet. Watch item — if development resumes and beta launches, it moves up.

UX / UI

Claude Talk to Figma

active42
arinspunk/claude-talk-to-figma516 1 evidence

Best Claude-specific write-access option. DXT installer reduces setup to one click for Claude Desktop. If you're all-in on Claude Code or Claude Desktop and need write-access to Figma without paying for Dev Mode, this is the natural pick.

UX / UI

Onlook

active72
onlook-dev/onlook25K+ 2 evidence

Different lane from Figma MCP tools — not a Figma bridge, but a direct design-in-code editor. 24,918 stars and two HN hits above 200 points make this the most-starred tool in the UX/UI category. Best for frontend teams where designers work directly in the Next.js + Tailwind codebase and want to eliminate the Figma-to-code translation step.

UX / UI

Excalidraw MCP

activeOfficial47
excalidraw/excalidraw-mcp3.4K+ 1 evidence

Clear diagramming lane leader. 3,371 stars and official Excalidraw org backing make it the default for AI-assisted diagramming. Different workflow from design-to-code — listed in UX/UI for visibility but doesn't compete with Figma MCPs.

UX / UI

Firecrawl MCP Server

activeOfficial70
firecrawl/firecrawl-mcp-server5.8K+ 91%8 evidence

#1 in product-business-development by volume and evidence. 50.6K npm/wk (highest in category), 83% benchmark accuracy (AIMultiple), F1 0.638 (SearchMCP). Three independent sources confirm extraction superiority.

Product / Business DevelopmentSearch & News

Exa MCP Server

activeOfficial62
exa-labs/exa-mcp-server4.0K+ 11 evidence

#6 in product-business-development, #2 research tool. Best for semantic discovery ('find companies like X'). npm declining (9.3K/wk, down from 11.2K). AIMultiple benchmark tested extraction, not discovery — 23% undersells Exa's actual strength.

Product / Business DevelopmentSearch & News

MCP Atlassian

active61
sooperset/mcp-atlassian4.7K+ 7 evidence

#2 overall, #1 enterprise operating surface for teams on the Atlassian stack. Highest stars (4.6K), deepest tool set (72 tools), 139K PyPI weekly downloads (primary channel). On-prem/Data Center support is the key moat vs official Rovo MCP.

Product / Business Development

Notion MCP Server

activeOfficial70
makenotion/notion-mcp-server4.1K+ 7 evidence

#1 operating surface for product teams — 47K+ npm weekly downloads (17x Atlassian) is the strongest objective adoption signal. Notion 3.3 Custom Agents ecosystem positions MCP as multi-tool hub.

Product / Business Development

Slack MCP Server

active66
korotovsky/slack-mcp-server1.5K+ 3 evidence

Communication lane, watch — metrics unverified. PulseMCP shows 1.6K/wk and #859 globally (2026-03-19), far below earlier catalog claims of 35.9K npm/week. Official Slack MCP GA and 50+ launch partners are confirmed, but current traction signals are weak. Platform relationship is real; MCP adoption is not yet the mechanism the market is using.

Product / Business Development

Google Workspace MCP

active58
taylorwilsdon/google_workspace_mcp1.9K+ 4 evidence

#6 overall. Best broad operating skill for teams living in Google Workspace. Breadth is real (100+ tools, 12+ services), but low npm adoption (604/week), community-only status, and existential threat from Google's own CLI weigh against it.

Product / Business Development

OpenHands

active88
OpenHands/OpenHands69K+ 12%11 evidence

Category leader. No other tool has the combination of open-source community (69K stars, 455 contributors), real download volume (1M/month), venture backing ($18.8M), hardware partnerships (AMD), and benchmark leadership (SWE-bench Verified 72%). Gap to #2 is enormous.

Teams of Agents / Multi-Agent OrchestrationSoftware Factories

Ralph Loop Agent

activeOfficial44
vercel-labs/ralph-loop-agent720 7 evidence

Best reference when the team wants a crisp loop pattern instead of a huge agent platform. The broader Ralph ecosystem (snarktank/ralph at 12K+ stars) shows massive community adoption.

Teams of Agents / Multi-Agent Orchestration

SWE-agent

stale68
SWE-agent/SWE-agent19K+ 2 evidence

Research/academic reference only. Princeton pedigree and 79.2% SWE-bench Verified on Opus 4.5 scaffold give it strong benchmark credibility. But no release in 10 months (last: v1.1.0, 2025-05-22) puts it outside the production cadence of all active tools. Use as a benchmark scaffold reference, not as a production coding CLI.

Coding CLIs / Code AgentsTeams of Agents / Multi-Agent Orchestration

Claude Code

activeOfficial90
anthropics/claude-code80K+ 121%17 evidence

The #1 coding CLI agent. Leads SWE-bench Pro standardized (45.89%), wins independent head-to-heads on reasoning depth, ~4% of GitHub commits. Rate limits are the #1 complaint. Costs 2-3x more per task than Codex CLI due to higher token consumption.

Coding CLIs / Code AgentsTeams of Agents / Multi-Agent Orchestration

Aider

active91
Aider-AI/aider42K+ 11%4 evidence

#7 coding CLI — strong verification story (191K PyPI/week, 5.7M lifetime), but Codex CLI at 2.49M npm/week and Gemini CLI at 678K/week have overtaken Aider's download rank. Category pressure is real: HN thread 'Claude Code with Sonnet 4 is so good I have stopped using Aider' (#44154020). Best for Python devs who want fine-grained model control and git-native workflow. v0.86.2 (2026-02-12) is 5 weeks behind competitors shipping daily.

Coding CLIs / Code AgentsSoftware Factories

Continue (Continuous AI)

active68
continuedev/continue32K+ 2 evidence

Best pick for teams that want background AI agents enforcing code quality on PRs, not real-time autocomplete. The pivot repositioned it away from individual devs toward team CI workflows.

Coding CLIs / Code AgentsTeams of Agents / Multi-Agent Orchestration

OpenCode

active83
sst/opencode125K+ 6 evidence

#5 coding CLI — active again after a development gap. v1.2.27 released 2026-03-16 with OpenAI official partnership (post-Anthropic block). 124,766 stars make this the largest AI coding repo by raw count. The unauthenticated RCE was fixed in v1.1.10+ but the corporate blocking incident and security history require explicit disclosure. Best for developers who want maximum model flexibility with a proven star-signal community.

Coding CLIs / Code Agents

Codex CLI

activeOfficial85
openai/codex66K+ 8%7 evidence

#2 coding CLI. Terminal-Bench 77.3% (GPT-5.3-Codex) and 3-4x more token-efficient. Sandbox-first safety is a genuine differentiator. Free with ChatGPT removes the cost barrier. Trails Claude Code by ~5pp on SWE-bench Pro standardized (41.04% vs 45.89%).

Coding CLIs / Code Agents

Gemini CLI

activeOfficial86
google-gemini/gemini-cli98K+ 15%8 evidence

#3 coding CLI — best free entry point. 1K req/day free tier is unmatched, 1M context is the largest. SWE-bench Pro standardized 43.30% (#2, only 2.59pp behind Claude Code). File deletion incident (AI Incident Database #1178) and tool-calling weaknesses are real concerns.

Coding CLIs / Code Agents

Cline (cline.bot)

watch64
59K+ 5 evidence

#4 coding CLI — strongest in the IDE-embedded-agent segment with 3.35M VS Code installs, 5M total across editors, $32M funding (Emergence Capital), and named enterprise customers (Salesforce, Samsung, SAP). The supply chain incident (v2.3.0 'OpenClaw') is a documented trust flag that still applies. A credible third-party security audit would remove the primary concern. Best for developers who live in VS Code and want agentic assistance without leaving the editor.

Coding CLIs / Code AgentsSoftware Factories

GitHub Copilot CLI

activeOfficial10
github/gh-copilot1.1K+ 3 evidence

#5 coding CLI — distribution is the killer advantage (15M paid Copilot subscribers). Multi-model approach and Enterprise Agent Control Plane are unique. But CVE-2026-29783 (arbitrary code execution 2 days after GA) and no published benchmark scores are serious concerns.

Coding CLIs / Code Agents

Amp (Amp Inc.)

active28
N/A 2 evidence

Watch — corporate spin-out from Sourcegraph to independent Amp Inc. (March 2026) is a material change. The original sourcegraph/amp GitHub link returns 404. Tool still ships (36K npm downloads/week) under ampcode.com. Sub-agent architecture (Oracle, Librarian, Painter) remains the most sophisticated in the category. Update or verify all links before recommending.

Coding CLIs / Code AgentsSoftware Factories

Goose (Block)

active68
block/goose33K+ 2 evidence

Best choice for teams who want a fully open, auditable, provider-agnostic agentic CLI that is not controlled by a model vendor. Linux Foundation governance adds institutional credibility. No published benchmarks and limited verified install data keep it in Tier 2.

Coding CLIs / Code Agents

Crush (Charmbracelet)

active68
charmbracelet/crush22K+ 2 evidence

Best choice for developers who live in the terminal and want a polished, non-VS-Code-dependent experience. Charmbracelet's proven track record (Bubble Tea, 25K+ apps) is a strong prior. No published benchmark scores — community quality signal is the main trust anchor.

Coding CLIs / Code Agents

Qwen Code (Alibaba / QwenLM)

activeOfficial68
QwenLM/qwen-code21K+ 2 evidence

Best cost=zero option for developers who want a serious open-weight model without paying per token. The 70.6% SWE-bench Verified figure (Qwen3-Coder-Next) is the strongest open-weight score in the category and anchors the strongest local/on-prem story. Alibaba/Chinese cloud provenance is a consideration for enterprise and GovCloud use cases.

Coding CLIs / Code Agents

Junie CLI (JetBrains)

watchOfficial28
N/A 1 evidence

Too early to rank — no public artifact (no repo, no benchmark, no independent review). JetBrains has 11M+ paid IDE seats; if the JetBrains installed base converts, this becomes a serious Tier 2 contender within 60 days. Check back after first independent comparisons.

Coding CLIs / Code Agents

Auggie CLI (Augment Code)

watchOfficial43
153 2 evidence

Highest SWE-bench Pro number in the category (51.80% on Augment scaffold), but the scaffold is not standardized — they used the same Opus 4.5 model that scores 45.89% on SEAL's standardized setup. The architecture/scaffolding advantage is credible and meaningful. Cannot rank above tools with millions of verified installs on a single blog-post benchmark. Watch for: public GA release and independent SWE-bench reproduction.

Coding CLIs / Code Agents

Browser Use

active76
browser-use/browser-use81K+ 10 evidence

Best default for agents that need to see and interact with real web pages end-to-end.

Web Browsing / Browser Automation

Playwright MCP

activeOfficial84
microsoft/playwright-mcp29K+ 7 evidence

Best MCP-native browser option for teams in Microsoft's ecosystem. The accessibility-snapshot approach is more reliable than vision-based alternatives for structured data extraction.

Web Browsing / Browser Automation

Stagehand

activeOfficial76
browserbase/stagehand22K+ 7 evidence

Best pick when the team wants TypeScript-native browser automation with the simplest possible API surface.

Web Browsing / Browser Automation

Chrome DevTools MCP

activeOfficial86
ChromeDevTools/chrome-devtools-mcp30K+ 5 evidence

Best MCP-native option for debugging, performance analysis, and Chrome-specific development workflows. Complements Playwright MCP rather than replacing it.

Web Browsing / Browser Automation

Vercel Agent Browser

active80
vercel-labs/agent-browser24K+ 4 evidence

Best pick when token efficiency is the primary constraint — long autonomous sessions, cost-sensitive deployments, or agents hitting context limits.

Web Browsing / Browser Automation

Skyvern

active80
Skyvern-AI/skyvern21K+ 6 evidence

Best pick for enterprise workflow automation on websites without APIs — form filling, data entry, procurement. Overkill for developer/coding agent browser tasks.

Web Browsing / Browser Automation

Lightpanda

active71
lightpanda-io/browser22K+ 3 evidence

Best infrastructure pick for teams running browser automation at scale who need to reduce costs. Drop-in CDP replacement for Chrome headless in pipelines. Still beta — not a user-facing agent tool.

Web Browsing / Browser Automation

Vibium

watch59
VibiumDev/vibium2.7K+ 3 evidence

Below cut line — only 2.7K stars, not production-ready by creator's own characterization. But founder pedigree (Selenium, Appium creator), 443-point HN thread, near-daily releases, and W3C standards-first architecture demand close tracking. The only Lane 2 tool with credible cross-browser (Firefox + Safari) story.

Web Browsing / Browser Automation

BrowserOS / Nxtscape

watch58
nxtscape/nxtscape10.0K+ 3 evidence

Lane 4 anchor (consumer agentic browsers) — 9,982 stars, YC-backed, 47 releases in ~10 months (v0.43.0 on 2026-03-12), AGPL-3.0. MCP server integration (31 tools) in v0.42.0, Skills/Memory/SOUL.md in v0.43.0. Demonstrably not vaporware. Thin contributor count (~10) and AGPL license limit enterprise adoption.

Web Browsing / Browser Automation

Salesforce MCP

activeOfficial54
salesforcecli/mcp319 3 evidence

#2 CRM lane, behind HubSpot. 26.6K npm/week is primarily Salesforce DX developer tooling (SFDX developers), not CRM product teams. Agentforce-gated open CRM access keeps PulseMCP low (~2K/wk). Right choice only for existing Salesforce Enterprise customers already committed to Agentforce.

Product / Business Development

Atlassian Rovo MCP

activeOfficial43
444 2 evidence

#7 in product-business-development. Official backing and 12+ AI client partnerships are strong trust signals. Cloud-only limitation and 10x star gap vs sooperset community server keep it behind.

Product / Business Development

Airtable MCP Server

active42
domdomegg/airtable-mcp-server428 1 evidence

#8 in product-business-development. Niche but real — 2.5K npm/week shows genuine adoption. Fills the gap between spreadsheets and databases.

Product / Business Development

Linear MCP Server

active33
jerhadf/linear-mcp-server347 4 evidence

#1 in project/product management lane. Feb 2026 PM upgrade (triage, backlog prioritization, initiative creation, milestone management) elevated Linear from dev-only to full PM surface. ~2,743 combined npm/wk (mcp-linear 2,074 + linear-mcp 669), 12.9K PulseMCP/wk, #123 global.

Product / Business Development

HubSpot MCP

activeOfficial23
HubSpot/mcp-server3 4 evidence

#1 CRM lane for startup/SMB teams. 10.7K npm/wk (ahead of Exa), 12K PulseMCP/wk (6x Salesforce), 335K all-time (#93 global). Open ecosystem approach is closing the CRM gap vs Salesforce's Agentforce gate. Key caveat: currently read-only.

Product / Business Development

Zapier MCP

activeOfficial34
23 3 evidence

#1 business automation/connector lane. 35.7K PulseMCP/wk (#46 global), 966K all-time. Unique breadth: 7,000+ apps in a single hosted MCP. No-install surface makes it the only MCP accessible to non-technical business users.

Product / Business Development

PostHog MCP

activeOfficial73
PostHog/posthog32K+ 4 evidence

#1 product analytics lane (no competition). 5.7M all-time PulseMCP visits (#5 globally across all MCP servers) with 20.6K/wk current. Zero catalog competition. Monorepo integration signals long-term product investment, not a side project.

Product / Business Development

Emdash

active51
2,741 3 evidence

Best orchestration layer. Only tool offering Best-of-N agent comparison and triple issue-tracker integration. Ry Walker Tier 1 independent validation outweighs modest star count.

Teams of Agents / Multi-Agent Orchestration

Superset

active66
superset-sh/superset7.4K+ 3 evidence

Best pure multiplexer. Highest raw community traction among orchestrators (7.2K stars, 512 PH). Privacy-first (Apache 2.0, zero telemetry, BYOK). Simple and focused.

Teams of Agents / Multi-Agent Orchestration

Factory AI (Droids)

active53
610 5 evidence

Enterprise-only with legitimate backing, but zero grassroots signal and an unverified benchmark claim. Large enterprises with white-glove support needs may benefit; not for individual developers or small teams.

Teams of Agents / Multi-Agent OrchestrationSoftware Factories

GitHub Copilot Coding Agent

activeOfficial43
N/A 4 evidence

The enterprise default for async autonomous coding. Wins on distribution and integration depth (lives inside GitHub where most code already is), not raw capability. SWE-bench 56.0% trails Claude Code (80.8%) by 25 points — pragmatic default, not best tool.

Software Factories

Cursor Automations

active41
N/A 3 evidence

Most innovative architecture in the category — event-driven triggers are genuinely new. But 12 days old with zero independent validation. Needs 30+ days of production evidence before confidence can increase.

Software Factories

Devin (Cognition)

active38
N/A 4 evidence

Highest-funded pure-play, but the gap between self-reported (67% merge rate) and independent results (15% success) is the defining data point. Business metrics are strong; product evidence on complex tasks is weak.

Software Factories

Jules (Google)

active34
N/A 3 evidence

Google-backed with a unique proactive scanning feature, but every independent review says 'not yet a daily driver.' Notable absence of SWE-bench scores. Best for free experimentation.

Software Factories

Augment Code / Intent Agent

active38
N/A 2 evidence

Strongest enterprise contender outside the top 3 by funding and claimed benchmark. The 70.6% SWE-bench score (if verified) would rank it #2. But public trust signals are weak relative to claims — no open-source repo, no public GitHub activity, near-zero community signal. Would move to #2 with a verified, auditable benchmark submission.

Software Factories

Replit Agent 4

active48
N/A 2 evidence

Architecturally impressive and the ChatGPT integration could make it the dominant tool for non-developers building apps. For professional software factories (issue → PR, CI integration, repository-scale refactors), it falls short of the top tier. Production deletion incident (CEO apologized) is a trust signal issue.

Software Factories

oh-my-claudecode

watch58
Yeachan-Heo/oh-my-claudecode10K+ 2 evidence

Watch list. Extraordinary star growth (10K in 10 weeks) but zero independent validation — no HN posts, no Reddit discussion, no reviews. Cannot rank until independent evidence appears. Potential star inflation flag.

Teams of Agents / Multi-Agent Orchestration

Composio Agent Orchestrator

watch58
ComposioHQ/composio27K+ 1 evidence

Watch list. Autonomous CI fix feature is genuinely novel and unique in category. Too early to rank at 1 month old with no HN engagement. Revisit in 60 days.

Teams of Agents / Multi-Agent Orchestration

Spine Swarm

watch38
N/A 2 evidence

Watch list. Benchmark leader in multi-agent research (GAIA, DeepSearchQA) but not a coding tool. Does not belong in coding-specific rankings unless it ships coding features.

Teams of Agents / Multi-Agent Orchestration

LangGraph

active78
langchain-ai/langgraph27K+ 3 evidence

The production default for Python multi-agent teams. Highest download volume in category (40.2M/month, 7× #2 Python competitor), most independently-verified Fortune 500 deployments, and best-in-class observability via LangSmith. Steeper learning curve than CrewAI — accept the tradeoff consciously.

Teams of Agents / Multi-Agent Orchestration

CrewAI

active73
crewAIInc/crewAI47K+ 4 evidence

Second-largest Python deployment footprint with real Fortune 500 adoption. Consistently rated 'fastest to prototype' — not marketing copy but backed by independent consensus. The right default for role-based business workflow automation teams prioritizing speed. Trade off: observability less mature than LangGraph without AMP Suite.

Teams of Agents / Multi-Agent Orchestration

OpenAI Agents SDK

activeOfficial71
openai/openai-agents-python20K+ 3 evidence

Best for OpenAI-committed teams building linear handoff chains — lowest friction path if locked into OpenAI models. Pre-1.0 API, no state persistence, no MCP. Do not choose if you need model-agnosticism, state persistence, or a stable API.

Teams of Agents / Multi-Agent Orchestration

Mastra

active73
mastra-ai/mastra22K+ 4 evidence

For JS/TS teams, this is not a comparison decision — it is the default. No other serious TypeScript-native multi-agent framework at scale. 442-pt HN thread is the strongest community validation signal in the entire frameworks category. Custom license (not MIT/Apache) — review before production use.

Teams of Agents / Multi-Agent Orchestration

Google Agent Development Kit (ADK)

activeOfficial68
google/adk-python18K+ 2 evidence

Strong download velocity for its age — GCP-native deployment and multi-language commitment give it a longer runway than single-language frameworks. Best for teams already on GCP/Vertex AI. No independently-verified production case studies outside Google-controlled publications.

Teams of Agents / Multi-Agent Orchestration

AWS Strands Agents SDK

watchOfficial53
strands-agents/sdk-python5.3K+ 2 evidence

AWS Bedrock teams only. High claimed downloads but anomalous download/star ratio (1,038 vs CrewAI 122) and zero HN organic discussion despite 14M cumulative downloads raises CI/CD pipeline inflation concern. Official AWS tooling is genuine advantage for Bedrock teams; lock-in penalty is high for everyone else.

Teams of Agents / Multi-Agent Orchestration

smolagents (HuggingFace)

watch78
huggingface/smolagents26K+ 2 evidence

Research and experimentation only. LocalPythonExecutor must NOT be used in production under any circumstances. Two independent security firms (JFrog + NCC Group) confirmed this. Docker or E2B sandboxing is an architectural requirement. Best for: evaluating CodeAgent paradigm, HuggingFace model experimentation, academic research.

Teams of Agents / Multi-Agent Orchestration

n8n

active83
n8n-io/n8n180K+ 2 evidence

Clear #1 for workflow automation with AI nodes. If your goal is to wire together SaaS tools with AI orchestration, n8n is the answer. If you're building an agent system from code, use LangGraph or CrewAI instead. Do not mix these use cases.

Teams of Agents / Multi-Agent Orchestration

Brave Search API

activeOfficial64
brave/brave-search-mcp-server797 5 evidence

#1 search API for AI agents — fastest latency (669ms), highest independent benchmark score (14.89), SOC 2 Type II attested. Default recommendation for general-purpose agent search.

Search & News

SearXNG

active78
searxng/searxng27K+ 3 evidence

#4 in search-news — the only option where no query ever leaves your infrastructure. 26,644 stars, active development. Not independently benchmarked on quality, but unmatched on privacy and cost.

Search & News

Tavily

activeOfficial48
N/A 4 evidence

#5 in search-news — biggest distribution advantage via LangChain default integration. AIMultiple 13.67 (#5) shows meaningful gap below top-4. Nebius acquisition adds strategic uncertainty.

Search & News

Jina Reader

staleOfficial43
jina-ai/reader10K+ 4 evidence

#6 in search-news — simplest possible single-URL extraction. ReaderLM-v2 is strong for edge/on-device. But OSS repo has had no commits for 10+ months and Firecrawl is a strict superset. On a downward trajectory.

Search & News

Crawl4AI

active83
unclecode/crawl4ai62K+ 3 evidence

#7 in search-news — the open-source self-hosted choice. 62K stars, Apache-2.0, actively maintained (v0.8.5 released 2026-03-18). Three independent comparisons show lower success rate vs Firecrawl (89.7% vs 95.3%) but wins on license, cost, and developer control.

Search & News

Parallel AI Search

activeOfficial38
2 evidence

Below cut line in search-news. Top-4 AIMultiple quality (14.21) but 20x Brave's latency makes it async-only. Closed source, no MCP server. Best deep research API if latency doesn't matter.

Search & News

You.com Search API

activeOfficial28
2 evidence

Below cut line in search-news. Research API launched Feb 26, 2026 with strong self-reported benchmark claims (#1 DeepSearchQA). Too new for independent verification — no HN traction, no AIMultiple placement. If claims verified, jumps to ranked list immediately.

Search & News

Linkup

watchOfficial28
3 evidence

Below cut line in search-news. 86K weekly downloads suggest real adoption despite 43 GitHub stars. Strong angel backing (Datadog CEO, Mistral CEO). All benchmark claims self-reported — no independent verification yet.

Search & News

Perplexity Sonar API

activeOfficial48
2 evidence

Below cut line in search-news. Consumer brand doesn't translate to API performance. AIMultiple #7 (12.96), 11K+ ms latency, BrowseComp 8%. HumAI shows highest accuracy (87%) but at cost of speed.

Search & News

Valyu DeepSearch

watchOfficial28
1 evidence

Below cut line in search-news. Unique proprietary data angle (50+ sources: SEC, clinical trials) but almost all evidence is self-reported. Minimal traction. Needs independent verification.

Search & News

Serper

activeOfficial28
1 evidence

Below cut line in search-news. Budget Google SERP wrapper. No semantic understanding, no independent index. Useful only when cost is the primary constraint.

Search & News

Spider Cloud

active38
niconiahi/spider2.3K+ 1 evidence

Below cut line in search-news. Niche high-volume crawling tool. Benchmark claims entirely self-reported. Tiny community (2.3K stars).

Search & News

Google Grounding with Search

activeOfficial28
1 evidence

Below cut line in search-news. Platform lock-in (Gemini API only). Not a standalone search API. If Gemini Deep Research exits preview with MCP support, deep research lane shifts dramatically.

Search & News

Bright Data MCP

activeOfficial51
2.2K+ 3 evidence

Below cut line in search-news. #1 in Browser MCP Benchmark (100% extraction success, 90% automation). 2,214 stars, 60+ MCP tools. Not a search API — web access infrastructure. The only option for scraping sites with aggressive anti-bot defenses.

Search & News

Hyperbrowser MCP

watchOfficial30
hyperbrowserai/mcp749 3 evidence

Below cut line in search-news. Strong 90% browser automation (tied Bright Data #1) but 63% web extraction (worst among ranked). GitHub stale 4+ months (last push Nov 2025). Watchlist — strong HN traction and stealth capabilities, but maintenance risk.

Search & News

ScrapeGraphAI

activeOfficial68
ScrapeGraphAI/Scrapegraph-ai23K+ 4 evidence

Below cut line in search-news. 23K stars, 194 HN pts (strongest HN score in category), unique LLM graph pipeline approach. But only 14.6K weekly PyPI downloads vs Firecrawl's 752K — star count likely inflated by viral novelty. Stars-to-downloads ratio 1,580:1 vs Firecrawl's 81:1.

Search & News

Kimi Code (Moonshot AI)

watchOfficial38
7K+ 1 evidence

Watch — real download signal (124K PyPI/week) and strong model capability (K2.5, HN 388 pts). Limited Western ecosystem integration and no SWE-bench Pro or Terminal-Bench scores. Best for Chinese developer ecosystem or teams using Moonshot AI models.

Coding CLIs / Code Agents

Kilo Code (Kilo-Org)

watch43
17K+ 1 evidence

Watch — real download signal (131K npm/week) and meaningful seed funding ($8M). OpenRouter-native architecture is a genuine differentiator for teams wanting model flexibility without managing API keys. Needs stronger differentiation beyond OpenRouter integration.

Coding CLIs / Code Agents

Cursor (Anysphere)

activeOfficial28
N/A 1 evidence

Reference only in coding-clis — Cursor is primarily an IDE, not a CLI agent. $29.3B valuation and strong adoption are real, but closed-source, paid, and IDE-first puts it outside the terminal-native category. Best for developers who want a polished commercial IDE with integrated AI.

Coding CLIs / Code Agents

Warp (Warp Technologies)

activeOfficial43
26K+ 1 evidence

Reference only in coding-clis — Warp is an AI terminal, not a coding CLI agent. Strong UX and 75.8% SWE-bench Verified are real signals, but 4,350 open issues and closed-source licensing are concerns. Best for developers who want an AI-first terminal experience rather than a code agent.

Coding CLIs / Code Agents

Kiro (AWS)

watchOfficial38
N/A 2 evidence

Watch — spec-driven approach is architecturally sound and GovCloud presence is meaningful for regulated industries. But too early to rank higher without a verified benchmark, user count, or independent case study. Outage controversy unresolved.

Coding CLIs / Code AgentsSoftware Factories

RooCode (RooVeterinaryInc)

active53
RooVeterinaryInc/roocode22K+ 2 evidence

#6 coding CLI — the 5.0/5 VS Code rating on 1.37M installs is the strongest quality signal in the IDE-embedded segment. Cline fork means it inherits a proven codebase while differentiating on team governance and multi-model flexibility. Best for teams adopting Cline-style agentic coding with stricter governance requirements.

Coding CLIs / Code Agents

Know a skill we're missing?

Submit evidence and we'll run the full research pipeline.

SUBMIT →