skillpack.co
All solutions

Claude Code

active

Anthropic's official agentic coding CLI. v2.1.81 (Mar 20) shipped `--bare`, smarter worktree resume, and improved MCP OAuth while the repo crossed 82,204 stars and logged ~14 commits/week across 10+ maintainers. Terminal-native, tool-use-driven, with deep file system + shell access, #1 SWE-bench Pro standardized (45.89%), ~4% of GitHub public commits (SemiAnalysis), $2.5B annualized revenue. 8M+ npm weekly downloads. Opus 4.6 with 1M context.

Score 98trending
Claude Code in action

Where it wins

#1 SWE-bench Pro standardized (45.89%) — the authoritative benchmark now that Verified is saturated

~4% of GitHub public commits — strongest real-usage signal (SemiAnalysis), $2.5B annualized revenue (fastest enterprise SaaS to $1B ARR — Constellation Research)

8M+ npm weekly downloads — 3x Codex, 12x Gemini. Installs are the hardest metric to game

Opus 4.6 with 1M context — highest-capability model available in a CLI

HN peak engagement 2,127 pts — unmatched community mindshare

~95% first-pass correctness (Educative.io) — roughly double Gemini CLI's 50-60%

Wins independent blind code quality tests 67% of the time across independent evaluations

Independent daily quality monitoring with no degradation detected (MarginLab, p<0.05)

Real-world daily cost ~$6 avg, <$12 at p90

Where to be skeptical

Rate limits are the #1 complaint (1,085 pts HN thread dedicated to this)

Costs $200+/month at heavy usage — 2-3x more expensive per task than Codex CLI due to 3-4x higher token consumption

CVE-2025-59536 (CVSS 8.7) — hooks injection RCE, fixed v1.0.111. CVE-2026-21852 — API token exfiltration, fixed. Two patched security incidents

Terminal-Bench 2.0: 74.7% (#3) — improved but still trails Gemini CLI (78.4%) and Codex CLI (77.3%)

Tied to Anthropic models only

Anthropic blocked third-party subscription use, angering users (625 pts HN)

Quality regression perception: 'Claude Code is being dumbed down?' HN thread (1,085 pts, 702 comments, Feb 2026) — MarginLab monitoring shows no statistical regression but community trust is a live issue

Editorial verdict

The #1 coding CLI agent. Leads SWE-bench Pro standardized (45.89%), wins independent head-to-heads on reasoning depth, ~4% of GitHub commits. Rate limits are the #1 complaint. Costs 2-3x more per task than Codex CLI due to higher token consumption.

How to get started

Install via npm install -g @anthropic-ai/claude-code (or run curl -fsSL https://claude.ai/install.sh | bash), then run claude init inside the repo you want it to manage to link git + MCP configs. Start every task with claude plan --issue <ticket> so it drafts the plan before you approve claude apply to write diffs.

Videos

Reviews, tutorials, and comparisons from the community.

Claude Code Tutorial #1 - Introduction &amp; Setup

Net Ninja·2025-08-21

Claude Code Tutorial for Beginners

Kevin Stratvert·2025-12-22

Claude Code - Full Tutorial for Beginners

Tech With Tim·2026-02-27

Related

Public evidence

strong2026-03-20
v2.1.81 release ships `--bare`, smarter worktree resume, improved MCP OAuth

Latest release notes confirm the new `--bare` flag, worktree resumption improvements, MCP OAuth fixes, and multi-shell channel support — proof that Anthropic continues to ship practical workflow upgrades every week.

82,204 stars, 6,873 forks, 100 commits in prior 7 weeksAnthropic release engineers
strong2026-01-07
HN 46532075 + GitHub issue #16673 show outages resolved in public

Users documented a v2.1.0 startup failure ('update to claude 2.1.0 then run claude. see the error.'). Maintainers acknowledged it and patched it quickly (https://github.com/anthropics/claude-code/issues/16673), reinforcing the transparent reliability story.

181 HN points, 76 commentsHacker News community & Anthropic maintainers
strong2026-03-22
74-like "How to Master Claude Code" thread keeps onboarding viral

Step-by-step walkthrough of read/edit/run primitives sparked comparisons with Copilot CLI and Gemini CLI in replies, proving the CLI workflow is still the reference onboarding story for newcomers.

74 likes, 8 repliesAI workflow evangelist community on X
strong2025-10
Claude Code 2.0 — 842 points, 413 comments on HN

Major front-page HN thread on the Claude Code 2.0 release. Community consensus positioned it as the standard agentic coding CLI all others are measured against.

842 HN points, 413 commentsHN community (submitted by polyrand)
strong2026-03
SWE-bench Pro standardized: Claude Code #1 at 45.89%

Claude Code leads SWE-bench Pro standardized at 45.89% vs Codex CLI's 41.0%. SWE-bench Verified is saturated (top 5 within 1 point). Pro standardized is the fair comparison — custom scaffold scores (Codex 56.8%) are not comparable.

Authoritative standardized benchmark (SEAL)SWE-bench team (standardized evaluation)
strong2026-03
SemiAnalysis: ~4% of GitHub public commits, $2.5B annualized revenue

~4% of public GitHub commits, projected 20%+ by EOY 2026. 42,896x growth in 13 months. $2.5B annualized revenue. The hardest real-usage metric in the category.

Industry analysis (SemiAnalysis — premier semiconductor/AI research)SemiAnalysis (independent research firm)
strong2026-03
npm weekly downloads: 8,028,705 — 3x Codex CLI, 12x Gemini CLI

8M+ weekly downloads is the hardest-to-game traction metric in the category. 3x Codex CLI (2.6M), 12x Gemini CLI (647K). Installs correlate with actual developer usage, not social media hype.

npm registry datanpm registry (official package statistics)
strong2026-03
Educative practical test: Claude Code 1h17m single-shot vs Gemini CLI 2h02m

Claude Code completed a React project in 1h17m single-shot. Gemini CLI took 2h02m with multiple retries. Demonstrates Claude Code's reasoning advantage on practical tasks beyond synthetic benchmarks.

Independent practical comparison testEducative (independent developer education platform)
strong2026-03
Morph 15-agent comparison: Claude Code — 'deepest reasoning on hard problems'

Morph's independent Tier 1 classification. Called Claude Code 'deepest reasoning on hard problems — best for complex refactors, unfamiliar codebases, architectural decisions.' Only Claude Code and Codex CLI made Tier 1.

Independent 15-agent comparison studyMorph Engineering Team (independent AI infrastructure company)
strong2026-02
HN: 'Claude Code is being dumbed down?' — 1,085 pts, 702 comments (live trust issue)

High-engagement thread questioning whether Claude 4.5 has regressed in quality. MarginLab's independent monitoring found no statistically significant degradation (p<0.05), but community perception of quality regression is real and ongoing. The trust cost of this thread is itself a factor.

1,085 HN points, 702 comments — top thread in the coding CLI categoryHN community
strong2026-03
Morphllm independent test: 'best AI coding agent for most developers'

Morphllm's independent comparison of 15 coding agents: Claude Code classified as 'best AI coding agent for most developers.' Same model can score 17 problems apart in different agents — scaffold maturity is the key differentiator.

Independent 15-agent comparison studyMorphllm (independent AI evaluation platform)
moderate2026-02
SemiAnalysis: 42,896x growth in 13 months, 4% of GitHub public commits

"4% of GitHub public commits are being authored by Claude Code right now... 42,896x growth in 13 months." (Dylan Patel, SemiAnalysis, Feb 2026). The single most cited real-usage metric in the category.

Industry analysis (SemiAnalysis — Dylan Patel)SemiAnalysis (independent research firm)
strong2026-02-26
Apple Xcode 26.3: Claude Code natively integrated — platform distribution play

First native IDE integration by a major platform vendor. Developers get full Claude Code in Xcode — including subagents, background tasks, and plugins. Gives distribution to all macOS/iOS developers.

369 HN pts, Apple Newsroom + Anthropic BlogApple (official partnership), Anthropic (official announcement)

Raw GitHub source

GitHub README peek

Constrained peek so you can sanity-check the source material without leaving the site.

Claude Code

Claude Code is an agentic coding tool that lives in your terminal, understands your codebase, and helps you code faster by executing routine tasks, explaining complex code, and handling git workflows -- all through natural language commands. Use it in your terminal, IDE, or tag @claude on Github.

Learn more in the official documentation.

<img src="https://raw.githubusercontent.com/anthropics/claude-code/main/demo.gif" />

Get started

[!NOTE] Installation via npm is deprecated. Use one of the recommended methods below.

For more installation options, uninstall steps, and troubleshooting, see the setup documentation.

  1. Install Claude Code:

    MacOS/Linux (Recommended):

    curl -fsSL https://claude.ai/install.sh | bash
    

    Homebrew (MacOS/Linux):

    brew install --cask claude-code
    

    Windows (Recommended):

    irm https://claude.ai/install.ps1 | iex
    

    WinGet (Windows):

    winget install Anthropic.ClaudeCode
    

    NPM (Deprecated):

    npm install -g @anthropic-ai/claude-code
    
  2. Navigate to your project directory and run claude.

Plugins

This repository includes several Claude Code plugins that extend functionality with custom commands and agents. See the plugins directory for detailed documentation on available plugins.

Reporting Bugs

We welcome your feedback. Use the /bug command to report issues directly within Claude Code, or file a GitHub issue.

Connect on Discord

Join the Claude Developers Discord to connect with other developers using Claude Code. Get help, share feedback, and discuss your projects with the community.

Data collection, usage, and retention

When you use Claude Code, we collect feedback, which includes usage data (such as code acceptance or rejections), associated conversation data, and user feedback submitted via the /bug command.

How we use your data

See our data usage policies.

Privacy safeguards

We have implemented several safeguards to protect your data, including limited retention periods for sensitive information, restricted access to user session data, and clear policies against using feedback for model training.

For full details, please review our Commercial Terms of Service and Privacy Policy.

View on GitHub →