NewsletterBlogGlossary

How Claude Code Is Reshaping Engineering at Ramp, Shopify, Spotify, and More

How engineering teams at Ramp, Rakuten, Brex, Wiz, Shopify, and Spotify are using Claude Code to ship faster, reduce toil, and rethink developer workflows.

DEV
ShareXLinkedIn

How Claude Code Is Reshaping Engineering at Ramp, Shopify, Spotify, and More

A year after launching as a research preview, Claude Code is no longer an experiment — it's infrastructure. A roundup thread cataloging adoption at Ramp, Rakuten, Brex, Wiz, Shopify, and Spotify reveals a consistent pattern: teams aren't just using Claude Code for autocomplete. They're restructuring how engineers spend their time, what counts as "developer work," and how fast code moves from idea to production. The results offer a practical blueprint for any engineering org evaluating agentic coding tools.

What Happened

The thread, compiled by @_catwu, aggregates public statements and engineering blog insights from six companies spanning fintech, e-commerce, security, and music streaming. Each company adopted Claude Code at different scales and for different use cases, but the through-line is clear: measurable acceleration of shipping velocity with fewer context switches for engineers.

At Ramp, engineers report using Claude Code for everything from boilerplate generation to complex refactoring across their fintech codebase. Brex, operating in a similar financial infrastructure space, has integrated Claude Code into daily development workflows where compliance-sensitive code still requires human review but the scaffolding and test generation are largely automated.

Shopify — one of the earliest and most vocal adopters — has embedded Claude Code deeply into its engineering culture. CEO Tobi Lütke has publicly stated that AI tool proficiency is now a baseline expectation for Shopify engineers, and Claude Code is a primary tool in that stack.

Rakuten represents the enterprise-scale Asian market adoption, while Wiz, the cloud security unicorn, demonstrates that even in security-critical domains, teams trust Claude Code with meaningful portions of their workflow. Spotify rounds out the list, bringing the tool into one of the world's most complex content and recommendation platforms.

This wave of adoption coincides with Claude Code hitting a significant milestone: 4% of all public GitHub commits are now authored by Claude Code, according to SemiAnalysis.

Why It Matters

The significance isn't that big companies are trying AI coding tools — everyone is. It's how they're using them and what organizational changes follow.

Three patterns emerge across these six companies:

1. Toil elimination, not replacement. None of these teams replaced engineers with Claude Code. Instead, they offloaded the work engineers hate: writing boilerplate, generating test scaffolding, migrating between API versions, updating documentation. The time recaptured goes into architecture decisions, code review, and system design — the high-leverage work that experienced engineers do best.

2. Code review becomes the bottleneck. When engineers can generate code 3-5x faster, the review queue explodes. Several of these companies have invested in better review tooling and processes in parallel with Claude Code adoption. New features like HTTP hooks and the upcoming /simplify skill help address this by automating parts of the review cycle.

3. Junior engineers level up faster. With Claude Code handling syntax and boilerplate, junior developers spend more time understanding why code is structured a certain way rather than struggling with how to write it. The tool acts as a real-time mentor, showing idiomatic patterns and explaining trade-offs when asked.

For the competitive landscape, this enterprise adoption creates a flywheel. More production usage generates better feedback, which drives better tooling — like Claude Code Remote, scheduled tasks in Cowork, and the recently launched memory on free plans. Competitors like Cursor and GitHub Copilot face a gap not just in model capability but in the workflow integration layer.

Technical Deep-Dive

What separates enterprise Claude Code usage from individual adoption is the configuration and customization layer. Teams at this scale rely heavily on three mechanisms:

CLAUDE.md project files. Each repository gets a CLAUDE.md that defines build commands, environment constraints, coding standards, and workflow rules. This ensures every engineer on the team gets consistent Claude Code behavior without sharing prompt templates in Slack. The file lives in version control and goes through code review like any other code.

Skills system. The skills/ directory pattern lets teams encode domain-specific knowledge — API design guidelines, test patterns, documentation standards — into reusable SKILL.md files. A fintech team at Ramp or Brex can have a skills/compliance-review/SKILL.md that enforces regulatory requirements every time Claude touches payment-related code.

Hooks and automation. The recently launched HTTP hooks system lets teams build custom guardrails: blocking commits that don't pass linting, triggering security scans on generated code, or routing certain types of changes to specific reviewers. This is the infrastructure layer that makes "Claude Code in production" actually safe.

The /simplify and /batch skills announced this week address two remaining pain points. /simplify reviews changed code for reuse opportunities and quality issues — automating part of the code review cycle. /batch enables running the same operation across multiple files or repositories, critical for large-scale migrations that enterprise teams face regularly.

Performance-wise, Claude Code's one-year anniversary post notes usage spanning "weekend projects to Mars rover drive planning" — a range that speaks to both the tool's flexibility and its reliability under production constraints.

What You Should Do

  1. Start with CLAUDE.md, not prompts. If your team is using Claude Code without a project configuration file, you're getting generic output. Spend an hour documenting your build commands, coding standards, and key constraints in a CLAUDE.md at your repo root.

  2. Identify your top 3 toil tasks. Look at what your engineers complain about most — test writing, migration scripts, boilerplate. Build a skill for each one with a concrete few-shot example.

  3. Invest in review, not generation. The bottleneck will shift to code review within weeks. Set up the /simplify skill and establish clear review guidelines for AI-generated code.

  4. Measure before and after. Track PR cycle time, not lines of code. The companies seeing the best results measure time-to-merge, review turnaround, and developer satisfaction — not raw output volume.

  5. Try Claude Code Remote if your team works across environments. The ability to edit production code from any device changes the on-call experience significantly.

Related: Today's newsletter covers the broader AI news landscape. See also: Claude Code vs Cursor for a detailed feature comparison.


Found this useful? Subscribe to AI News for daily AI briefings.