Generative AI February 13, 2026

How Spotify engineers use Claude Code and Honk to stop writing code by hand

Spotify says some of its best developers haven’t written code by hand since December. Normally that would read like stage-managed exec talk. The details make it harder to dismiss. The internal setup, called Honk, lets engineers ask Claude Code from S...

How Spotify engineers use Claude Code and Honk to stop writing code by hand

Spotify’s AI dev stack is getting serious: iOS changes from Slack, Claude Code in the loop, and a very different job for senior engineers

Spotify says some of its best developers haven’t written code by hand since December. Normally that would read like stage-managed exec talk. The details make it harder to dismiss.

The internal setup, called Honk, lets engineers ask Claude Code from Slack to change the iOS app, build it, send the result back to Slack, and move it toward merge. Spotify’s example is blunt enough to stick: an engineer commuting in the morning can message Slack from a phone, ask for a bug fix or feature change, get a build back, and keep things moving before arriving at the office.

That’s a bigger shift than “AI helps with coding.” Plenty of teams have copilots in the editor. Far fewer have an agent wired into branching, testing, signing, packaging, and release controls. Spotify apparently does.

The company also has enough shipped work behind the claim to make it worth a look. Spotify says it rolled out more than 50 new features in 2025 and kept shipping in early 2026 with additions like Prompted Playlists, Page Match for audiobooks, and About This Song. Honk is part of the explanation for that pace.

The hard part is the plumbing

Slack is just the entry point. The hard part sits behind it.

For this to work on a production mobile app, Honk almost certainly acts as an orchestration layer between chat and the repo. The model needs repo context, tool access, a branch workflow, CI hooks, code-signing controls, and some policy layer that limits what it can touch.

The shape is familiar to anyone who’s spent time around internal developer platforms:

  • a Slack bot receives the request and authenticates the user
  • an orchestrator packages the task with repo context and rules
  • Claude Code edits code, likely with repo-aware indexing and structured tool calls
  • the system runs linting, unit tests, maybe UI tests and static analysis
  • a branch and PR get created with summaries and diffs
  • CI builds the iOS app, probably through something like fastlane
  • the artifact gets sent back through Slack via TestFlight or a similar internal channel
  • a human reviews and approves the merge

That’s a real delivery path. The model has moved beyond autocomplete and into a bounded software workflow.

And yes, the Slack part is replaceable. Swap in Teams, a web console, or a CLI and the core idea still holds. The important change is that the model can act inside guardrails instead of waiting in the IDE for a human to copy and paste.

Why this sounds plausible now

A year ago, this would have sounded flimsier.

The models got better at long-context work and tool use. They still drift, miss edge cases, and occasionally produce garbage with confidence. But with narrow instructions, solid repo context, and hard checks around them, they’re less likely to vandalize a codebase than they were in 2023.

The rest of the stack also caught up. Code search across large repos, language-aware parsing, CI integration, feature flags, canary patterns, policy engines, and artifact provenance were already standard in strong platform teams. Agents fit into that machinery. They depend on it.

That’s why Spotify’s claim feels believable now. The model isn’t being trusted in the abstract. It’s being boxed into a system designed to catch bad behavior.

There’s an obvious catch. This works best in engineering orgs that already have discipline. Clear ownership. Good tests. Reliable build pipelines. Strong release controls. If your repo is a junk drawer and CI is red three days a week, an AI agent won’t rescue you. It will create bad pull requests faster.

Senior engineers are moving up the stack

The more interesting change is the job itself.

If Spotify’s workflow holds up, senior engineers spend less time typing out view controllers, wiring API clients, or fixing repetitive bugs. More of their time goes into deciding what change should happen, checking whether the model’s approach makes sense, reviewing diffs, and managing risk.

That’s efficient. It also changes where engineering judgment sits.

The bottleneck shifts toward review quality, policy, and tests. Weak senior reviewers get expensive fast. Strong ones can supervise a lot more change because the AI handles the repetitive work. Teams that still think in lines changed or tickets closed are going to miss the point. Once deployment frequency rises, the better signals are failure rate, rollback speed, and whether the team can explain what the agent actually did.

The “developers haven’t written code” line is catchy and a little sloppy. They’re still doing engineering. They’re doing less transcription.

The weak spot is test quality

Agent-driven delivery sounds good until you ask the simple question: how does the system know the change is safe?

Mostly, it doesn’t. It knows the tests passed.

That puts a bright light on the test suite. If coverage is shallow, UI tests are brittle, or contract tests are missing around important boundaries, the model will happily make changes that clear local checks and still break production. The agent only looks competent if the safety net is real.

Mobile teams get extra headaches here. iOS builds come with signing, provisioning, environment-specific config, and all the usual flaky edges. If AI-generated code is moving through that path, signing keys and build credentials need tight isolation. The agent should not hold long-lived secrets directly. Short-lived tokens, controlled service boundaries, and narrow permissions are the sane setup.

Security teams will also care about the audit trail. Every agent action should be logged: prompt, files changed, commands run, tests executed, approvals granted, model version used. If something ships and goes sideways, “Claude probably changed it” won’t survive postmortem.

This is where software supply chain work stops looking like paperwork. Provenance records, SBOMs, and SLSA-style attestations matter when an autonomous system is generating and shipping code.

Spotify’s data point matters too

Spotify also said it’s building a non-commoditized dataset around music understanding. That sounds like standard earnings-call language. In this case, it’s a fair point.

Coding agents are getting interchangeable fast. Durable advantage sits in data and product context. Spotify has years of listening behavior, mood signals, geography, time-of-day patterns, skip behavior, playlist intent, and the messy edge cases around how people describe music. “Good workout music” isn’t one thing. It shifts by culture, audience, and context. A general-purpose model doesn’t know that on its own.

That matters on the product side with features like Prompted Playlists and song explanations. It matters on moderation too. Spotify says it’s labeling AI-generated music through metadata and policing spam. That’s not a side concern. If generative music floods the catalog, recommendation quality degrades fast unless the platform can classify provenance and filter low-value junk.

For developers, the lesson is simple enough: the model layer commoditizes faster than the data layer.

What other teams should copy

Most companies shouldn’t try to clone Spotify’s full workflow in one go. Start narrower.

Pick change classes where failure is cheap:

  • docs and internal tooling
  • localization updates
  • low-risk UI copy or layout changes behind feature flags
  • dependency bump PRs with mandatory review
  • boilerplate test generation where humans still own the assertions

Then tighten the system around those flows. Good auth. Role-based controls. Ownership maps. Explicit allowlists for which repos and files the agent can touch. Hard blocks around auth, payments, cryptography, infrastructure, and privacy-sensitive code unless a human gives direct approval.

Then fix testing debt. First, not later.

An AI agent plugged into CI will expose weak engineering fundamentals quickly. That’s useful, even if it’s mildly embarrassing.

The practical takeaway

Spotify’s Honk setup looks like an early version of what a lot of large engineering orgs will build: chat-triggered or task-triggered agents that make bounded changes, run through CI, return artifacts, and stop at a human approval gate.

The notable part is that Spotify seems to have wired the model into the full path from request to build well enough that senior engineers trust it often enough to change how they work.

That’s the threshold worth watching. The question is whether your team can let AI touch the release pipeline without losing control.

Keep going from here

Useful next reads and implementation paths

If this topic connects to a real workflow, these links give you the service path, a proof point, and related articles worth reading next.

Relevant service
AI model evaluation and implementation

Compare models against real workflow needs before wiring them into production systems.

Related proof
Internal docs RAG assistant

How model-backed retrieval reduced internal document search time by 62%.

Related article
Claude Code Rate Limits Now Include Weekly Quotas. What It Means for Teams

Anthropic has added two weekly rate limits to Claude Code on top of the existing five-hour rolling limit. For teams that lean on Claude Code for long coding sessions, refactors, agent loops, or CI-driven generation, that means a hard weekly ceiling n...

Related article
Anthropic launches Claude Code Review to manage AI-generated pull requests

Anthropic has launched Code Review inside Claude Code, now in research preview for Claude for Teams and Claude for Enterprise. The timing makes sense. AI assistants are churning out pull requests faster than most teams can review them, and a lot of t...

Related article
Why OpenAI moved from Cursor to Windsurf in a reported $3B deal

OpenAI reportedly tried to buy Anysphere, the company behind Cursor, before moving into acquisition talks with Windsurf at roughly $3 billion. That sequence matters more than the deal chatter. It suggests OpenAI is looking past the most popular codin...