Episode 001: The Agentic Software Development Revolution

18 minutes
agentic software developmentAI coding agentsCrewDeckAI agent orchestrationgit worktree isolation
0:00--:--

Episode 001: The Agentic Software Development Revolution

[ALEX]: What if you could run twenty AI developers simultaneously, each with their own workspace, and never worry about a single merge conflict? That's the promise of agentic software development in 2026. And today, we're going to dig into whether that promise is actually becoming real. Welcome to the very first episode of CloudNNJ. I'm Alex.

[JAMIE]: And I'm Jamie. Thanks for being here for episode one. We've got a big topic today. Agentic software development — what it actually means, why it's different from the copilot era we just lived through, and how teams are starting to manage this new world where AI agents aren't just suggesting code, they're writing entire features autonomously.

[ALEX]: So let's set the stage. By the end of 2025, roughly eighty-five percent of developers were regularly using AI tools for coding. That's according to a JetBrains survey of nearly twenty-five thousand developers. But here's the thing — most of that usage was still in copilot mode. Tab to accept a suggestion. Maybe a chat window for questions. The tool was reactive, right?

[JAMIE]: Exactly. And 2025 changed that in a really fundamental way. We saw the rise of what people are calling agentic coding tools. Claude Code, Cursor's agent mode, GitHub Copilot's new coding agent, Devin 2.0, OpenHands — these tools don't just suggest lines of code. They reason about your entire codebase, modify multiple files, run terminal commands, execute tests, and iterate based on feedback. All while you go get coffee.

[ALEX]: The New Stack called 2025 the "Agentic CLI Era." And the numbers back that up. The AI agents market is projected to grow from about eight billion dollars in 2025 to over fifty-two billion by 2030. That's a forty-six percent compound annual growth rate. But what's interesting to me isn't the market size — it's how the developer's role is changing.

[JAMIE]: Yeah, this is the part that I think a lot of engineers are still processing. When you have one AI agent, it feels like a really powerful tool. You give it a task, it writes some code, you review it, you move on. But the moment you try to run two agents, or five, or twenty — suddenly you're not a developer anymore. You're an engineering manager for a team of AI workers. And you've got none of the tooling that human engineering managers take for granted.

[ALEX]: That's such a great way to frame it. Gartner reported a fourteen hundred percent surge in multi-agent system inquiries from the first quarter of 2024 to the second quarter of 2025. They're predicting that forty percent of enterprise applications will embed AI agents by the end of 2026, up from less than five percent in 2025. The demand is real. But the orchestration problem is equally real.

[JAMIE]: So here's the thing about orchestration. Running a single agent is easy. You've got Claude Code or Cursor, you give it a prompt, it goes to work. But when you try to run multiple agents in parallel, you hit these really practical problems. First, where does each agent work? If they're all in the same working directory, they'll step on each other's files. Second, how do you track what each agent is doing? You can't just have twenty terminal windows open and try to keep track mentally. Third, how do you route the right task to the right agent? Not every agent is great at everything.

[ALEX]: And this is where we start seeing a gap in the current tooling landscape. If you look at the competitive matrix, it's pretty clear. Linear and Jira are phenomenal at project management, but they have zero AI agent execution capabilities. On the other side, Cursor and GitHub Copilot are excellent at AI-powered coding, but they have no project management layer. There's no Kanban board. There's no task lifecycle. There's no way to organize work across multiple agents and track it the way you'd track work across a human team.

[JAMIE]: Right. And that gap is exactly what CrewDeck is going after. They're positioning themselves as "JIRA for AI Agents" — and I have to say, when I first heard that tagline, I was skeptical. But when you actually look at what they've built, it clicks. It's a desktop application where AI agents are first-class team members. You've got a visual Kanban board with tasks flowing through To Do, In Progress, In Review, Document, and Done columns. But instead of assigning tasks to human developers, you're assigning them to AI agents.

[ALEX]: Let's talk about the technical architecture, because I think that's where this gets really interesting. The foundation of the whole system is git worktree isolation. For listeners who might not be familiar, git worktrees are a feature of Git that lets you check out multiple branches simultaneously in separate directories, all sharing the same underlying repository. So instead of one working directory where you're constantly switching branches, you can have twenty separate working directories, each on its own branch.

[JAMIE]: And this is becoming the de facto standard for parallel AI development. There was a really popular article on Medium late last year called "Git Worktrees: The Secret Weapon for Running Multiple AI Coding Agents in Parallel." Cursor 2.0 introduced multi-agent support via worktrees back in October 2025. The Nx team wrote about how worktrees changed their AI agent workflow. It's everywhere. But here's what CrewDeck does differently — they automate the entire worktree lifecycle. When you create a task and assign it to an agent, CrewDeck automatically provisions a worktree, creates a feature branch, runs the agent in that isolated directory, and when the agent finishes, it commits, pushes, and opens a pull request. You never touch a worktree command manually.

[ALEX]: And they claim support for up to twenty concurrent sessions. That's twenty agents, each in their own worktree, each with their own terminal session, all running simultaneously. Now, I want to be clear — we haven't independently verified that number under production load. But the architecture supports it because each worktree is a completely isolated workspace. There's no file locking, no race conditions on the working directory.

[JAMIE]: The other piece of their architecture that I find technically impressive is the terminal execution engine. Most AI coding tools, when they need to run a command, they use something like Node's child process exec function. It's a simple API — you pass in a command string, you get back the output when it finishes. CrewDeck uses PTY-based terminal sessions instead. PTY stands for pseudo-terminal. It's the same technology that powers your actual terminal emulator. What this means in practice is that agents can receive user input mid-execution. You can interact with the agent while it's working — answer a prompt, provide clarification, redirect it. It's not a fire-and-forget system.

[ALEX]: That's a meaningful distinction. Think about when an AI agent hits a decision point — maybe it's not sure which authentication library to use, or it encounters an ambiguous requirement. With a simple exec approach, the agent either guesses or fails. With interactive PTY sessions, the agent can pause and ask you, or you can jump in and redirect. It's that human-in-the-loop capability that enterprise teams are demanding. And that actually connects to a broader industry trend.

[JAMIE]: Yeah, the human-in-the-loop versus full autonomy debate is one of the most active conversations in the space right now. IEEE Spectrum ran a great piece asking whether coding AI tools will ever reach full autonomy. The consensus across the industry seems to be what people are calling "bounded autonomy." You want your AI agents to work independently on well-defined tasks, but you need approval gates, configurable autonomy levels, and clear audit trails. CrewDeck's task lifecycle captures that well — the agent builds the feature, but then the task moves to "In Review," where a human reviews the pull request before anything gets merged.

[ALEX]: Speaking of review, let me highlight another feature that I think is underappreciated — their multi-agent debate capability. This was introduced in a recent PR. The idea is that for complex architectural decisions, you can have multiple AI agents literally debate each other. Say you're deciding between a microservices architecture and a monolith. You assign the systems architect agent, the security auditor agent, and the backend developer agent to a structured debate. Each agent proposes an approach, critiques the others' proposals, and then the system synthesizes a consensus. The output is a markdown document committed to a docs-debate folder in your repo.

[JAMIE]: That's a really creative approach to the single-model bias problem. When you only have one AI agent making decisions, you get whatever biases that agent has baked in. By having multiple specialized agents weigh in from different perspectives, you get a richer analysis. I'm not going to claim it replaces a real architecture review with experienced engineers, but as a starting point for discussion? That's genuinely useful.

[ALEX]: Now, let's talk about multi-tool support, because this ties into another huge industry trend. CrewDeck supports Claude Code, Cursor, Gemini CLI, and GitHub Copilot. You can pick the right tool for the right task. Maybe you use Claude's Opus model for complex architectural work and Gemini for quick documentation updates. The platform auto-detects which tools you have installed, lets you set defaults at the project level, and even includes a command migration service that translates CLI flags between different tools.

[JAMIE]: This is smart because the industry is clearly moving toward multi-model strategies. Andreessen Horowitz reported that thirty-seven percent of enterprises are now using five or more AI models, up from twenty-nine percent the year before. The winning architecture, according to their research, is multi-model routing with fallbacks. You use Claude for reasoning-heavy work, GPT for content generation, Gemini Flash for high-volume cheap tasks. Nobody wants to be locked into a single vendor's roadmap. And with Anthropic's Model Context Protocol and Google's Agent-to-Agent protocol establishing interoperability standards, the ecosystem is moving toward openness.

[ALEX]: There's also the agent marketplace, which I think is one of CrewDeck's most strategically interesting features. It's essentially an App Store for AI development agents. You can browse pre-built agents organized by category — backend, frontend, DevOps, testing, documentation, security — install them with one click, rate them, and track updates. The agents are defined as markdown files with YAML frontmatter, which means they're version-controlled and travel with your repository.

[JAMIE]: The marketplace model is really compelling because it creates network effects. More agents attract more users, more users attract more agent creators, more creators build more agents. It's the same flywheel that made the iOS App Store and npm so powerful. And for the ecosystem, it means that specialized knowledge gets packaged and shared. If someone builds a really effective database optimization agent, every team in the marketplace benefits from it.

[ALEX]: Let's zoom out for a second and talk about what I think is the biggest challenge facing this whole space — enterprise adoption. KPMG's latest AI Pulse survey found that eighty percent of leaders cite cybersecurity as the single greatest barrier to achieving their AI strategy goals. Seventy-five percent prioritize security, compliance, and auditability as the most critical requirements for agent deployment. And here's a sobering stat from the research — even the best-performing large language model produces secure and correct code only about fifty-six percent of the time without security-specific prompting.

[JAMIE]: And this is where CrewDeck's local-first architecture is actually a significant advantage. Everything runs on your machine. Your code never leaves your local environment for processing by CrewDeck itself — the AI tools obviously call their respective APIs, but the orchestration layer is entirely local. For companies worried about code exfiltration or compliance requirements around data residency, that matters a lot. They've also got agent observability built in — real-time tracing of agent decisions, tool usage, and reasoning — which addresses the "black box" problem that makes enterprise security teams nervous.

[ALEX]: Their roadmap shows enterprise features landing in Q4 of this year — SSO, SAML integration, audit logging, fine-grained permissions. That's table stakes for any enterprise tool, but the fact that they're planning for it shows they understand the path from developer tool to team platform to enterprise solution.

[JAMIE]: So here's my honest take, and I want to be balanced here. CrewDeck is doing something genuinely novel — combining project management with AI agent execution in a way that nobody else really is right now. The git worktree isolation is the right architectural bet. The PTY-based execution engine is technically superior to what most competitors offer. And the multi-tool support means you're not locked in. But they're also pre-revenue, early stage, and competing against companies with massive distribution advantages. Linear could add AI agent features. GitHub could bolt on better project management. That competitive window they're talking about — six to twelve months — is real, and it's ticking.

[ALEX]: Absolutely. And I'd add that the orchestration problem they're solving is going to get harder, not easier, as agents become more capable. Today, you might run five or ten agents. Next year, it could be fifty. The tooling needs to scale not just technically but cognitively — engineers need to be able to understand and manage what dozens of agents are doing without getting overwhelmed. The Kanban board metaphor is a good start, but I think the next frontier is more intelligent orchestration — maybe an AI that manages the other AIs.

[JAMIE]: They actually have that with their swarm execution mode. Instead of running agents sequentially, you designate an orchestrator agent that dynamically coordinates the other agents, deciding who works on what and synthesizing the results. It's still early, but it points toward a future where you don't manually assign tasks at all — you describe what you want at a high level, and the platform figures out how to decompose it, assign it, execute it, and deliver it.

[ALEX]: So if you're an engineering leader evaluating agentic tools right now, what should you be thinking about? Let me throw out a few criteria.

[JAMIE]: First, isolation. However you're running AI agents, they need their own workspaces. Git worktrees are the current best practice. If a tool doesn't offer that, you'll hit merge conflicts at scale. Second, observability. You need to see what your agents are doing — not just the terminal output, but the reasoning behind their decisions. When a task fails, you need to debug it without guessing. Third, flexibility. Don't lock yourself into one AI model or one AI tool. The landscape is moving too fast. Whatever you adopt should support multiple models and multiple tools.

[ALEX]: And fourth, I'd add governance. Think about your approval gates. Where do humans stay in the loop? At what point does an agent need permission to proceed? The answer will be different for every team and every task type, but your tooling should make that configurable, not hardcoded.

[JAMIE]: Well said. And fifth — and this might be the most practical advice — start small. You don't need twenty agents on day one. Start with one or two, get comfortable with the workflow, understand the failure modes, and then scale up. The teams that succeed with agentic development are the ones that treat it like managing a junior engineering team. You wouldn't hire twenty juniors on the same day and expect everything to go smoothly.

[ALEX]: That's a perfect analogy to end on. Agentic software development is real, it's here, and it's going to reshape how engineering teams work in 2026 and beyond. The tooling is still maturing — and CrewDeck is one of the most interesting entries in the space because they're solving the orchestration layer rather than trying to build yet another AI coding agent. Whether they can maintain that lead as the big players start moving in the same direction — that's the story we'll be watching.

[JAMIE]: Absolutely. Thanks for joining us on this first episode of CloudNNJ. If this resonated with you, share it with your engineering team. And if you're experimenting with agentic workflows, we'd love to hear about your experience. Until next time.

[ALEX]: Until next time. Happy building.