GSD Is the Missing Piece For Claude Code

AAI LABS
Computing/SoftwareSmall Business/StartupsInternet Technology

Transcript

00:00:00AI coding frameworks keep multiplying, and every one of them claims to be the best way
00:00:04to build with agents.
00:00:05But the best way to build with agents doesn't entirely depend on the framework you pick.
00:00:09There are other factors people forget to evaluate, and then they get frustrated when the framework
00:00:14doesn't fit their project.
00:00:15But that's not the framework's problem, that's a selection problem.
00:00:18Each framework works well for the use case it was designed for, and there's no one size
00:00:22fits all here.
00:00:23Now our team has previously covered AI coding frameworks on the channel, but recently we
00:00:28came across one that's been gaining a lot of traction lately.
00:00:31The reason we're making a video on this is not because it's another best framework that
00:00:35is going to cause you to leave the other ones.
00:00:37It's because when we tested it, we found it to be genuinely different, and it fits use
00:00:41cases that these frameworks don't.
00:00:43We've covered multiple frameworks in previous videos, like BMAD and superpowers.
00:00:48Picking the wrong one means you're either over-engineering or under-preparing.
00:00:51Today we're going over this framework called GSD, which is short for Get Shit Done, and
00:00:56you'll understand why.
00:00:57By the end of this video, you'll understand where to use which framework, starting with
00:01:00the first one which is GSD.
00:01:02You use GSD where you aren't really sure what to build, and you don't want to pre-plan everything
00:01:06because the requirements might change in the future.
00:01:09Now this doesn't mean that you don't know what to build at all.
00:01:11It means that the product you want requires a lot of experimentation in each step.
00:01:15This is great if you want to quickly build out MVPs for experimental projects as quickly
00:01:19as possible.
00:01:20With GSD, it does ask you about a broader scope, but unlike the BMAD method, it doesn't really
00:01:25lock you in.
00:01:26Using the initial scope, it plans each implementation step by step.
00:01:30This means that you're not locked into the later phases because they have not been pre-planned
00:01:34by the system in extensive detail.
00:01:35So if you're building a custom solution that hasn't really been built before, it would
00:01:39be better if you go ahead with GSD.
00:01:41Let's say if you were building an on-screen interview assistant, like Cluely.
00:01:45You would need to figure out a lot of things as to how the user experience would look like
00:01:50or how to make sure that all screen sharing platforms don't detect it during screen sharing.
00:01:54A lot of things might need experimentation, with no way to determine them before the planning.
00:01:59BMAD, on the other hand, takes the opposite approach.
00:02:02It's a step-by-step framework that creates thorough documentation for each phase before
00:02:06implementing anything.
00:02:07You need to be sure about what you're building at the very start, and it's not like it
00:02:11messes up.
00:02:12It has a whole research department as a module where the creator of BMAD has poured in all
00:02:17the context about these different roles, such as a business analyst and design thinker that
00:02:21make sure that the product you want to make is thought about from all angles.
00:02:25But it's all preloaded, and based on that you make your PRD and architecture documents,
00:02:30and then based on those you get sharded tasks that the implementation team just needs to
00:02:34follow.
00:02:35When we made our video on the BMAD system, we praised it because of how locked up it was.
00:02:39The system is so well built in the prompts that it doesn't really let the agent deviate
00:02:43from the tasks.
00:02:44But over an extended period of usage, we've noticed that when you actually need a change
00:02:48in the requirements, that is when the system becomes unstable, because even the best of
00:02:52the models actually miss little details when changing the requirements.
00:02:56Another complaint that people mention is that it takes too much time to plan out everything
00:03:00before any implementation actually starts, but that's because it shouldn't be used
00:03:04for those sorts of projects.
00:03:05You should be using it when you're absolutely sure of the requirements and just want the
00:03:09system built without any loopholes in it, because all the specs that it makes are heavily interlinked.
00:03:14Or when you want to build conventional systems, such as a custom CRM solution for a client
00:03:19or your own custom community platform.
00:03:21Now, Superpowers is based on TDD and the whole point of the ideology is that you know what
00:03:25you are building.
00:03:26TDD is important where the cost of an edge case is too high to miss, and this cost doesn't
00:03:30include a simple Stripe integration with the Next.js app, but let's say something like
00:03:34an agentic platform where AI agents take actions on behalf of your users and a wrong action
00:03:39can't be undone and is too costly.
00:03:41In terms of usability, it's like GSD where the planning is done feature by feature with
00:03:46just an outline of what the project needs to be.
00:03:49But since tests are generated first, it doesn't really give you that room to experiment with
00:03:53different things and isn't ideal for projects where the planning is done beforehand.
00:03:57But another thing that you can do is that if the project lies in both domains is that you
00:04:02can implement a version of the project with GSD where the main functionality is completed
00:04:06and then you can continue the rest of the app by onboarding Superpowers on the app.
00:04:10We do have separate videos on both Superpowers and the BeMad method, and they would be linked
00:04:15down in the description below if you want to explore these in more detail.
00:04:19GSD was also built with preventing context rot in mind, by using sub-agents, spawning
00:04:23separate processes for isolated tasks, keeping the main agent's context clean so that it can
00:04:28focus on what actually matters without drifting.
00:04:31Majority of the AI coding agents now support sub-agents, so it doesn't matter if you use
00:04:35Claude code with this or not.
00:04:37But Claude also received another update recently, the new Opus 4.6 with a 1 million context
00:04:42window.
00:04:43So a lot of techniques that we've taught on this channel to actively manage your context
00:04:46don't really matter that much anymore.
00:04:49So to install it, you're going to copy the command and paste it in the project folder
00:04:53where you are working.
00:04:54Then select which agent you want to install it for.
00:04:57In my case it was Claude, so I installed it for that.
00:04:59You then have to choose what should be the scope of the installation.
00:05:02I prefer project level because each project may use a different framework and this keeps
00:05:06the configuration scoped to the one chosen for this project.
00:05:10Since we were using Next.js for building, we installed it in our newly initialized project.
00:05:15And once installed, the GSD framework should show up in your .claud folder in the form of
00:05:19agents, commands, and hooks.
00:05:21If you are not using Claude, they will be stored in the .agent folder in the root of the project.
00:05:26It has multiple agents intended for different specific tasks, but unlike normal markdown
00:05:30based prompts, everything is structured in XML format, and this is done on purpose.
00:05:34Claude models are known to work better with XML formatted instructions since it lets them
00:05:39parse the structure more easily, so it's a performance optimization towards whatever
00:05:43agent that you're using.
00:05:44But before we move forwards, let's have a word from our sponsor, Genspark.
00:05:48Most people are currently stuck juggling multiple AI subscriptions, paying individual fees for
00:05:52ChatGPT, Claude, and various media tools.
00:05:55Genspark is an all-in-one AI workspace that hit $200 million ARR in just 11 months by consolidating
00:06:02everything into a single subscription.
00:06:03It uses a super agent system to choose the best model for any task.
00:06:07It handles your entire workflow, from building AI slides and automating AI sheets, to generating
00:06:12AI media, and even making real-world phone calls with the "call for me" feature.
00:06:16You can even turn your research into AI pods to listen to on the go.
00:06:20Plus, use Speakly to turn a 15-minute recording into 3,000 words of perfect text instantly.
00:06:26The wildest part?
00:06:27Genspark offers unlimited AI chat and image features throughout 2026.
00:06:31Manobanana2, GPT-Image, Flux, Seedream, Gemini 3.1 Pro, GPT 5.2, Claude Opus 4.6, and more
00:06:40top models are all included with unlimited access.
00:06:43Hit the link in the pinned comment to start building with Genspark.
00:06:46To start using GSD, you're going to use this new project command.
00:06:50By running this command, the agent started working according to the initialization prompt
00:06:54and started with exploring the codebase first.
00:06:56Since I had initialized the Next.js app already, it detected that there was code present and
00:07:01asked if it should explore the existing codebase first.
00:07:04We told it to skip the mapping to avoid wasting time.
00:07:06If you were working on an existing project, it would be better to ask it to map the codebase
00:07:11first, but since it contained the blank template, I skipped it.
00:07:14After that, it will ask you to tell it the app idea you want to work on.
00:07:18After which, it asks a lot of questions regarding the target audience of your app, its features,
00:07:22the scope of the project, and more.
00:07:24But the question-answer session GSD conducts is noticeably different from that of the superpowers.
00:07:29Superpowers try to extract edge cases from you during this phase.
00:07:32GSD does not, and it focuses more on understanding what to build, not stress testing how it might
00:07:38break.
00:07:39Once it gathers enough information from the session, it creates the project.md inside the
00:07:43dot planning folder, which contains the description, out of scope items, context constraints, and
00:07:48key decisions all deduced from the questioning.
00:07:50But here's where the context rot prevention shows up in practice.
00:07:54It keeps project.md deliberately short and focused, so the agents wouldn't deviate from
00:07:59the main aims by getting buried in documentation.
00:08:01It tracks all the work through git, but it doesn't use a straightforward git commit.
00:08:05It runs a script underneath that performs checks before committing, ensuring everything is up
00:08:10to the standards before actually using git to commit.
00:08:12Once planning was done, GSD moved to the research phase.
00:08:16It spawns multiple agents researching different aspects of the app in parallel, all running
00:08:20in the background.
00:08:21After the research completes, the research synthesizer agent kicks in.
00:08:24Since all these agents are tuned to use appropriate models for the tasks, the synthesizer agent
00:08:29uses the sonnet model, not the heavier opus.
00:08:32GSD has a proper reference mapping each subagent to the right model based on the weight of work
00:08:36it handles so that it doesn't waste extra efforts.
00:08:39It condenses the research findings and also flagged potential issues, things to watch out
00:08:44for that could hinder the work long term.
00:08:46After the research is synthesized, the requirements stage begins.
00:08:49It asks you targeted questions about the MVP, identifying which features are truly essential
00:08:55for version 1.
00:08:56Because GSD is focused on fast delivery, it makes sure V1 contains only what is necessary.
00:09:01After it confirms the MVP with you, it then generates the roadmap structure, which you
00:09:05have to approve, and your approval marks the project initialization as complete.
00:09:09Also, if you are enjoying our content, consider pressing the hype button, because it helps
00:09:14us create more content like this and reach out to more people.
00:09:17Now that the initialization phase is over, the next step is to implement the plan.
00:09:22In the previous phase, it had broken the project down into 4 phases based on the requirement.
00:09:26It will give you two ways to work, either skip discussions or proceed with discussions.
00:09:31Discussion for this framework means questioning sessions with the agent where it ensures it
00:09:35understands the plan properly.
00:09:37We chose to discuss first so that we can clarify the requirements as much as we can, but you
00:09:41can choose to skip discussions if you think the previous questioning session was enough
00:09:45for it to build the app.
00:09:47Once the discussion session is over, it creates a context.md file in the phases folder inside
00:09:52the .planning folder.
00:09:54This file contains all the details of what we have discussed and mapped the phase we just
00:09:58talked about with the agent.
00:09:59But true to GSD's philosophy, this file is also concise in length so that Claude can stay
00:10:04focused on what actually matters.
00:10:06Next, you begin the planning of phase 1.
00:10:08The planning phase starts with the research based on the context.md file it just created
00:10:13using a dedicated research agent with the sonnet model exploring different aspects.
00:10:17But we thought it would use context 7 or actually look up against the docs properly.
00:10:21Instead, it was using web search with the year 2025 in the search keywords which it shouldn't
00:10:27have been doing.
00:10:28It would have been more controlled if we had connected context 7 MCP so that its research
00:10:32was better grounded.
00:10:33So when you are using make sure to connect with such MCP for better grounding.
00:10:37It created a research.md file and documented all of its research, including the confidence
00:10:42level for its research, which indicated how trustworthy are its sources.
00:10:46After the research, it created a plan.
00:10:48These plans listed in detail the dependencies of each phase and all the requirements with
00:10:52proper IDs.
00:10:53But here's the thing, this is where GSD is different because instead of just writing the
00:10:57plan one way like others do, it actually cross verifies the plan across multiple dimensions
00:11:02and validates whether the plan is actually implementable and aligned with the goals.
00:11:06It uses dedicated planning and verifying plan agents, where the planning agent creates the
00:11:11plans and the verifier keeps checking them, sending warnings back.
00:11:14So it basically performs the adversarial planning on its own without us having to control any
00:11:19of that manually.
00:11:20Once the plan was final and all the steps had passed, it committed and broke the plan
00:11:24down into two waves.
00:11:25It actually breaks down into as many waves as required and parallelizes those that are
00:11:30independent, so that sub agents can handle them in parallel.
00:11:33It started implementing the project waves using the dedicated agent to carry out the plan
00:11:37it created.
00:11:38Once it is finished, it verifies the checkpoint using proper playwright testing underneath,
00:11:43creating scripts, deleting them after using to clean the folder, and performing multiple
00:11:47automated tasks under the hood.
00:11:49It then gave us a summary of what was built and provided instructions on how to verify
00:11:53it for ourselves.
00:11:54The first iteration builds the app that is just a placeholder, with all the initial elements
00:11:58visible to show how the app would actually look.
00:12:01Further iterations will build one aspect of the app at a time, leading to a complete app
00:12:05at the end of the cycle.
00:12:06The entire iteration used 138,000 tokens, which was actually not much given a 1 million token
00:12:12context window.
00:12:13But with an agent with a 200,000 context window, it would signal that it's time to compact.
00:12:18But since this relies on documentation of phases, even if we clear the context, the agents would
00:12:23know where to start.
00:12:25When we approved the work, it ran multiple tests and marked wave 2 as complete as well.
00:12:29After that, it spawned the GSD verifier again and cross-checked whether the implementation
00:12:34actually matched the initial goal.
00:12:36Once verified, it marked phase 1 as complete, including the verifications, and prompted us
00:12:41to proceed to the next phase of the application.
00:12:43Now once we ran the app through all of the phases, the previously placeholder features
00:12:47were all functional and working as intended.
00:12:49GSD works well when you're working on a large-scale application with multiple features, but it's
00:12:54overkill if the app you're building is much simpler and doesn't need the heavy lifting.
00:12:59For simpler apps, Claude or any other agent on its own is sufficient and doesn't need this
00:13:03thorough planning.
00:13:04However, if you want controlled execution with proper planning without wasting too much
00:13:08effort on this phase, you should definitely go for GSD.
00:13:11Now this is one framework, but oftentimes there are use cases where the existing ones
00:13:16don't hold up and you need to build one of your own.
00:13:18For that, you need to know certain principles before actually building your own.
00:13:22We covered those principles in a previous video that would help you build better workflows.
00:13:26You might end up seeing that video on the end screen so you can just click on it instead
00:13:30of looking it up.
00:13:31That brings us to the end of this video.
00:13:33If you'd like to support the channel and help us keep making videos like this, you can do
00:13:37so by joining AI Labs Pro.
00:13:38As always, thank you for watching and I'll see you in the next one.

Key Takeaway

GSD provides a balanced, modular approach to AI-driven development by prioritizing fast delivery, experimental flexibility, and automated verification over the rigid, exhaustive pre-planning found in other frameworks.

Highlights

GSD (Get Shit Done) is a flexible AI coding framework designed for experimental projects where requirements are likely to change.

Unlike the rigid BMAD method, GSD plans implementation step-by-step to avoid locking users into a pre-defined, extensive detail that might fail during requirement shifts.

The framework prevents context rot by using sub-agents and spawning separate processes for isolated tasks, keeping the main agent's focus clean.

GSD employs adversarial planning where a 'verifier' agent constantly checks the 'planning' agent's work to ensure implementability and goal alignment.

The system uses XML-formatted instructions to optimize performance for Claude models, which parse structured data more effectively than standard markdown.

Post-implementation, GSD utilizes Playwright testing under the hood to verify checkpoints and ensure the code matches the initial goals.

Timeline

Introduction to AI Coding Frameworks and Selection

The speaker introduces the growing landscape of AI coding frameworks and emphasizes that project success depends on selecting the right tool rather than just the 'best' one. Many developers face frustration because they pick frameworks that do not align with their specific project needs, leading to over-engineering or under-preparation. The video introduces GSD (Get Shit Done) as a unique alternative that fills gaps left by popular methods like BMAD and Superpowers. This section highlights that each framework has a specific design intent, and understanding these differences is crucial for efficient agentic workflows. By the end of this segment, the viewer is primed to learn why GSD is considered a 'missing piece' in the current ecosystem.

GSD vs. BMAD: Flexibility vs. Structure

This section contrasts the GSD framework with the BMAD method, focusing on their respective approaches to planning and experimentation. GSD is recommended for MVPs and custom solutions like interview assistants where user experience requires iterative discovery and pre-planning is difficult. In contrast, BMAD is described as a high-structure framework that creates exhaustive documentation and sharded tasks before any implementation begins, making it ideal for conventional systems like CRMs. However, the speaker notes that BMAD can become unstable when requirements change because the interlinked specifications are hard for models to update without missing details. GSD avoids this 'lock-in' by planning implementation steps dynamically as the project progresses.

Superpowers and Context Management

The speaker explores the Superpowers framework, which is rooted in Test-Driven Development (TDD) and is best suited for high-stakes projects where edge cases are costly. While Superpowers offers safety, it lacks the experimental room provided by GSD, though the two can be used sequentially to build and then stabilize an app. A major technical highlight in this section is GSD's strategy for preventing 'context rot' through the use of sub-agents and isolated task processing. Interestingly, the speaker mentions that with the release of Claude Opus 4.6 and its 1-million-token context window, some manual context management techniques are becoming less critical. This context allows the agent to maintain a broader understanding of the codebase without losing track of specific implementation details.

Installation and Structure of GSD

Instructions for installing GSD are provided, involving a simple terminal command and selection of the target agent, such as Claude. The speaker recommends a project-level installation to keep configurations scoped to specific project needs, noting that GSD files appear in the .claud or .agent folders. The framework's internal architecture relies on agents, commands, and hooks structured specifically in XML format to enhance parsing performance for Claude models. This structural choice is a deliberate performance optimization, as Claude handles XML better than standard markdown-based prompts. This technical setup ensures that the sub-agents can operate with high precision and low error rates during the execution phase.

Sponsorship: Genspark AI Workspace

The video features a segment on Genspark, an all-in-one AI workspace that aims to eliminate the need for multiple AI subscriptions by consolidating various top-tier models. The platform uses a 'super agent' system to select the most appropriate model for a given task, whether it's generating slides, automating sheets, or making real-world phone calls. Notable features mentioned include 'Speakly' for voice-to-text conversion and 'AI pods' for listening to research on the go. The sponsor offer includes unlimited access to flagship models like Gemini 3.1 Pro and GPT 5.2 through 2026. This section serves as a practical example of the 'agentic' trends discussed throughout the framework analysis.

Project Initialization and Research Phase

The GSD workflow begins with a project initialization command that maps the existing codebase and conducts a unique interview session with the user. Unlike Superpowers, which focuses on edge cases, GSD's questioning session focuses on understanding the target audience, features, and project scope. It then generates a concise 'project.md' file in a dedicated planning folder to keep the agent focused without burying it in excessive documentation. The framework tracks progress via Git, but it employs an underlying script to perform standard checks before any commits are finalized. This ensures that the documentation remains a lightweight and functional guide for the agents rather than a source of confusion.

Parallel Research and Requirement Analysis

During the research phase, GSD spawns multiple agents to investigate different aspects of the application simultaneously in the background. A 'research synthesizer' agent, typically running on a lighter model like Sonnet to save resources, condenses these findings and flags potential long-term hindrances. The requirements stage follows, where GSD identifies the truly essential features for a Version 1 (V1) release to ensure fast delivery. The user is then asked to approve a roadmap structure, which marks the completion of the initialization phase. This parallelized approach demonstrates GSD's efficiency in gathering intelligence before moving into the actual coding waves.

Implementation, Adversarial Planning, and Testing

Implementation is divided into phases and waves, with GSD offering users the choice to participate in clarifying discussions or proceed automatically. A standout feature of GSD is 'adversarial planning,' where a dedicated verifier agent checks the plans against the goals and sends warnings if the implementation seems unaligned. Once a plan is validated, sub-agents handle independent tasks in parallel and use Playwright testing to verify checkpoints and clean up temporary scripts. The first iteration often results in a functional placeholder UI, which is then refined in subsequent cycles into a fully functional feature set. The speaker notes that an entire cycle used approximately 138,000 tokens, which is well within the limits of modern large-context models.

Conclusion and Strategic Framework Advice

In the final section, the speaker summarizes that GSD is a powerful tool for large-scale applications with multiple features but might be 'overkill' for very simple apps. For basic projects, a standard AI agent without such rigorous planning frameworks is often sufficient. The speaker encourages viewers to understand the principles behind these workflows, as they may eventually need to build their own custom frameworks for specific use cases. The video concludes with a recommendation to watch a previous video on building better AI workflows and an invitation to join AI Labs Pro for further support. This wrap-up reinforces the theme that framework selection is a strategic decision based on project complexity and the need for controlled execution.

Community Posts

View all posts