For CTOs and Engineering Leaders
How CTOs Can Use AI Without Losing Their Edge
You can generate code faster with Copilot, but your team cannot skip the work of understanding whether that code should exist at all. The real risk is not AI tools themselves. The risk is treating tool output as a substitute for the architectural thinking that separates a good technology decision from a costly one.
These are suggestions. Your situation will differ. Use what is useful.
Know What You Are Not Outsourcing
Your job is to make architecture choices when the trade-offs are real and the right answer is unclear. AI tools are excellent at writing code that fits a given spec. They are useless at deciding whether microservices are the right choice for your organisation, or whether you should build or buy a piece of infrastructure. When you ask Claude or ChatGPT about build versus buy, you are asking a tool that has no knowledge of your constraints, your team's skills, or your actual cost structure. The tool will give you a reasonable-sounding answer. That answer belongs in a document you circulate to your team for criticism, not in a decision you make alone.
- ›Before using an AI tool to evaluate a major architecture decision, write down the three constraints that matter most for your organisation. Then check whether the AI's answer addresses those constraints specifically.
- ›When your team proposes an AI-generated solution to a design problem, ask them to present the alternative they rejected and why. If they cannot answer, they used the tool as a shortcut to thinking.
- ›Keep a running list of architecture decisions that went wrong in your organisation. Review each one. Ask whether an AI tool would have prevented the mistake or caused it.
Require Your Engineers to Debug What They Do Not Understand
Cursor and Copilot can write working code very quickly. They cannot explain why the code works, and they cannot help when the code fails in a way the tool did not anticipate. If your engineers can use an AI pair programmer but cannot debug the output when it breaks in production, you have created a knowledge gap that will compound over time. Your team needs to spend enough time with the generated code to develop instinct about what it does and why it does it that way. That instinct is what separates engineers who can follow a tool's output from engineers who can judge it.
- ›Require code review to include one question about the generated code: what would break if we changed X. If the reviewer cannot answer, the code stays in review.
- ›When an AI-generated component fails in testing or production, make it a learning event for the team. Have the engineer who wrote it lead a session on what the AI missed and how to spot similar gaps.
- ›Set a rule that AI-generated infrastructure code must be tested in a staging environment by someone other than the person who generated it. The goal is to catch surprises before they reach production.
Use AI to Speed Up Work You Already Know How to Do
The safest use of these tools is the narrowest use. When you ask Copilot to generate a helper function based on a pattern your team uses every day, you are getting a speed increase on work you already understand completely. When you ask ChatGPT to suggest an entire system design you have never built before, you are getting a plausible-sounding answer with no way to verify its soundness. The cognitive risk grows as the scope of the question grows. Your team should use AI tools to eliminate tedious work. They should not use them to replace the thinking work that you hired them to do.
- ›Define a list of tasks your team performs regularly that are repetitive and low-risk: boilerplate code, test fixtures, configuration files, documentation templates. These are safe targets for AI acceleration.
- ›For any new architectural pattern your team has not used before, require a design review before anyone writes code. An AI tool can generate code faster, but it cannot replace the human judgment needed to evaluate a new approach.
- ›Measure the time your team saves with AI tools on routine tasks. Also measure the time spent fixing, debugging, or redesigning AI-generated work on complex problems. Let the numbers guide your policy.
Build Your Team's Ability to Reason From First Principles
The engineers most at risk from AI tools are those who never learned to solve problems without them. If your team has always had access to a fast AI assistant, they may lack the deep experience of working through a hard problem with incomplete information and no easy answers. That experience builds the instinct that lets someone spot a plausible-sounding but wrong suggestion. You need to create space in your engineering culture for the kind of thinking that cannot be automated. Code review, architectural design sessions, and postmortems are good places to practise this thinking out loud.
- ›In engineering interviews, include a problem where the candidate must explain their reasoning step by step. If they jump straight to a solution without articulating why they rejected alternatives, their instinct is weak.
- ›Schedule a monthly architecture discussion where your team debates a major decision without opening any AI tools. The goal is to hear different opinions and see how people reason under uncertainty.
- ›When you hire senior engineers, look for people with experience in at least two very different technology environments. That experience is a foundation for first-principles thinking that transfers across tools.
Treat Vendor Claims About AI Capabilities as Marketing, Not Truth
When GitHub or Anthropic tells you that their tools increase engineering productivity by 40 percent, that number comes from a study that measures one specific thing in one specific context. It does not tell you how much time your team will actually save, whether the saved time offsets the time spent managing the risks, or whether the kind of work your team does is even the kind that benefits from this tool. You need the ability to evaluate those claims independently, without relying on a vendor's analysis. That means some of your senior engineers must test these tools themselves and report honestly on what works and what does not.
- ›Before adopting any new AI tool, have your strongest engineers spend two weeks using it on real work. Ask them to track time spent generating code versus time spent fixing, reviewing, or rewriting it.
- ›When a vendor claims a productivity gain, ask them for the specific metric and the specific context. Then design a small internal test that mimics your actual work patterns. Your results may differ significantly.
- ›Keep your licensing decisions separate from your enthusiasm decisions. It is reasonable to be excited about a tool. It is different to commit your entire engineering team to it based on marketing claims.
Key principles
- 1.AI tools are fastest at work you already understand completely. They are least reliable at work that requires new judgement.
- 2.Your organisation's architecture decisions must be made by people who understand your constraints, not by people reading an AI tool's output.
- 3.Engineers who use AI tools but cannot debug the output are becoming dependent on the tool rather than developing mastery.
- 4.The cognitive risk is not that AI tools will make bad decisions. The risk is that your team will stop developing the instinct to recognise when a tool has made a bad decision.
- 5.Vendor metrics about AI productivity are measured in isolation. Your job is to measure whether these tools actually improve outcomes in your specific context.
Key reminders
- Require your architecture review process to include a question about alternatives. If the team cannot explain what they rejected and why, the decision goes back to the drawing board.
- Create a list of failed architecture decisions from your organisation's history. Ask your team whether an AI tool would have prevented each failure or caused it. This builds instinct.
- Use AI tools to automate tedious work that your team already knows how to do. Do not use them to replace the thinking work that distinguishes good engineers from average ones.
- When evaluating a new AI tool, run an internal pilot with your strongest engineers. Their honest assessment matters far more than a vendor's benchmark.
- Schedule regular sessions where your team reasons through hard problems without using AI tools. This is where instinct develops and where you spot weaknesses in their first-principles thinking.