Product Management with GenAI: What Changes and What Doesn't

Last Updated: April 2026

I’m seeing more and more products that were clearly built with the help of GenAI. They’re full of features that don’t fit well together. Some of the products I’ve seen don’t solve a real problem.

“Even when building is cheap, building the wrong thing is still expensive.”

At Product Camp Ottawa 2026, I asked 3 questions:

  1. How many of you work in organizations that shipped a feature in the last quarter with AI-generated code? About 1/3 to 1/2 of the people in the room
  2. How many of those features were built from a hypothesis that had been tested with real users before production code was written? Only 3 people raised their hands, roughly 15% of those who shipped with GenAI
  3. How many of you have spoken to an actual end user, human to human, in the last month? Nearly half, a good sign. However, this is different from what I see working with teams.

I learned that people at Product Camp are better than most at talking to their users. However, like the rest of the world, they’re struggling to ensure that the right thing is being built.

Before GenAI appeared, most teams I met had a bottleneck at Quality Assurance.

Diagram showing a bottleneck at Quality Assurance before GenAI
Before GenAI: the bottleneck sits at Quality Assurance.

With GenAI, this problem doesn’t change. We still have more features than we can possibly test. If the bottleneck shifts, it’s only because we now have an even worse Code Review problem.

Diagram showing GenAI pushing more work into Code Review and Quality Assurance
With GenAI: more code generated upstream of the same bottlenecks.

Worse, in many cases, features are getting rejected in the review process or fail in testing. The bottleneck is the place where we don’t have enough people to do the work. Failures discovered here are the most expensive because we wasted the time of the people doing the review and testing. Many of these failures are features that should never have been built or ones with a poor user experience.

Diagram showing rejected features and failed tests accumulating at the bottleneck
Failures found at the bottleneck are the most expensive.

Pressure to use GenAI everywhere it can be used is increasing. No one would suggest that putting more cars on the highway upstream of a bottleneck is a good idea. Yet that is exactly what we’re doing when we generate code. (Classic Theory of Constraints)

You end up with feature bloat that confuses users and feels unfinished. By increasing the number of features, we increase the attack surface and with it the security risk. (The more features we have, the more places an attacker can try to get in.)

“More features, faster, isn’t winning.”

How should Product Management Respond?

Every feature that is built is a hypothesis: “If we build this feature, it will solve a specific problem for a group of users.”

Agile Product Management has long had a set of Discovery tools to help with this:

  • Lean Startup: good for testing big hypotheses: should we build this product or feature?
  • Lean UX: what version of the feature will solve the problem? (It also puts emphasis on improving usability and user experience)
  • Continuous Discovery: what feature(s) are most important right now?
  • Behaviour Driven Development: exploring the acceptance criteria for the feature before implementation, so that we build the right thing. Generating code without doing this is a big risk
  • Impact Mapping: which feature will have the biggest impact for the business?

Even without GenAI, all of these tools would help a team build the right product for the right users. GenAI makes these tools easier to use, but it’s not magic.

What changes with GenAI?

Across all of these tools, I see that GenAI helps:

  • Make prototypes easier to build
  • Faster to run Lean Startup type experiments: Wizard of Oz, Fake Door, etc.
  • Possible to run multiple experiments at once
  • Analysis of large amounts of user data: customer interviews, surveys, analytics, etc.

GenAI also creates some risks:

RiskResolution
Prototypes seem polished and complete, so they get mistaken for certaintyGrayscale: build the prototype without colour. No one can mistake it for the real thing. I learned about this over 30 years ago, doing usability work. When we show a UI that looks good, we get feedback about font and colour. When it is deliberately simple, we get far more information about the usability.
Outsourcing all analysis to GenAI risks missing important insights and introducing bias and hallucinations1Do your own analysis before getting the AI to do one. This way, you’ll have a better understanding of the data. Use the AI as a cross-check for your completed analysis, not to replace it.
Using AI agents to simulate real usersDon’t. Talk to real users, even if only three of them, instead.
Pressure to go faster means that we skip key validation stepsRemind the people pressuring you that going faster will cost more later, when you need to revert a feature that doesn’t meet a user need or has poor usability.

The biggest risk is going through the motions: theatre, going through the discovery process without learning. Ask: what did we learn this week that changed our plan? If nothing, it’s theatre.

What doesn’t change?

The need for:

  • Customer contact, real world conversations and testing with people
  • Collaboration with the team building the product

GenAI speeds up the mechanics of Discovery. Speeding up the mechanics is great, but it means we need to spend more time thinking and learning.

To compete in this market, stop promising to build features just because a stakeholder asked for them. Have a clear testable hypothesis. Building great products still comes down to the same things: talking to real users, running experiments, and listening to the feedback.

What hypothesis will you test next? How will you use GenAI to help? What will you do to offset the risks?

Footnotes

  1. AI Generated Personas are a very specific example of this: “Generative AI personas considered harmful? Putting forth twenty challenges of algorithmic user representation in human-computer interaction” Danial Amin, Joni Salminen, Bernard J. Jansen, Joongi Shin, Dae Hyun Kim

Mark Levison

Mark Levison

Mark Levison has been helping Scrum teams and organizations with Agile, Scrum and Kanban style approaches since 2001. From certified scrum master training to custom Agile courses, he has helped well over 8,000 individuals, earning him respect and top rated reviews as one of the pioneers within the industry, as well as a raft of certifications from the ScrumAlliance. Mark has been a speaker at various Agile Conferences for more than 20 years, and is a published Scrum author with eBooks as well as articles on InfoQ.com, ScrumAlliance.org and AgileAlliance.org.

Get Certified

Explore what Scrum is and how to make it work for you in our Scrum Certification training. Hands-on learning will guide you to improve teamwork, deliver quick feedback, and achieve better products and results.

Registration is now open for workshops:

About this course

Focuses on the role of the team and the ScrumMaster. Get the skills and practical experience necessary to improve teamwork, take the exam, and advance your career with a certification that is in high demand today. Often the best fit for anyone new to Scrum.

Learning and Benefits

Relatable Scenarios

Learn on-the-job applications of key Scrum concepts, skills, principles, along with practical solutions that you can apply the next day for difficult, real-life situations.

Respected Certification

Everything you need to earn your Scrum Alliance® ScrumMaster certification, including exam fee and membership, and so much more.

Practical Exercises

With focus on the challenges that real teams face, and tools to dig deeper. You don’t need more boring Scrum theory. You need something you can sink your teeth into to see immediate results.

Jargon-Free Learning

This workshop is not just for software development or people with a computer science degree. We’ve helped many non-software teams with Scrum.

Career Advancement

Use Scrum knowledge to standout at work, get paid more, and impress your customer, all without burning out.

GenAI for Systems Thinking

Learn the basics of using Generative AI as a tool to support Systems Thinking in your ScrumMaster role. Explore how to leverage GenAI to help uncover patterns, and think more deeply about the systems your team operates in.

Ongoing Support

Your learning doesn’t stop when the workshop ends. You get lifetime access to all course materials, plus a followup email series designed to reinforce your learning objectives and help you apply what you’ve learned on the job.