Virtual Workshop
Product Demo
AI impact: Measure what matters
GitHub Copilot here, Cursor there, Claude Code somewhere else. AI across tools is adopted, but is it actually working?
Join us live for a live demo on connecting AI activity to delivery outcomes, and apply a new framework for validating and scaling AI impact on speed, predictability, and DevEx.
(All registrants will receive the recording after the demo.)
Session
8am PDT on March 26, 2026
Speakers
Ben Lloyd Pearson
Director, Developer Experience, LinearB
Ofer Affias
Senior Director of Product, LinearB
About the demo
The real challenge isn’t adoption; it’s understanding what happens after.
In this session, you’ll learn how to connect AI activity to commits, PRs, and delivery outcomes, and walk away with an operating model to prove and scale impact.
Measure impact
Compare delivery, throughput, and quality by AI involvement, tool, workflow, repos, and team.
Turn metrics into action
Establish cadences to evaluate and scale without sacrificing predictability or developer health
Answer whether your AI investments are working
Your leadership wants ROI. Your team wants clarity. This session shows you how to get the data and operating model to deliver both.
Reserve your spot
How much of your code is AI-assisted, and where is it contributing to code, reviews, and PRs?
Which tools are making the most impact across your delivery pipeline?
Which users, teams, and workflows are making the most AI impact?
Is AI improving throughput, or simply shifting work from coding to review and rework?