AWS sits at the heart of the generative AI boom, powering everything from LLM training runs to global-scale inference. In this sweeping conversation, AWS CEO Matt Garman discusses the future of work, engineering, open vs. closed models, and why agentic workflowsānot just raw modelsāwill be where the next wave of value is created.
Garman is bullish on the economic upside of AI, skeptical of doomer narratives, and refreshingly candid about infrastructure bottlenecks, engineering culture, and how Amazon uses its own silicon to support customer choice.
Key Moments from the Interview
01:00 ā White Collar Bloodbath or Utopia?
Garmanās optimistic view on AI, jobs, and productivity.
04:15 ā Hiring in the Age of AI
Why more productivity doesnāt mean fewer people.
08:52 ā 80% of AWS Developers Use AI
How AI is changing developer workflows at Amazon.
12:55 ā Should You Still Study Engineering?
Advice for the next generation in a fast-changing tech landscape.
15:46 ā Infrastructure Bottlenecks
From silicon to power, whatās actually constraining AI growth?
18:12 ā Where AI Usage Is Growing Most
Training matters, but inference drives demand.
20:05 ā AWS Silicon Strategy
Graviton, Tranium, Inferentia, and why Annapurna matters.
27:57 ā Serving the Model Ecosystem
Bedrock, specialization, and what AWS looks for in new models.
33:39 ā Open vs. Closed Source
How AWS views the trade-offs and partnerships.
36:24 ā Will AWS Build a Frontier Model?
On Nova, customer choice, and competition with partners.
41:33 ā Benchmarks Are Breaking
Why standardized evaluations may not matter much longer.
47:13 ā The Future of Agents
The biggest opportunity in AIāand how AWS is enabling it.
Full Interview: White Collar Jobs, Hyperscalers, AI Coding, Open vs Closed, Agents, and more! (Matt Garman)
In His Own Words: What Matt Garman Revealed
The Real Impact of AI on Work (01:00)
AI will remove toil, not jobs.
Most white collar jobs today involve work no one actually wants to do. AI can take that away and let people focus on what matters.
Hiring Wonāt Slow Down (04:15)
Productivity gains will fuel more opportunityānot layoffs.
Thereās no mass unemployment because Excel exists. AI is more disruptive, but the same principle applies: weāll move up the value chain.
Developer Usage: Over 80% (08:52)
AI coding tools are mainstream inside AWS.
North of 80% of our developers are using AI in some part of their workflowāunit tests, documentation, full-on agentic coding.
Engineering Still Matters (12:55)
But itās about mindset, not memorizing syntax.
If you think youāre going to specialize in one thing for 30 yearsāyouāre wrong. Learn how to think, how to learn, and how to build.
Bottlenecks Keep Shifting (15:46)
Today itās silicon. Tomorrow? Could be power.
Thereās no one bottleneck. Solve one, and the next appears. Thatās the nature of scaling infrastructure at AWS.
Inference Is King (18:12)
Training headlines. Inference dominates cost.
Most of the growth today is inference. Thatās what drives compute usageāthe moment a customer hits an app or asks a question.
AWS Silicon: A 10-Year Bet (20:05)
From Nitro to Graviton to Tranium, Annapurna made it all possible.
Ten years later, the Annapurna team still works with us. Our best acquisition ever.
Why Model Choice Matters (27:57)
AWS isnāt betting on one model to rule them all.
We want customers to use the best model for the job. Thatās why Bedrock includes everything from Anthropic to Writer to Luma AI.
Open vs. Closed Isnāt the Real Question (33:39)
The real value comes from customization.
Whether itās open weights or fine-tuned APIs, customers want to shape models to their own workflows. Thatās what matters.
Will AWS Build a Frontier Model? (36:24)
Nova is the startābut partnerships will still thrive.
Weāve built a muscle: compete where necessary, partner everywhere else. Thatās how weāve run AWS for 19 years.
Benchmarks Are Nearing Irrelevance (41:33)
Models have learned how to ace them.
Benchmarks work well for SSDs. They donāt capture complex behavior. Weāre already seeing models saturate MMLU, AM24, and others.
Agent Workflows Are the Next Platform Shift (47:13)
AWS is going all-in on agent infrastructure.
We launched Agent Core in Bedrock to make enterprise agents scalable, secure, and auditable. Thatās where real AI ROI will come from.
Key Takeaways
AI Will Enhance Work, Not Replace It
Garman sees AI as an augmentation toolāespecially in creative, analytical, and engineering workflows.
Agent Infrastructure Is the Next Big Platform
From secure runtimes to memory and gateways, AWS is building the scaffolding for enterprise-scale agents.
Model Diversity > One-Model Dominance
Rather than an omni-model future, AWS is betting on specializationāoffering customers choice through Bedrock.
Open vs. Closed Is a Spectrum
Customization, not philosophy, is what customers care about. Whether via open weights or closed APIs, itās about fit.
Benchmarks Are Losing Signal
As models converge on test scores, real-world performance and UX integration matter more than academic leaderboards.
Silicon Strategy Is Core to AWSās Edge
Owning the full stackāfrom chip to UIāgives AWS flexibility, pricing control, and customer alignment unmatched by competitors.
Full Transcript
01:00 ā Future of Work / White-Collar Jobs
MB: I wanted to talk first about the future of work and white-collar work. Thereās a spectrumāāwhite-collar bloodbathā to utopia. Where do you fall and why?
MG: Iām on the optimistic side. In tech thereās never been a more exciting time. Across industry, the advances in AI have enormous potential to increase efficiency, effectiveness, and enablement at work.
A lot of what AI promises is taking away the toil in day-to-day jobs. Today, the vast majority of time isnāt spent on what people get excited aboutāitās not āput my numbers in this system,ā āpull the report,ā or ācollate information.ā That overhead takes a large percentage of time. AI can really help shrink that, letting people focus on the creative and analytical parts they love. That drives value for companies and people.
Iām very optimistic. This is not āno one has a job and robots run the world.ā Companies and people become more efficient, and people spend more time on what theyāre excited about.
03:05 ā Hiring in a High-Productivity Era
MB: If you automate large swaths of tasks, you can do more as a company. Does that mean you wonāt hire anymore? How do you think about hiring when every person is more productive?
MG: It dependsāthereās no single answer for every company. Historically, efficiency gains come with a transition. The critical thing is for people to be flexible, willing to learn, and accept jobs will evolve. The job from two years ago wonāt be identical two years from now.
There isnāt mass unemployment because we have computers, automation, or robotics. There are lots of jobsāoften higher-paying. The economy is bigger; on average people are better off. Think of Excel: many used to spend time doing calculations. Excel didnāt eliminate those jobs; it changed them. AI is more disruptive than Excel, but the analogy holds: people move to higher-value work.
People are worried; I donāt minimize that. Embrace the technology. The more you do, the better off youāll be. AI has potential to transform every industry, company, and job. It transforms, not replaces. If you donāt lean in, you might be out of a job; if you do, it makes you better/faster and lets you do more of what you like. Thatās better for companies, people, and economic growth.
05:51 ā Speed of Change
MB: Pessimists say AIās speed is different from prior shifts. Will speed really affect the white-collar market?
MG: Itās a rapidly evolving space; people will have to move faster. For developers worried coding tools will make them unnecessary, I think weāll need more developers, not fewer. The job changes: maybe you wonāt author Java code; youāll deconstruct problems, coordinate agents, and build systems.
The part you may not do in two or three years is authoring Java codeātools will be great at that. But pulling it together, reviewing, deciding itās not quite right, coordinating agentsāthat becomes the developerās job. That person drives more value. These tools unlock creativity: turning ideas into action takes time; if we unlock that, good developers become even more valuable.
MB: As a leader, if someone becomes 5ā10x more productive, the last thing you want is fewer of them.
MG: Exactly. You invest more, because the ROI compounds. I donāt understand the math for wanting fewer of those people.
08:52 ā How Much Code Is āWritten by AIā?
MB: How much of AWSās code is written by AI? And define āwritten.ā
MG: āLines of code written by AIā is a silly metricālines arenāt the point (fewer can be better). The last metric I saw: north of 80% of our developers use AI in their workflowsāunit tests, docs, writing code, or agentic flows via things like QCLI or our Neuro IDE. That number goes up weekly.
MB: Are engineers proactively upskilling, or do you run programs?
MG: Both. Amazonās large and not homogeneous, but most developers are curious. The number who havenāt used an AI coding tool rounds to zero. Using it to transform the job vs. using piecesāthatās where education matters. Thereās a learning curve: how to change work, when it accelerates vs. slows you down.
First-gen tools can be linear: you āvibe code,ā it gives you code, but itās not what you want; hard to āgo back.ā With an agentic coding-first mentality, you start with a spec, then work with the tool to build parts of it. As you vibe code, it updates the specābut the spec remains the source of truth you can modify.
This also guides junior developers toward best practices. Some leaders told me, āWith AI we can replace all junior people.ā Thatās the dumbest thing Iāve ever heardātheyāre least expensive, most leaned into AI tools, and you need a pipeline of talent. Keep hiring out of college; teach them to decompose problems and build software. Tools like Q can coach good practice and help juniors collaborate with seasoned engineers.
12:55 ā Should Students Study Engineering?
MB: Do you recommend engineering as a career to someone entering college?
MG: Yesāthough kids should study what theyāre passionate about. The emphasis should be: think for yourself, develop critical reasoning, creativity, and a learning mindset. With todayās pace, if you learn one thing and plan to ride it for 30 years, it wonāt be valuable 30 years from now. Learn how to learn and how to think. Engineering is great for systems thinking and problem decomposition.
MB: Internally, how do you measure success using AI?
MG: We donāt have a magic new measure. Some of it is productivity. We encourage experimentationātools, methods, setups. Previously, large systems needed many people focusing on pieces. Now, tools enable smaller, faster pods with broader scope. Startups move fast because of structureābig orgs can, too, if they organize as small pods. Weāre leaning into that.
15:46 ā Capacity & Infrastructure Buildout
MB: Looking 2ā5 years out: what are the bottlenecksāsilicon, energy?
MG: Think of The Goal (book) on production lines: thereās always a bottleneckāsolve it and the next appears. All of those (chips, power, networking) are bottlenecks at some point. Recent shortage: NVIDIA chips; as that eases, power could be next. Itās hard to get everything in sync at the growth rate weāre seeing; all require capital.
Our job is to think 1, 3, 5 years out: ensure enough power, capacity, and network for customers. Sometimes weāre wrong, sometimes weāre shortābut we take that on so customers donāt have to.
18:07 ā Where Growth Is Coming From
MB: Where is growthāRL, inference, training?
MG: Inference drives most usage. Training and fine-tuning are important, but compute demand is dominated by the end-user interactionāquestions, app usage, workflows. Infrastructure doesnāt care if itās RL, FT, or inferenceāsame silicon serves multiple needs (networking can differ for big pipelines). Tranium is great for training and inference (despite our naming); NVIDIA chips are similarly versatile.
20:05 ā AWS Custom Silicon (Nitro, Graviton, Inferentia, Tranium)
MB: Custom silicon is a differentiator. What sets yours apart vs. Google TPUs?
MG: We start with customers: breadth of choiceācapabilities, cost points, trade-offs. No single best solution for all workloads.
Timeline:
Nitro (ā10 years ago): offloaded virtualization (network, storage, hypervisor) to custom cards; customers got bare-metal performance and better security.
Graviton (ARM CPUs): enterprise-ready with Graviton2; today Graviton4 is ~20% faster than the best x86 and 20% cheaperāa strong value. We still sell tons of Intel and AMD, because customer choice.
AI accelerators (~5 years ago):
Inferentia for inferenceāAlexa cut inference costs ~70%.
Tranium1 taught us software ecosystem lessons.
Tranium2 is now in market.
Many customers still prefer NVIDIA (CUDA is excellent). Othersāe.g., Anthropicālean into Tranium; we also use Tranium under the hood to power Bedrock models in serverless inference.
MB: Would you sell your chips to third parties (like rumors of Google selling TPUs)?
MG: Never say never. Today, selling only in AWS simplifies everything: one environment (AWS data center, server, network), fewer SKUs/firmware paths. Selling merchant silicon adds complexityābut itās conceivable someday.
MB: Annapurna seems underappreciated. What did you see back then (pre-genAI)?
MG: They were mission-driven and matched Amazonās culture. We wanted to offload virtualization to a card; nobody had that product. Annapurna was building a network card with ARM cores. We co-designed to also offload EBS virtualization. They were smart, scrappy, customer-focused, thought big. We acquired themābest acquisition weāve made. Most of that team is still at Amazon 10 years later.
27:57 ā Which Models AWS Serves
MB: When deciding which models to serve on Bedrock, how do you choose?
MG: We want choice: the best models customers might want. Thereāll be a handful of frontier models (expensive), but lots of purpose-built ones. Recently we added Writer (agentic workflows) and 12 Labs (video understanding). We have Poolside (coding), Stability (image/video), Luma AI (video), and more. Ideally weād support everyoneālike AWS Marketplace. Some remain proprietary elsewhere, but our goal is breadth.
29:52 ā One Omni-Model vs. Many Specialized Models
MB: Some predict one giant omni-model. I think weāre headed to specialization. Your view?
MG: From day one we believed customers will use many models. Today they do: a large general model for planning/reasoning; specialty models for specific workflows; fine-tuned models on proprietary data. Trade-offs on cost/capability lead to mixture-of-experts in practice.
Enterprises want models that deeply understand their dataābetter fine-tuning, domain knowledge, customer context. And as we move into an agent-focused world, the model is critical but not sufficient. You need scaffolding, workflows, memory, audit logs, and domain-specific pieces. Most ROI will come from agent workflows doing real work.
33:39 ā Open vs. Closed Source
MB: You serve Anthropic (closed) and plenty of open-weight models. How do you think about open vs. closedāand partnerships with Anthropic, OpenAI, Meta?
MG: Most so-called āopenā are open weights, not OSS. The key for customers is customizationābring your own data, tailor the model, run custom workflows. Whether via open weights (e.g., Llama, Mistral) or APIs for fine-tuning/distillation on āclosedā models (e.g., Nova), the value is the same: fit to your use case. Over time, everyone will want customization; different vendors will enable it differently.
MB: Will AWS build a truly frontier model?
MG: We think choice matters. Weāre investing in Nova to offer differentiated capabilities while maintaining deep partnerships (Anthropic, etc.). Compete where it helps customers; partner everywhere elseāan AWS muscle weāve built over 19 years.
MB: Pricingāwill models race to cost of silicon + electricity?
MG: Unlikely. Models arenāt commodities today; why tomorrow? Cloud wasnāt commoditized after a year or two either. There are differences in availability, features, UX. Ask customers: Llama, Claude, GPTāare they commodities? No. Even open models differ (Mistral vs. Llama). Vendors must keep innovating; thereās real value to be captured.
MB: Open models seem ~3ā6 months behind closed source. Will that persist?
MG: Thereās no inherent advantage to open vs. closed; itās a choice. China labs open-sourced their best while behind; thatās why some feel āopen lags.ā But models like DeepSeek and Qwen are impressive; customers love them. OpenAI released open models (smaller than GPT-5). Anthropic hasnāt open-sourced. Meta open-sources all Llama models. Whether they remain behind is about execution, not openness.
41:33 ā Benchmarks
MB: Benchmarks are saturatedāAM24, A2, MMLU. Weāre down to single-digit gains. Do we need new benchmarks?
MG: Benchmarks are great for commodities (e.g., SSD throughput). The more complex the system, the worse benchmarks become. Early database benchmarks (TPC-C) faded; people test their own workloads. Weāll likely move the same way with models. Itās easy to train models to ace benchmarks; that doesnāt make them best overall.
MB: Your personal test?
MG: I prefer testing inside applicationsāhow well it synthesizes research into coherent docs, generates ideas, interacts. Speed matters too. Integration/UX matters a lotāPerplexity is a good example with thoughtful UI and visible āthinking.ā Latency is critical for real-time consumer use; for enterprise workflows (e.g., payroll agents), accuracy can trump speed and async is fine. We launched auditable reasoning (math-proof-like verification) applied to LLMsāpowerful when you need correctness over latency.
47:13 ā Agents
MB: Where are you seeing repeatable, accurate agent implementations with positive ROI?
MG: Several:
Agentic coding is a major unlockādevelopers build more, faster.
Enterprise agents were hard to build/run at scale. We launched Agent Core in Bedrock: building blocks for scalable, secure, auditable, measurable agents.
Secure, serverless runtime that scales to zero and up to thousands.
Short- and long-term memory built in.
Agent Gateway for auth with other agents/systems; hosted MCP server support.
Observability hooks to AWS or third-party tools.
Open framework: works with any model (Gemini, OpenAI, Bedrock), any stack (Strands, LangChain, etc.).
With scaffolding easier, we see use cases across processing, individual productivity, marketing, sales, and industry workflows. Many are still human-in-the-loop, but the path to more autonomy is clear.
50:31 ā Final Advice
MB: Many worry about being automated away. Words of encouragement?
MG: Make yourself valuable. AI and agents amplify employees; theyāre valuable through you. If youāre great at marketing, do marketingānot pulling campaign plumbing. If youāre great at building apps, be great at building, not at memorizing a language. Learn the tools, focus on customer problems, keep learning. I have very little worry that jobs just disappear and robots run everything.
MB: Thank you, Matt.
MG: Thank youāthis was really fun
Enjoyed this conversation?
Follow us on X and subscribe to our YouTube channel for more interviews with the people building the future of AI.

