Header image for AI Engineer Summit 2025: Agents at Work - Engineering Track

AI Engineer Summit 2025: Agents at Work - Engineering Track

AI Engineer Summit 2025: Agents at Work - Engineering Track

Executive Summary:

The 2025 AI Engineer Summit marked a pivotal moment in the evolution of AI agents, with a clear shift from theoretical capabilities to practical implementation. While speakers celebrated unprecedented advances in reasoning models and tool use, a consistent narrative emerged across presentations: developing reliable, production-ready AI agents requires not just advanced models but also robust engineering practices, thoughtful evaluation frameworks, and organizational alignment. The most successful agent implementations focused on specific domains where they could deliver measurable value, with speakers highlighting coding assistance, customer service, research, and finance as areas with proven product-market fit. A growing consensus emerged that successful AI implementation requires balancing high-capability models with pragmatic constraints, as organizations grapple with reliability challenges, cost considerations, and human-AI collaboration dynamics. Above all, the summit demonstrated that AI engineering is evolving from simply utilizing models to building systems that scale intelligently and improve with minimal human intervention.

Key Takeaways:

  • 2025 is poised to be "the year of agents" with capability improvements in reasoning models, lower compute costs, and better tool use capabilities
  • Successful agent implementation requires focusing on specific high-value use cases rather than attempting to build general-purpose solutions
  • Evaluating agents is fundamentally different from evaluating models, requiring multi-dimensional metrics that consider cost, latency, and reliability
  • The gap between model capabilities in controlled settings and their performance in complex real-world environments remains substantial
  • Organizations are shifting from rigid, deterministic systems to more flexible architectures that can improve automatically as models get smarter
  • Effective agents must understand both what users are doing in the moment and develop memory about user preferences and organizational guidelines
  • Reinforcement learning is emerging as a key approach for improving agent reliability, especially for complex multi-step tasks
  • Voice AI and multimodal agents present unique challenges but open new possibilities for more natural human-AI collaboration
  • Domain-specific models trained on proprietary data consistently outperform general-purpose approaches for enterprise use cases
  • Education and AI literacy will be critical as AI agents become more prevalent, with growing efforts to teach AI concepts from childhood

Speaker Landscape:

The conference featured an exceptionally diverse set of perspectives spanning the complete AI agent implementation ecosystem. Leaders from established organizations included Anju Kambadur (Bloomberg), John Crepezzi (Jane Street), Mukund Sridhar and Aarush Selvan (Google Gemini), Soumith Chintala (Meta PyTorch), and Rahul Sengottuvelu (Ramp). Specialized AI companies were represented by Barry Zhang (Anthropic), Karina Nguyen (OpenAI), Mike Conover (BrightWave), and Zack Reneau-Wedeen (Sierra). Academic perspectives came from Sayash Kapoor (author of AI Snake Oil) and Will Brown (Morgan Stanley researcher). The summit also included founders of emerging AI tools including Kevin Hou (Windsurf), Kyle Corbitt (OpenPipe), and Mustafa Ali (Method Financial). This blend of enterprise practitioners, specialized AI companies, researchers, and startup founders created a comprehensive view of how agents are being implemented across different domains and organizational contexts.

Thematic Analysis:

From Capability to Reliability: The Path to Production-Ready Agents

The conference opened with Swyx (AI Engineer Foundation) framing 2025 as "the perfect storm for AI agents" with converging advancements in reasoning, compute, engineering optimizations, infrastructure, and hardware. Yet a tension emerged between theoretical capabilities and practical implementation challenges. Sayash Kapoor (AI Snake Oil) highlighted how many headline-grabbing agent capabilities fail to translate into reliable performance, with examples ranging from legal research tools with high hallucination rates to supposed AI scientists that failed basic reproducibility tests.

This reliability gap was echoed by several enterprise speakers. Anju Kambadur (Bloomberg) emphasized that in finance, "precision, comprehensiveness, speed, throughput, and availability" are non-negotiable requirements that current agent technologies don't consistently deliver. John Crepezzi (Jane Street) described how they built custom models for working with OCaml code because off-the-shelf models were insufficient. As Mike Conover (BrightWave) put it, "winning systems will perform end-to-end RL over tool use calls" to optimize for global outcomes rather than local decisions.

The path forward emphasized systematic evaluation and incremental improvement. Kyle Corbitt (OpenPipe) demonstrated how Method Financial scaled to 500 million agent deployments by carefully measuring error rates, latency, and cost across different models before fine-tuning a small, reliable model optimized for their specific use case. Barry Zhang (Anthropic) advised a simple architecture focused on three components—environment, tools, and system prompt—that could be iteratively improved rather than trying to build complex systems from the start.

Agents as Augmentation: New Paradigms for Human-AI Collaboration

A recurring theme across presentations was reconceptualizing agents not as autonomous entities but as collaborative extensions of human capabilities. Karina Nguyen (OpenAI) outlined an evolution from "models trained on RL and chain of thought using real-world tools" toward "co-innovators" that collaborate with humans on creative tasks. She highlighted how Canvas was designed with collaborative affordances like multiplayer editing and contextual search that enable humans and AI to work together effectively.

Kevin Hou (Windsurf) demonstrated how their editor tracks a "unified timeline" of both human and AI actions, allowing their agent to understand what developers are doing and continue work seamlessly. Zack Reneau-Wedeen (Sierra) described how their voice agents for customer service operate within a semi-autonomous framework where bots handle most calls but can escalate to humans when needed. This creates a virtuous cycle where the system "gets even better at that sort of call" over time.

Speakers also emphasized that empowering humans remains the ultimate goal. Rahul Sengottuvelu (Ramp) argued that "what is truly scarce in the world is engineer time," making it worthwhile to run models 10,000 times more compute-intensive if they save human effort. Soumith Chintala (Meta PyTorch) advocated for personal, local, private AI agents that can deeply augment humans without privacy risks, stressing that "your personal agent is so personal to you and so intimate" that users should maintain control rather than delegating to cloud services.

Engineering for Intelligence: Building Systems That Scale with Smarter Models

A forward-looking theme emerged around designing systems that automatically improve as underlying models get better. Rahul Sengottuvelu (Ramp) presented a framework distinguishing "classical compute" from "fuzzy compute" (neural networks), arguing that systems should maximize the latter since "if you did nothing, absolutely nothing... the big labs are still working, spending billions of dollars making those models better." He demonstrated an experimental email client where the LLM itself acted as the backend, rendering UI and handling user interactions without traditional software engineering.

Will Brown (Morgan Stanley) discussed how reinforcement learning is becoming crucial for agent development, describing how models like DeepSeek's R1 demonstrate that "the long chain of thought... actually emerges as a byproduct" of training with appropriate reinforcement signals. He presented his GRPO implementation that helps developers fine-tune task-specific agents with custom reward functions.

John Crepezzi (Jane Street) described their "Aid" framework which provides a unified backend for multiple editor integrations, allowing them to swap in new models or context-building strategies without changing frontends. Similarly, Kevin Hou (Windsurf) explained how their agent architecture was designed to "scale with intelligence" so that "if the models get better, our product gets better," including removing chat interfaces in favor of pure agent interactions.

Evaluating Agents: Beyond Benchmark Performance to Real-World Value

Multiple speakers addressed the challenge of evaluating agent performance in ways that translate to actual business value. Sayash Kapoor (AI Snake Oil) criticized over-reliance on static benchmarks, noting that Cognition's Devin agent "did very well on SWE Bench" but was "only successful at three out of twenty tasks" in real-world testing. He advocated for multi-dimensional metrics that consider not just accuracy but also cost and reliability, showing how these metrics revealed that Claude 3.5 performed as well as GPT-4o on some tasks at 1/10th the cost.

Anju Kambadur (Bloomberg) highlighted organizational approaches to evaluation, explaining that Bloomberg's agents undergo rigorous testing with "remediation workflows and circuit breakers" to catch errors before they impact financial data. Mike Conover (BrightWave) similarly described how their research agent includes detailed citation tracking and "receipts" so users can validate information sources.

Kyle Corbitt (OpenPipe) provided a practical framework for evaluation, recommending targeted testing of error rates, latency, and cost for specific production tasks rather than relying on generic benchmarks. This approach helped Method Financial identify that fine-tuned 8B parameter models outperformed much larger models for their specific use case while dramatically reducing costs and latency.

Voice and Multimodal Agents: The Next Frontier

Voice AI emerged as an important frontier for agent development. Nick Karyotakis (SuperDial) outlined the challenges of building reliable voice agents, noting that "voice UI itself is not that new" but modern approaches have shifted from "prescriptive to descriptive development." He described how SuperDial builds phone agents that handle insurance verification calls by traversing phone trees, extracting information, and escalating to humans when needed.

Zack Reneau-Wedeen (Sierra) shared how Sierra's voice assistants handle customer service for brands like Chubbies, SiriusXM, and ADT by creating a "responsive design" approach where "it's the same agent code" operating across different channels and modalities. He emphasized that for voice agents, "latency all of a sudden matters so much more" than with text interfaces.

Mukund Sridhar and Aarush Selvan (Google) presented Gemini Deep Research as a multimodal research agent that "can browse the web as much as it needs" to produce comprehensive answers. They highlighted engineering challenges including long-running tasks, iterative planning, noisy web environments, and context management. Karina Nguyen (OpenAI) similarly discussed how multimodal capabilities in Canvas enable more natural collaboration between humans and AI on creative tasks.

Education and Access: Building the Next Generation of AI Engineers

The conference closed with perspectives on democratizing AI engineering. Stefania Druga presented Cognimates, a platform that teaches children about AI by allowing them to train their own models and build applications. She emphasized that "kids are actually like little scientists" who can formulate and test hypotheses about how AI works, and that early exposure helps "demystify the intelligence" of AI systems. She highlighted that AI literacy is now part of the EU AI Act, requiring providers to "ensure a sufficient level of AI literacy" among users.

Soumith Chintala (Meta PyTorch) advocated for open models, noting that while each major lab only improves their own model, "open models are improving themselves in coordination across board." He predicted that "open models would actually start getting better than closed models per dollar of investment" as the open-source community reaches critical mass.

Multiple speakers touched on making agent development more accessible. Will Brown shared a simple reinforcement learning implementation that went viral because "it was one file of code... really simple... and it invited modification." Kyle Corbitt (OpenPipe) emphasized that fine-tuning is becoming simpler, making it viable for smaller teams to build custom models optimized for their specific use cases.

Notable Quotations:

On the State of AI Agents:

"Gartner hates us. Gartner thinks we've hit the peak. So it's only downhill from here, guys. Sorry to inform you that AI engineering is over." — Swyx, AI Engineer Foundation (joking about the Gartner hype cycle)

"I feel compelled to say something about the death of the traditional software engineering job, so I'll leave you with these last few words. If you're in SWE, pivot to AIE." — Mustafa Ali, Method Financial

"The blank page when you start to write. The effect in Scratch it's called Cold Start. We see a lot of students that go to the platform and they really don't know where to start. So it's very helpful for that too." — Stefania Druga, Google

On Reliability Challenges:

"In reality, you're saying every hundred requests, 20 of them are just completely wrong." — Sayash Kapoor (quoting Waseem Alshikh at Writer)

"Language models are already capable of very many things. But if you trick yourself into believing this means a reliable experience for the end user, that's when products in the real world go wrong." — Sayash Kapoor, AI Snake Oil

"Building in some of this guardrail, I just think is good sense and that almost makes you go faster as you factor out individual agents and each agent can evolve without having these handshake signals." — Anju Kambadur, Bloomberg

On Practical Implementation:

"Don't build agents for everything. Keep it as simple for as long as possible. And finally, as you iterate, try to think like your agent, gain their perspective, and help them do their job." — Barry Zhang, Anthropic

"We want you to spend time on things that you are good at, right? The things that make us all excited, which is shipping products, building great features, and generally just shipping code." — Kevin Hou, Windsurf

"If your budget per task is around 10 cents, for example, you're building a high volume customer support system, that only affords you 30 to 50,000 tokens." — Barry Zhang, Anthropic

On Evaluation and Improvement:

"To become successful engineers, you need a reliability shift in your mindset. To think of yourselves as the people who are ensuring that this next wave of computing is as reliable for end users as possible." — Sayash Kapoor, AI Snake Oil

"If there are false positives in your verifiers, the model performance sort of bends downwards, simply because the more you try, the more likely it is you'll get a wrong answer." — Sayash Kapoor, AI Snake Oil

"Benchmark performance very rarely translates into the real world." — Sayash Kapoor, AI Snake Oil

On Future Directions:

"The way I'm thinking about it is the kind of interface to AGI is blank canvas that kind of self-morphs into your intent." — Karina Nguyen, OpenAI

"When I need a personal local agent, I definitely want voice mode because sometimes I want to talk to it and not actually type out everything I want to say." — Soumith Chintala, Meta PyTorch

"The bitter lesson is just so powerful, and exponential trends are so powerful that you can just hitch the ride." — Rahul Sengottuvelu, Ramp

Emerging Questions:

  • How do we design evaluation frameworks that measure agent performance in real-world conditions rather than controlled environments?
  • What is the optimal balance between model capabilities and engineering scaffolding in production agent systems?
  • How can we ensure AI agents remain trustworthy when given greater autonomy and access to sensitive information?
  • What new user interfaces and interaction paradigms will emerge to support effective human-AI collaboration?
  • How will the economics of agent development evolve as compute costs decrease but expectations for reliability increase?
  • Will domain-specific agents continue to dominate, or will general-purpose agents eventually become viable?
  • How should agent systems handle failure cases and know when to escalate to humans?
  • What governance models and safety frameworks need to be developed as agents gain more agency?
  • How will the rise of open models and improved fine-tuning tools change the competitive landscape?
  • What educational approaches are needed to prepare the next generation for an agent-driven world?

Resources:

Agent Development Frameworks:

  • PipeCat - Open source voice AI orchestration framework by Daily (mentioned by Nick Karyotakis)
  • OpenPipe - Platform for building, training, and deploying fine-tuned open source models (presented by Kyle Corbitt)
  • Tensor Zero - Structured and typed LLM endpoints for production (mentioned by Nick Karyotakis)
  • DSPy - Framework for optimizing prompts through automation (mentioned by Will Brown)
  • LangFuse - Self-hostable observability platform for LLM applications (mentioned by Nick Karyotakis)

Models and Research:

  • R1 - DeepSeek's open source reasoning model inspired by OpenAI's o1 (discussed by Will Brown)
  • LLAMA 3.1 - Meta's open source language model (mentioned by multiple speakers)
  • GPTO Algorithm - Simple RL algorithm for fine-tuning LLMs (presented by Will Brown)
  • Rubric Engineering - Will Brown's approach to designing reward functions for LLM fine-tuning
  • gr.inc - Ross Taylor's project addressing reasoning gaps in open models (mentioned by Soumith Chintala)

Enterprise Tools:

  • Aid - Jane Street's AI development environment providing unified backend for editor integrations
  • WindSurf - Agentic editor with background understanding of development context
  • BrightWave - Knowledge agent for due diligence and financial research
  • Sierra - Voice AI platform for customer service and conversation automation
  • Gemini Deep Research - Google's research agent for comprehensive web exploration

Educational Resources:

  • Cognimates - Platform for teaching children about AI through visual programming
  • AI Literacy - EU AI Act now requires AI literacy as part of regulation
  • Scratch - Visual programming environment for children with 100+ million users

Next Steps & Future Outlook:

Industry Direction:

  • The industry is moving from an experimentation phase to production implementation, with more focus on reliability engineering than model innovation
  • Voice and multimodal agents will become more prevalent as interaction paradigms evolve beyond text interfaces
  • Open source models will continue to close the gap with proprietary models, driven by community collaboration
  • Fine-tuning and reinforcement learning will become standard practices for optimizing agents for specific tasks
  • Local, private agents will emerge as alternatives to cloud-based services for applications requiring deeper personalization

Action Items:

  • For Technical Leaders: Implement multi-dimensional evaluation frameworks that consider cost, latency, and reliability; design systems that improve automatically as models get better
  • For Product Managers: Focus on specific high-value use cases rather than attempting to build general-purpose agents; create interfaces that enable effective human-AI collaboration
  • For Organizations: Develop appropriate governance frameworks for agents with increasing autonomy; invest in data infrastructure and educational initiatives
  • For AI Engineers: Learn reinforcement learning techniques for agent development; focus on building systems that scale with model improvements rather than requiring constant reengineering
  • For Educators: Integrate AI literacy into curricula from childhood; develop tools that help students understand AI concepts through hands-on experience

The 2025 AI Engineer Summit revealed an industry in transition from promising technology to practical implementation. As agents move from controlled environments to production systems, the focus is shifting from raw capabilities to reliability engineering, from benchmark performance to real-world value, and from general-purpose solutions to domain-specific applications. The most successful organizations are not just deploying the latest models but building systems that can continuously improve as AI capabilities advance. With continued progress in reinforcement learning, multimodal interfaces, and open source development, 2025 is poised to be the year when AI agents move from experimental projects to essential business tools.


Feature Image Prompt:

Generate an image. The aesthetic should be cyberpunk with colors of neon pink, blue and purple. Do not add any people. Imagine a futuristic, rain-soaked cityscape where towering, neon-drenched skyscrapers pulse with holographic control panels and streaming digital data. Interwoven circuit-like networks and abstract, glitch-inspired schematics hover over reflective surfaces, symbolizing the dynamic evolution of AI from abstract theory to reliable, production-ready systems. Vivid neon hues and bold, fragmented geometries blend to create an atmosphere of cutting-edge engineering and relentless innovation.