This show is your guidebook to building scalable and maintainable AI systems. You will learn how to architect AI applications, apply AI to your work, and the considerations involved in building or customizing new models. Everything that you need to know to deliver real impact and value with machine learning and artificial intelligence.
Support the show!Listen in your favorite app:
FountainHere are shows you might like
Summary In this episode of the AI Engineering Podcast Emmanouil (Manos) Koukoumidis, CEO of Oumi, about his vision for an open platform for building, evaluating, and deploying AI foundation models. Manos shares his journey from working on natural language AI services at Google Cloud to founding Oumi with a mission to advance open-source AI,…
Summary In this episode of the AI Engineering Podcast Emmanouil (Manos) Koukoumidis, CEO of Oumi,…
16 March 2025 | 00:56:12
Summary In this episode of the AI Engineering Podcast Adil Hafiz talks about the Arch project, a gateway designed to simplify the integration of AI agents into business systems. He discusses how the gateway uses Rust and Envoy to provide a unified interface for handling prompts and integrating large language models (LLMs), allowing developers to…
Summary In this episode of the AI Engineering Podcast Adil Hafiz talks about the Arch project, a…
26 February 2025 | 00:31:25
Summary In this episode of the AI Engineering Podcast Ali Golshan, co-founder and CEO of Gretel.ai, talks about the transformative role of synthetic data in AI systems. Ali explains how synthetic data can be purpose-built for AI use cases, emphasizing privacy, quality, and structural stability. He highlights the shift from traditional methods to…
Summary In this episode of the AI Engineering Podcast Ali Golshan, co-founder and CEO of Gretel.ai,…
16 February 2025 | 00:54:21
Summary In this episode of the AI Engineering podcast Viraj Mehta, CTO and co-founder of TensorZero, talks about the use of LLM gateways for managing interactions between client-side applications and various AI models. He highlights the benefits of using such a gateway, including standardized communication, credential management, and potential…
Summary In this episode of the AI Engineering podcast Viraj Mehta, CTO and co-founder of TensorZero,…
22 January 2025 | 01:03:05
Summary In this episode of the AI Engineering Podcast Ron Green, co-founder and CTO of KungFu AI, talks about the evolving landscape of AI systems and the challenges of harnessing generative AI engines. Ron shares his insights on the limitations of large language models (LLMs) as standalone solutions and emphasizes the need for human oversight,…
Summary In this episode of the AI Engineering Podcast Ron Green, co-founder and CTO of KungFu AI,…
16 December 2024 | 00:55:13
Summary In this episode of the AI Engineering Podcast Jim Olsen, CTO of ModelOp, talks about the governance of generative AI models and applications. Jim shares his extensive experience in software engineering and machine learning, highlighting the importance of governance in high-risk applications like healthcare. He explains that governance is…
Summary In this episode of the AI Engineering Podcast Jim Olsen, CTO of ModelOp, talks about the…
01 December 2024 | 00:54:19
Summary In this episode of the AI Engineering Podcast, Vasilije Markovich talks about enhancing Large Language Models (LLMs) with memory to improve their accuracy. He discusses the concept of memory in LLMs, which involves managing context windows to enhance reasoning without the high costs of traditional training methods. He explains the…
Summary In this episode of the AI Engineering Podcast, Vasilije Markovich talks about enhancing…
25 November 2024 | 00:55:01
Summary In this episode of the AI Engineering Podcast, Tanner Burson, VP of Engineering at Prismatic, talks about the evolving impact of generative AI on software developers. Tanner shares his insights from engineering leadership and data engineering initiatives, discussing how AI is blurring the lines of developer roles and the strategic value of…
Summary In this episode of the AI Engineering Podcast, Tanner Burson, VP of Engineering at…
22 November 2024 | 00:52:58
Summary Machine learning workflows have long been complex and difficult to operationalize. They are often characterized by a period of research, resulting in an artifact that gets passed to another engineer or team to prepare for running in production. The MLOps category of tools have tried to build a new set of utilities to reduce that friction,…
Summary Machine learning workflows have long been complex and difficult to operationalize. They are…
11 November 2024 | 01:16:12
Summary With the growth of vector data as a core element of any AI application comes the need to keep those vectors up to date. When you go beyond prototypes and into production you will need a way to continue experimenting with new embedding models, chunking strategies, etc. You will also need a way to keep the embeddings up to date as your data…
Summary With the growth of vector data as a core element of any AI application comes the need to…
11 November 2024 | 00:53:50