logo
episode-header-image
Apr 13
43m 38s

MLA 024 Code AI MCP Servers, ML Engineer...

OCDevel
About this episode

Tool use in code AI agents allows for both in-editor code completion and agent-driven file and command actions, while the Model Context Protocol (MCP) standardizes how these agents communicate with external and internal tools. MCP integration broadens the automation capabilities for developers and machine learning engineers by enabling access to a wide variety of local and cloud-based tools directly within their coding environments.

Links

Tool Use in Code AI Agents

  • Code AI agents offer two primary modes of interaction: in-line code completion within the editor and agent interaction through sidebar prompts.
  • Inline code completion has evolved from single-line suggestions to cross-file edits, refactoring, and modification of existing code blocks.
  • Tools accessible via agents include read, write, and list file functions, as well as browser automation and command execution; permissions for sensitive actions can be set by developers.
  • Agents can intelligently search a project’s codebase and dependencies using search commands and regular expressions to locate relevant files.

Model Context Protocol (MCP)

  • MCP, introduced by Anthropic, establishes a standardized protocol for agents to communicate with tools and services, replacing bespoke tool integrations.
  • The protocol is analogous to REST for web servers and unifies tool calling for both local and cloud-hosted automation.
  • MCP architecture involves three components: the AI agent, MCP client, and MCP server. The agent provides context, the client translates requests and responses, and the server executes and responds with data in a structured format.
  • MCP servers can be local (STDIO-based for local tasks like file search or browser actions) or cloud-based (SSE for hosted APIs and SaaS tools).
  • Developers can connect code AI agents to directories of MCP servers, accessing an expanding ecosystem of automation tools for both programming and non-programming tasks.

MCP Application Examples

  • Local MCP servers include Playwright for browser automation and Postgres MCP for live database schema analysis and data-driven UI suggestions.
  • Cloud-based MCP servers integrate APIs such as AWS, enabling infrastructure management directly from coding environments.
  • MCP servers are not limited to code automation; they are widely used for pipeline automation in sales, marketing, and other internet-connected workflows.

Retrieval Augmented Generation (RAG) as an MCP Use Case

  • RAG, once standard in code AI tools, indexed codebases using embeddings to assist with relevant file retrieval, but many agents now favor literal search for practicality.
  • Local RAG MCP servers, such as Chroma or LlamaIndex, can index entire documentation sets to update agent knowledge of recent or project-specific libraries outside of widely-known frameworks.
  • Fine-tuning a local LLM with the same documentation is an alternative approach to integrating new knowledge into code AI workflows.

Machine Learning Applications

  • Code AI tooling supports feature engineering, data cleansing, pipeline setup, model design, and hyperparameter optimization, based on real dataset distributions and project specifications.
  • Agents can recommend advanced data transformations—such as Yeo-Johnson power transformation for skewed features—by directly analyzing example dataset distributions.
  • Infrastructure-as-code integration enables rapid deployment of machine learning models and supporting components by chaining coding agents to cloud automation tools.
  • Automation concepts from code AI apply to both traditional code file workflows and Jupyter Notebooks, though integration with notebooks remains less seamless.
  • An iterative approach using sidecar Python files combined with custom instructions helps agents access necessary background and context for ML projects.

Workflow Strategies for Machine Learning Engineers

  • To leverage code AI agents in machine learning tasks, engineers can provide data samples and visualizations to agents through Python files or prompt contexts.
  • Agents can guide creation and comparison of multiple model architectures, metrics, and loss functions, improving efficiency and broadening solution exploration.
  • While Jupyter Lab plugin integration is currently limited, some success can be achieved by working with notebook files via code AI tools in standard code editors or by moving between notebooks and Python files for maximum flexibility.

 

Up next
Yesterday
MLA 025 AI Image Generation: Midjourney vs Stable Diffusion, GPT-4o, Imagen & Firefly
The 2025 generative AI image market is a trade-off between aesthetic quality, instruction-following, and user control. This episode analyzes the key platforms, comparing Midjourney's artistic output against the superior text generation and prompt adherence of GPT-4o and Imagen 4, ... Show More
1h 12m
May 30
MLG 036 Autoencoders
Auto encoders are neural networks that compress data into a smaller "code," enabling dimensionality reduction, data cleaning, and lossy compression by reconstructing original inputs from this code. Advanced auto encoder types, such as denoising, sparse, and variational auto encod ... Show More
1h 5m
May 8
MLG 035 Large Language Models 2
At inference, large language models use in-context learning with zero-, one-, or few-shot examples to perform new tasks without weight updates, and can be grounded with Retrieval Augmented Generation (RAG) by embedding documents into vector databases for real-time factual lookup ... Show More
45m 25s
Recommended Episodes
Jan 2025
Erik Bernhardsson on Creating Tools That Make AI Feel Effortless
Today on No Priors, Elad chats with Erik Bernhardsson, founder and CEO of Modal Labs, a platform simplifying ML workflows by providing a serverless infrastructure designed to streamline deployment, scaling, and development for AI engineers. Erik talks about his early work on Spot ... Show More
23m 36s
Jul 2024
#225 The Full Stack Data Scientist with Savin Goyal, Co-Founder & CTO at Outerbounds
The role of the data scientist is changing. Some organizations are splitting the role into more narrowly focused jobs, while others are broadening it. The latter approach, known as the Full Stack Data Scientist, is derived from the concept of a full stack software engineer, with ... Show More
48m 44s
May 6
How OpenAI Builds AI Agents That Think and Act with Josh Tobin - #730
Today, we're joined by Josh Tobin, member of technical staff at OpenAI, to discuss the company’s approach to building AI agents. We cover OpenAI's three agentic offerings—Deep Research for comprehensive web research, Operator for website navigation, and Codex CLI for local code e ... Show More
1h 7m
Feb 2025
HN769: CI/CD Pipelines and Network Automation
Continuous Integration / Continuous Deployment (CI/CD) is a framework that developers use to help them manage and integrate frequent code changes. As network automation evolves, should network engineers adopt CI/CD? Guest Tony Bourke joins us to talk about CI/CD pipelines: what t ... Show More
57m 40s
Feb 2025
The Future of Data Engineering: AI, LLMs, and Automation
Summary In this episode of the Data Engineering Podcast Gleb Mezhanskiy, CEO and co-founder of DataFold, talks about the intersection of AI and data engineering. He discusses the challenges and opportunities of integrating AI into data engineering, particularly using large langua ... Show More
59m 39s
Aug 2023
AI Superpowers for Frontend Developers, with Vercel Founder/CEO Guillermo Rauch
Everything digital is increasingly intermediated through web user experiences, and now AI development can be frontend-first, too. Just ask Guillermo Rauch, the founder and CEO of Vercel, the company behind Next.js. In this episode of No Priors, hosts Sarah Guo and Elad Gil speak ... Show More
38m 13s
May 2024
MLOps + DevOps + Kubernetes with Annie Talvasto
Machine learning models need updating - what's the reliable way to do it? While in Romania, Richard sat down with Annie Talvasto to talk about her work helping to build DevOps practices around machine learning: Building repeatable processes for data ingestions, cleaning, organiza ... Show More
33m 17s
Aug 2024
Episode 201 - Introduction to KitOps for MLOps
Join Allen and Mark in this episode of Two Voice Devs as they dive into the world of MLOps and explore KitOps, an open-source tool for packaging and versioning machine learning models and related artifacts. Learn how KitOps leverages the Open Container Initiative (OCI) standard t ... Show More
33m 59s
Nov 2024
SE Radio 641: Catherine Nelson on Machine Learning in Data Science
Catherine Nelson, author of the new O’Reilly book, Software Engineering for Data Scientists, discusses the collaboration between data scientists and software engineers -- an increasingly common pairing on machine learning and AI projects. Host Philip Winston speaks with Nelson ab ... Show More
48m 19s
Nov 2024
Automate IAM policies creation across multiple accounts
In this episode of the AWS Developers Podcast, Seb and Ran Isenberg discuss the automation of IAM policies across multiple AWS accounts. They explore the challenges faced in managing security and access in a multi-account environment, the design and implementation of an automatio ... Show More
34m 20s