logo
episode-header-image
Apr 2025
43m 38s

MLA 024 Code AI MCP Servers, ML Engineer...

OCDevel
About this episode

Tool use in code AI agents allows for both in-editor code completion and agent-driven file and command actions, while the Model Context Protocol (MCP) standardizes how these agents communicate with external and internal tools. MCP integration broadens the automation capabilities for developers and machine learning engineers by enabling access to a wide variety of local and cloud-based tools directly within their coding environments.

Links

Tool Use in Code AI Agents

  • Code AI agents offer two primary modes of interaction: in-line code completion within the editor and agent interaction through sidebar prompts.
  • Inline code completion has evolved from single-line suggestions to cross-file edits, refactoring, and modification of existing code blocks.
  • Tools accessible via agents include read, write, and list file functions, as well as browser automation and command execution; permissions for sensitive actions can be set by developers.
  • Agents can intelligently search a project's codebase and dependencies using search commands and regular expressions to locate relevant files.

Model Context Protocol (MCP)

  • MCP, introduced by Anthropic, establishes a standardized protocol for agents to communicate with tools and services, replacing bespoke tool integrations.
  • The protocol is analogous to REST for web servers and unifies tool calling for both local and cloud-hosted automation.
  • MCP architecture involves three components: the AI agent, MCP client, and MCP server. The agent provides context, the client translates requests and responses, and the server executes and responds with data in a structured format.
  • MCP servers can be local (STDIO-based for local tasks like file search or browser actions) or cloud-based (SSE for hosted APIs and SaaS tools).
  • Developers can connect code AI agents to directories of MCP servers, accessing an expanding ecosystem of automation tools for both programming and non-programming tasks.

MCP Application Examples

  • Local MCP servers include Playwright for browser automation and Postgres MCP for live database schema analysis and data-driven UI suggestions.
  • Cloud-based MCP servers integrate APIs such as AWS, enabling infrastructure management directly from coding environments.
  • MCP servers are not limited to code automation; they are widely used for pipeline automation in sales, marketing, and other internet-connected workflows.

Retrieval Augmented Generation (RAG) as an MCP Use Case

  • RAG, once standard in code AI tools, indexed codebases using embeddings to assist with relevant file retrieval, but many agents now favor literal search for practicality.
  • Local RAG MCP servers, such as Chroma or LlamaIndex, can index entire documentation sets to update agent knowledge of recent or project-specific libraries outside of widely-known frameworks.
  • Fine-tuning a local LLM with the same documentation is an alternative approach to integrating new knowledge into code AI workflows.

Machine Learning Applications

  • Code AI tooling supports feature engineering, data cleansing, pipeline setup, model design, and hyperparameter optimization, based on real dataset distributions and project specifications.
  • Agents can recommend advanced data transformations—such as Yeo-Johnson power transformation for skewed features—by directly analyzing example dataset distributions.
  • Infrastructure-as-code integration enables rapid deployment of machine learning models and supporting components by chaining coding agents to cloud automation tools.
  • Automation concepts from code AI apply to both traditional code file workflows and Jupyter Notebooks, though integration with notebooks remains less seamless.
  • An iterative approach using sidecar Python files combined with custom instructions helps agents access necessary background and context for ML projects.

Workflow Strategies for Machine Learning Engineers

  • To leverage code AI agents in machine learning tasks, engineers can provide data samples and visualizations to agents through Python files or prompt contexts.
  • Agents can guide creation and comparison of multiple model architectures, metrics, and loss functions, improving efficiency and broadening solution exploration.
  • While Jupyter Lab plugin integration is currently limited, some success can be achieved by working with notebook files via code AI tools in standard code editors or by moving between notebooks and Python files for maximum flexibility.

 

Up next
May 2025
MLG 034 Large Language Models 1
Explains language models (LLMs) advancements. Scaling laws - the relationships among model size, data size, and compute - and how emergent abilities such as in-context learning, multi-step reasoning, and instruction following arise once certain scaling thresholds are crossed. The ... Show More
50m 48s
May 2025
MLG 035 Large Language Models 2
At inference, large language models use in-context learning with zero-, one-, or few-shot examples to perform new tasks without weight updates, and can be grounded with Retrieval Augmented Generation (RAG) by embedding documents into vector databases for real-time factual lookup ... Show More
45m 25s
May 2025
MLG 036 Autoencoders
Auto encoders are neural networks that compress data into a smaller "code," enabling dimensionality reduction, data cleaning, and lossy compression by reconstructing original inputs from this code. Advanced auto encoder types, such as denoising, sparse, and variational auto encod ... Show More
1h 5m
Recommended Episodes
Jan 2025
Erik Bernhardsson on Creating Tools That Make AI Feel Effortless
Today on No Priors, Elad chats with Erik Bernhardsson, founder and CEO of Modal Labs, a platform simplifying ML workflows by providing a serverless infrastructure designed to streamline deployment, scaling, and development for AI engineers. Erik talks about his early work on Spot ... Show More
23m 36s
Jul 2025
Best IDEs and AI Coding Assistants in 2025: Complete Guide to VSCode, Cursor, GitHub Copilot & More
Are you ready for the future of software development? The world of IDEs and AI coding assistants is changing daily, and staying ahead means understanding the tools that are reshaping how we code. In this comprehensive 2025 guide, we dive deep into the most influential development ... Show More
9m 50s
Sep 18
From RAG to Relational: How Agentic Patterns Are Reshaping Data Architecture
SummaryIn this episode of the AI Engineering Podcast Mark Brooker, VP and Distinguished Engineer at AWS, talks about how agentic workflows are transforming database usage and infrastructure design. He discusses the evolving role of data in AI systems, from traditional models to m ... Show More
52m 58s
Jul 2024
#225 The Full Stack Data Scientist with Savin Goyal, Co-Founder & CTO at Outerbounds
The role of the data scientist is changing. Some organizations are splitting the role into more narrowly focused jobs, while others are broadening it. The latter approach, known as the Full Stack Data Scientist, is derived from the concept of a full stack software engineer, with ... Show More
48m 44s
Oct 11
Context Engineering as a Discipline: Building Governed AI Analytics
SummaryIn this episode of the Data Engineering Podcast, host Tobias Macey welcomes back Nick Schrock, CTO and founder of Dagster Labs, to discuss Compass - a Slack-native, agentic analytics system designed to keep data teams connected with business stakeholders. Nick shares his j ... Show More
51m 58s
May 2025
How OpenAI Builds AI Agents That Think and Act with Josh Tobin - #730
Today, we're joined by Josh Tobin, member of technical staff at OpenAI, to discuss the company’s approach to building AI agents. We cover OpenAI's three agentic offerings—Deep Research for comprehensive web research, Operator for website navigation, and Codex CLI for local code e ... Show More
1h 7m
Feb 2025
HN769: CI/CD Pipelines and Network Automation
Continuous Integration / Continuous Deployment (CI/CD) is a framework that developers use to help them manage and integrate frequent code changes. As network automation evolves, should network engineers adopt CI/CD? Guest Tony Bourke joins us to talk about CI/CD pipelines: what t ... Show More
57m 40s
Feb 2025
The Future of Data Engineering: AI, LLMs, and Automation
Summary In this episode of the Data Engineering Podcast Gleb Mezhanskiy, CEO and co-founder of DataFold, talks about the intersection of AI and data engineering. He discusses the challenges and opportunities of integrating AI into data engineering, particularly using large langua ... Show More
59m 39s
Aug 2023
AI Superpowers for Frontend Developers, with Vercel Founder/CEO Guillermo Rauch
Everything digital is increasingly intermediated through web user experiences, and now AI development can be frontend-first, too. Just ask Guillermo Rauch, the founder and CEO of Vercel, the company behind Next.js. In this episode of No Priors, hosts Sarah Guo and Elad Gil speak ... Show More
38m 13s
Sep 2024
AI Agents for Data Analysis with Shreya Shankar - #703
Today, we're joined by Shreya Shankar, a PhD student at UC Berkeley to discuss DocETL, a declarative system for building and optimizing LLM-powered data processing pipelines for large-scale and complex document analysis tasks. We explore how DocETL's optimizer architecture works, ... Show More
48m 24s