Apidog

All-in-one Collaborative API Development Platform

API Design

API Documentation

API Debugging

API Mocking

API Automated Testing

How to Use Kubernetes MCP Server (K8s MCP)

Simplify K8s management with AI! This tutorial guides you through setting up a Kubernetes MCP Server (K8s MCP) to automate tasks and gain insights. Start managing your cluster smarter today!

Ashley Goolam

Ashley Goolam

Updated on April 7, 2025

Have you ever found managing your Kubernetes (K8s) deployments to be a challenging task? With the help of AI, you can efficiently manage containers while maintaining control and security. This is where the Kubernetes MCP Server (K8s MCP) comes into play.

This tutorial will guide you through understanding and using a K8s MCP, which is a server implementing the Model Context Protocol (MCP) specifically designed for interacting with your Kubernetes clusters. This unlocks exciting possibilities for automating tasks, gaining insights, and even using AI agents to manage your deployments. Let's dive into the world of intelligent cloud management!

💡
Before you begin configuring the Kubernetes MCP Server, consider exploring Apidog—a powerful tool for designing, testing, and documenting APIs. Apidog simplifies API integration with structured models and seamless collaboration. When combined with Zapier MCP, it enhances automation and API management.
Apidog Ui image
button

What is a Kubernetes MCP Server (K8s MCP)?

A Kubernetes MCP Server is a bridge between your AI-powered tools and your Kubernetes cluster. It implements the Model Context Protocol (MCP) to provide a secure and structured way for AI to interact with your deployments.

Think of it this way: Kubernetes is the operating system for your cloud, and the K8s MCP is the translator that allows AI to understand and manage that operating system.

kubernetes official website

Instead of directly exposing your cluster to potentially untrusted AI models, the K8s MCP server acts as a gatekeeper, providing a controlled interface that allows AI to:

  • List available resources in your cluster (pods, deployments, services, etc.).
  • Get detailed information about specific resources.
  • Execute commands to manage your deployments (scale, update, restart, etc.).

All of this happens with proper authorization and auditing, ensuring that your cluster remains secure.

Kubernetes MCP Server Prerequisites

Before we get started, let's make sure you have everything you need:

1. A Kubernetes Cluster: You'll need access to a running Kubernetes cluster. This could be a local cluster (using Minikube or Rancher Desktop), a cloud-based cluster (on AWS, Azure, or Google Cloud), or any other Kubernetes environment.

2. kubectl Installed and Configured: The kubectl command-line tool is essential for interacting with your Kubernetes cluster.

  • Make sure you have kubectl installed and in your PATH (your operating system's list of directories where executable programs are searched for).
  • You must have a valid kubeconfig file with at least one context configured. This file typically lives in ~/.kube/config.
  • Verify that kubectl can connect to your cluster without any credential errors. Run kubectl get pods in a standard terminal. If this command fails, the K8s MCP server will also fail to connect.

3. Node.js and Bun: This specific K8s MCP implementation uses Node.js and the Bun package manager. Make sure you have both installed.

bun package installer

4. Helm v3 (Optional): Helm v3 needs to be installed and configured for the server, although it might not be used. You can download it from their official website.

helm
How to Use GPT4Free Python Library
This tutorial will guide you through the installation, setup, and usage of the gpt4free library, showing you how to leverage its capabilities for various AI-powered text generation tasks.

Installation and Configuration: Setting Up Your K8s MCP Server

1. Clone the Repository:

Open your terminal and run the following command:

git clone https://github.com/Flux159/mcp-server-kubernetes.git
cd mcp-server-kubernetes

This will download the code from GitHub and move you into the project directory.

2. Install Dependencies with Bun:

This project utilizes Bun, a fast JavaScript runtime, as its package manager. Make sure you have Bun installed, and then run:

bun install

This command reads the package.json file and installs all the necessary dependencies for the project.

3. Run the K8s MCP Server:

bun run dev # for development and file watching

This will start the K8s MCP server. It will automatically connect to the kubectl context you currently have configured. Make sure your connection to the K8s cluster is working or the server will have connection issues.

Important Notes: Since the server will automatically connect to your current kubectl context you will want to make sure that Helm is properly configured if you have a Helm chart in the project.

4. Local Testing with MCP Inspector:

For quick testing of the k8's MCP server, it is recommended to use the Model Context Protocol Inspector (@modelcontextprotocol/inspector) for local testing. This tool helps you visualize and interact with the MCP server's capabilities.

npx @modelcontextprotocol/inspector node dist/index.js

Follow the on-screen instructions in the terminal to access the Inspector link in your browser. This will allow you to explore the available MCP resources and test commands.

mcp inspector

Local Testing K8s MCP Server with Claude Desktop

To integrate this K8s MCP with Claude Desktop (or another AI tool), you'll need to configure Claude to communicate with the server.

  1. Locate Claude Desktop Configuration:

Find the "claude_desktop_config.json" file in Claude Desktop's settings (usually found in a Developer or Advanced section).

2. Add the MCP Server Configuration:

Add a new entry to the mcp servers section of the "claude_desktop_config.json" file:

{
  "mcpServers": {
    "k8s-mcp": { // Or choose a descriptive name
      "command": "node",
      "args": ["/path/to/your/mcp-server-kubernetes/dist/index.js"]
    }
  }
}

Important: Replace "/path/to/your/mcp-server-kubernetes/dist/index.js" with the actual absolute path to the "dist/index.js" file in your cloned repository.

3. Test with Claude Desktop:

Restart Claude Desktop. Now, you should be able to interact with your Kubernetes cluster through Claude using natural language commands. Start by asking Claude to list your pods or create a test deployment to see if the connection to the server works. If these work, it is safe to assume that the rest will work as well.

Example 1: "Can you list the pods in the default namespace"

list pods

Example 2: "Can you create and nginx pod in the default namespace and name it nginx-pod"

create pods

Example 3: "Can you delete the nginx pod"

delete pods

Security Considerations: Protecting Your Cluster

Security is paramount when integrating AI with your Kubernetes cluster.

  • Verify Functionality and Scope: Before trusting the AI to manage critical aspects of your cluster, carefully verify what actions it's actually capable of performing. Understand the permissions associated with your kubectl context or Service Account.

Conclusion

You've now taken the first steps towards exploring the possibilities of using AI to interact with your Kubernetes cluster! By setting up a K8s MCP server, you've opened the door for AI tools to potentially assist with management and provide insights. Remember to focus on security and carefully validate the functionality before relying on the AI for critical tasks.

button
Apidog all in one image

How to Use Resend API (A Beginners Guide)Viewpoint

How to Use Resend API (A Beginners Guide)

Email is a critical component of modern web applications, used for everything from user onboarding and notifications to password resets and marketing campaigns. However, building and managing a reliable email sending infrastructure can be complex and time-consuming. This is where email API services like Resend come in. Resend offers a developer-friendly platform designed to simplify the process of sending transactional and marketing emails. It provides robust APIs, detailed analytics, and excel

Mark Ponomarev

April 29, 2025

How to Run Qwen 3 Locally with Ollama & VLLMViewpoint

How to Run Qwen 3 Locally with Ollama & VLLM

The landscape of large language models (LLMs) is evolving at breakneck speed. Models are becoming more powerful, capable, and, importantly, more accessible. The Qwen team recently unveiled Qwen3, their latest generation of LLMs, boasting impressive performance across various benchmarks, including coding, math, and general reasoning. With flagship models like the Mixture-of-Experts (MoE) Qwen3-235B-A22B rivaling established giants and even smaller dense models like Qwen3-4B competing with previou

Ashley Innocent

April 29, 2025

How to Use Deepseek V3 with Cursor for FreeViewpoint

How to Use Deepseek V3 with Cursor for Free

Learn to use Deepseek V3 for FREE with Cursor in this step-by-step beginner's guide! Code a Python factorial function with AI. My takes: fast and cost-effective!

Ashley Goolam

April 29, 2025