# flux159_mcp-server-kubernetes **Repository Path**: mirrors/flux159_mcp-server-kubernetes ## Basic Information - **Project Name**: flux159_mcp-server-kubernetes - **Description**: MCP Server Kubernetes 是一个可以连接到 Kubernetes 集群并对其进行管理的服务器 - **Primary Language**: JavaScript - **License**: MIT - **Default Branch**: main - **Homepage**: https://www.oschina.net/p/flux159_mcp-server-kubernetes - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2025-04-07 - **Last Updated**: 2025-08-30 ## Categories & Tags **Categories**: Artificial-Intelligence, MCP **Tags**: None ## README # MCP Server Kubernetes [![CI](https://github.com/Flux159/mcp-server-kubernetes/actions/workflows/ci.yml/badge.svg)](https://github.com/yourusername/mcp-server-kubernetes/actions/workflows/ci.yml) [![Language](https://img.shields.io/github/languages/top/Flux159/mcp-server-kubernetes)](https://github.com/yourusername/mcp-server-kubernetes) [![Bun](https://img.shields.io/badge/runtime-bun-orange)](https://bun.sh) [![Kubernetes](https://img.shields.io/badge/kubernetes-%23326ce5.svg?style=flat&logo=kubernetes&logoColor=white)](https://kubernetes.io/) [![Docker](https://img.shields.io/badge/docker-%230db7ed.svg?style=flat&logo=docker&logoColor=white)](https://www.docker.com/) [![Stars](https://img.shields.io/github/stars/Flux159/mcp-server-kubernetes)](https://github.com/Flux159/mcp-server-kubernetes/stargazers) [![Issues](https://img.shields.io/github/issues/Flux159/mcp-server-kubernetes)](https://github.com/Flux159/mcp-server-kubernetes/issues) [![PRs Welcome](https://img.shields.io/badge/PRs-welcome-brightgreen.svg)](https://github.com/Flux159/mcp-server-kubernetes/pulls) [![Last Commit](https://img.shields.io/github/last-commit/Flux159/mcp-server-kubernetes)](https://github.com/Flux159/mcp-server-kubernetes/commits/main) [![Trust Score](https://archestra.ai/mcp-catalog/api/badge/quality/Flux159/mcp-server-kubernetes)](https://archestra.ai/mcp-catalog/flux159__mcp-server-kubernetes) MCP Server that can connect to a Kubernetes cluster and manage it. Supports loading kubeconfig from multiple sources in priority order. https://github.com/user-attachments/assets/f25f8f4e-4d04-479b-9ae0-5dac452dd2ed ## Usage with Claude Desktop ```json { "mcpServers": { "kubernetes": { "command": "npx", "args": ["mcp-server-kubernetes"] } } } ``` By default, the server loads kubeconfig from `~/.kube/config`. For additional authentication options (environment variables, custom paths, etc.), see [ADVANCED_README.md](ADVANCED_README.md). The server will automatically connect to your current kubectl context. Make sure you have: 1. kubectl installed and in your PATH 2. A valid kubeconfig file with contexts configured 3. Access to a Kubernetes cluster configured for kubectl (e.g. minikube, Rancher Desktop, GKE, etc.) 4. Helm v3 installed and in your PATH (no Tiller required). Optional if you don't plan to use Helm. You can verify your connection by asking Claude to list your pods or create a test deployment. If you have errors open up a standard terminal and run `kubectl get pods` to see if you can connect to your cluster without credentials issues. ## Usage with mcp-chat [mcp-chat](https://github.com/Flux159/mcp-chat) is a CLI chat client for MCP servers. You can use it to interact with the Kubernetes server. ```shell npx mcp-chat --server "npx mcp-server-kubernetes" ``` Alternatively, pass it your existing Claude Desktop configuration file from above (Linux should pass the correct path to config): Mac: ```shell npx mcp-chat --config "~/Library/Application Support/Claude/claude_desktop_config.json" ``` Windows: ```shell npx mcp-chat --config "%APPDATA%\Claude\claude_desktop_config.json" ``` ## Features - [x] Connect to a Kubernetes cluster - [x] Unified kubectl API for managing resources - Get or list resources with `kubectl_get` - Describe resources with `kubectl_describe` - List resources with `kubectl_get` - Create resources with `kubectl_create` - Apply YAML manifests with `kubectl_apply` - Delete resources with `kubectl_delete` - Get logs with `kubectl_logs` - Manage kubectl contexts with `kubectl_context` - Explain Kubernetes resources with `explain_resource` - List API resources with `list_api_resources` - Scale resources with `kubectl_scale` - Update field(s) of a resource with `kubectl_patch` - Manage deployment rollouts with `kubectl_rollout` - Execute any kubectl command with `kubectl_generic` - Verify connection with `ping` - [x] Advanced operations - Scale deployments with `kubectl_scale` (replaces legacy `scale_deployment`) - Port forward to pods and services with `port_forward` - Run Helm operations - Install, upgrade, and uninstall charts - Support for custom values, repositories, and versions - Template-based installation (`helm_template_apply`) to bypass authentication issues - Template-based uninstallation (`helm_template_uninstall`) to bypass authentication issues - Pod cleanup operations - Clean up problematic pods (`cleanup_pods`) in states: Evicted, ContainerStatusUnknown, Completed, Error, ImagePullBackOff, CrashLoopBackOff - Node management operations - Cordoning, draining, and uncordoning nodes (`node_management`) for maintenance and scaling operations - [x] Troubleshooting Prompt (`k8s-diagnose`) - Guides through a systematic Kubernetes troubleshooting flow for pods based on a keyword and optional namespace. - [x] Non-destructive mode for read and create/update-only access to clusters - [x] Secrets masking for security (masks sensitive data in `kubectl get secrets` commands, does not affect logs) ## Prompts The MCP Kubernetes server includes specialized prompts to assist with common diagnostic operations. ### k8s-diagnose Prompt This prompt provides a systematic troubleshooting flow for Kubernetes pods. It accepts a `keyword` to identify relevant pods and an optional `namespace` to narrow the search. The prompt's output will guide you through an autonomous troubleshooting flow, providing instructions for identifying issues, collecting evidence, and suggesting remediation steps. ## Local Development Make sure that you have [bun installed](https://bun.sh/docs/installation). Clone the repo & install dependencies: ```bash git clone https://github.com/Flux159/mcp-server-kubernetes.git cd mcp-server-kubernetes bun install ``` ### Development Workflow 1. Start the server in development mode (watches for file changes): ```bash bun run dev ``` 2. Run unit tests: ```bash bun run test ``` 3. Build the project: ```bash bun run build ``` 4. Local Testing with [Inspector](https://github.com/modelcontextprotocol/inspector) ```bash npx @modelcontextprotocol/inspector node dist/index.js # Follow further instructions on terminal for Inspector link ``` 5. Local testing with Claude Desktop ```json { "mcpServers": { "mcp-server-kubernetes": { "command": "node", "args": ["/path/to/your/mcp-server-kubernetes/dist/index.js"] } } } ``` 6. Local testing with [mcp-chat](https://github.com/Flux159/mcp-chat) ```bash bun run chat ``` ## Contributing See the [CONTRIBUTING.md](CONTRIBUTING.md) file for details. ## Advanced ### Non-Destructive Mode You can run the server in a non-destructive mode that disables all destructive operations (delete pods, delete deployments, delete namespaces, etc.): ```shell ALLOW_ONLY_NON_DESTRUCTIVE_TOOLS=true npx mcp-server-kubernetes ``` For Claude Desktop configuration with non-destructive mode: ```json { "mcpServers": { "kubernetes-readonly": { "command": "npx", "args": ["mcp-server-kubernetes"], "env": { "ALLOW_ONLY_NON_DESTRUCTIVE_TOOLS": "true" } } } } ``` ### Commands Available in Non-Destructive Mode All read-only and resource creation/update operations remain available: - Resource Information: `kubectl_get`, `kubectl_describe`, `kubectl_logs`, `explain_resource`, `list_api_resources` - Resource Creation/Modification: `kubectl_apply`, `kubectl_create`, `kubectl_scale`, `kubectl_patch`, `kubectl_rollout` - Helm Operations: `install_helm_chart`, `upgrade_helm_chart`, `helm_template_apply`, `helm_template_uninstall` - Connectivity: `port_forward`, `stop_port_forward` - Context Management: `kubectl_context` ### Commands Disabled in Non-Destructive Mode The following destructive operations are disabled: - `kubectl_delete`: Deleting any Kubernetes resources - `uninstall_helm_chart`: Uninstalling Helm charts - `cleanup`: Cleanup of managed resources - `cleanup_pods`: Cleaning up problematic pods - `node_management`: Node management operations (can drain nodes) - `kubectl_generic`: General kubectl command access (may include destructive operations) ### Helm Template Apply Tool The `helm_template_apply` tool provides an alternative way to install Helm charts that bypasses authentication issues commonly encountered with certain Kubernetes configurations. This tool is particularly useful when you encounter errors like: ``` WARNING: Kubernetes configuration file is group-readable. This is insecure. Error: INSTALLATION FAILED: Kubernetes cluster unreachable: exec plugin: invalid apiVersion "client.authentication.k8s.io/v1alpha1" ``` Instead of using `helm install` directly, this tool: 1. Uses `helm template` to generate YAML manifests from the Helm chart 2. Applies the generated YAML using `kubectl apply` 3. Handles namespace creation and cleanup automatically #### Usage Example ```json { "name": "helm_template_apply", "arguments": { "name": "events-exporter", "chart": ".", "namespace": "kube-event-exporter", "valuesFile": "values.yaml", "createNamespace": true } } ``` This is equivalent to running: ```bash helm template events-exporter . -f values.yaml > events-exporter.yaml kubectl create namespace kube-event-exporter kubectl apply -f events-exporter.yaml -n kube-event-exporter ``` #### Parameters - `name`: Release name for the Helm chart - `chart`: Chart name or path to chart directory - `repo`: Chart repository URL (optional if using local chart path) - `namespace`: Kubernetes namespace to deploy to - `values`: Chart values as an object (optional) - `valuesFile`: Path to values.yaml file (optional, alternative to values object) - `createNamespace`: Whether to create the namespace if it doesn't exist (default: true) ### Pod Cleanup with Existing Tools Pod cleanup can be achieved using the existing `kubectl_get` and `kubectl_delete` tools with field selectors. This approach leverages standard Kubernetes functionality without requiring dedicated cleanup tools. #### Identifying Problematic Pods Use `kubectl_get` with field selectors to identify pods in problematic states: **Get failed pods:** ```json { "name": "kubectl_get", "arguments": { "resourceType": "pods", "namespace": "default", "fieldSelector": "status.phase=Failed" } } ``` **Get completed pods:** ```json { "name": "kubectl_get", "arguments": { "resourceType": "pods", "namespace": "default", "fieldSelector": "status.phase=Succeeded" } } ``` **Get pods with specific conditions:** ```json { "name": "kubectl_get", "arguments": { "resourceType": "pods", "namespace": "default", "fieldSelector": "status.conditions[?(@.type=='Ready')].status=False" } } ``` #### Deleting Problematic Pods Use `kubectl_delete` with field selectors to delete pods in problematic states: **Delete failed pods:** ```json { "name": "kubectl_delete", "arguments": { "resourceType": "pods", "namespace": "default", "fieldSelector": "status.phase=Failed", "force": true, "gracePeriodSeconds": 0 } } ``` **Delete completed pods:** ```json { "name": "kubectl_delete", "arguments": { "resourceType": "pods", "namespace": "default", "fieldSelector": "status.phase=Succeeded", "force": true, "gracePeriodSeconds": 0 } } ``` #### Workflow 1. **First, identify problematic pods** using `kubectl_get` with appropriate field selectors 2. **Review the list** of pods in the response 3. **Delete the pods** using `kubectl_delete` with the same field selectors #### Available Field Selectors - `status.phase=Failed` - Pods that have failed - `status.phase=Succeeded` - Pods that have completed successfully - `status.phase=Pending` - Pods that are pending - `status.conditions[?(@.type=='Ready')].status=False` - Pods that are not ready #### Safety Features - **Field selectors**: Target specific pod states precisely - **Force deletion**: Use `force=true` and `gracePeriodSeconds=0` for immediate deletion - **Namespace isolation**: Target specific namespaces or use `allNamespaces=true` - **Standard kubectl**: Uses well-established Kubernetes patterns ### Node Management Tool The `node_management` tool provides comprehensive node management capabilities for Kubernetes clusters, including cordoning, draining, and uncordoning operations. This is essential for cluster maintenance, scaling, and troubleshooting. #### Operations Available - **`list`**: List all nodes with their status and schedulability - **`cordon`**: Mark a node as unschedulable (no new pods will be scheduled) - **`drain`**: Safely evict all pods from a node and mark it as unschedulable - **`uncordon`**: Mark a node as schedulable again #### Usage Examples **1. List all nodes:** ```json { "name": "node_management", "arguments": { "operation": "list" } } ``` **2. Cordon a node (mark as unschedulable):** ```json { "name": "node_management", "arguments": { "operation": "cordon", "nodeName": "worker-node-1" } } ``` **3. Drain a node (dry run first):** ```json { "name": "node_management", "arguments": { "operation": "drain", "nodeName": "worker-node-1", "dryRun": true } } ``` **4. Drain a node (with confirmation):** ```json { "name": "node_management", "arguments": { "operation": "drain", "nodeName": "worker-node-1", "dryRun": false, "confirmDrain": true, "force": true, "ignoreDaemonsets": true, "timeout": "5m" } } ``` **5. Uncordon a node:** ```json { "name": "node_management", "arguments": { "operation": "uncordon", "nodeName": "worker-node-1" } } ``` #### Drain Operation Parameters - `force`: Force the operation even if there are pods not managed by controllers - `gracePeriod`: Period of time in seconds given to each pod to terminate gracefully - `deleteLocalData`: Delete local data even if emptyDir volumes are used - `ignoreDaemonsets`: Ignore DaemonSet-managed pods (default: true) - `timeout`: The length of time to wait before giving up (e.g., '5m', '1h') - `dryRun`: Show what would be done without actually doing it - `confirmDrain`: Explicit confirmation to drain the node (required for actual draining) #### Safety Features - **Dry run by default**: Drain operations default to dry run to show what would be done - **Explicit confirmation**: Drain operations require `confirmDrain=true` to proceed - **Status tracking**: Shows node status before and after operations - **Timeout protection**: Configurable timeouts to prevent hanging operations - **Graceful termination**: Configurable grace periods for pod termination #### Common Use Cases 1. **Cluster Maintenance**: Cordon nodes before maintenance, drain them, perform maintenance, then uncordon 2. **Node Scaling**: Drain nodes before removing them from the cluster 3. **Troubleshooting**: Isolate problematic nodes by cordoning them 4. **Resource Management**: Drain nodes to redistribute workload For additional advanced features, see the [ADVANCED_README.md](ADVANCED_README.md). ## Architecture See this [DeepWiki link](https://deepwiki.com/Flux159/mcp-server-kubernetes) for a more indepth architecture overview created by Devin. This section describes the high-level architecture of the MCP Kubernetes server. ### Request Flow The sequence diagram below illustrates how requests flow through the system: ```mermaid sequenceDiagram participant Client participant Transport as Transport Layer participant Server as MCP Server participant Filter as Tool Filter participant Handler as Request Handler participant K8sManager as KubernetesManager participant K8s as Kubernetes API Note over Transport: StdioTransport or
SSE Transport Client->>Transport: Send Request Transport->>Server: Forward Request alt Tools Request Server->>Filter: Filter available tools Note over Filter: Remove destructive tools
if in non-destructive mode Filter->>Handler: Route to tools handler alt kubectl operations Handler->>K8sManager: Execute kubectl operation K8sManager->>K8s: Make API call else Helm operations Handler->>K8sManager: Execute Helm operation K8sManager->>K8s: Make API call else Port Forward operations Handler->>K8sManager: Set up port forwarding K8sManager->>K8s: Make API call end K8s-->>K8sManager: Return result K8sManager-->>Handler: Process response Handler-->>Server: Return tool result else Resource Request Server->>Handler: Route to resource handler Handler->>K8sManager: Get resource data K8sManager->>K8s: Query API K8s-->>K8sManager: Return data K8sManager-->>Handler: Format response Handler-->>Server: Return resource data end Server-->>Transport: Send Response Transport-->>Client: Return Final Response ``` See this [DeepWiki link](https://deepwiki.com/Flux159/mcp-server-kubernetes) for a more indepth architecture overview created by Devin. ## Publishing new release Go to the [releases page](https://github.com/Flux159/mcp-server-kubernetes/releases), click on "Draft New Release", click "Choose a tag" and create a new tag by typing out a new version number using "v{major}.{minor}.{patch}" semver format. Then, write a release title "Release v{major}.{minor}.{patch}" and description / changelog if necessary and click "Publish Release". This will create a new tag which will trigger a new release build via the cd.yml workflow. Once successful, the new release will be published to [npm](https://www.npmjs.com/package/mcp-server-kubernetes). Note that there is no need to update the package.json version manually, as the workflow will automatically update the version number in the package.json file & push a commit to main. ## Not planned Adding clusters to kubectx. ## Star History [![Star History Chart](https://api.star-history.com/svg?repos=Flux159/mcp-server-kubernetes&type=Date)](https://www.star-history.com/#Flux159/mcp-server-kubernetes&Date) ## 🖊️ Cite If you find this repo useful, please cite: ``` @software{Patel_MCP_Server_Kubernetes_2024, author = {Patel, Paras and Sonwalkar, Suyog}, month = jul, title = {{MCP Server Kubernetes}}, url = {https://github.com/Flux159/mcp-server-kubernetes}, version = {2.5.0}, year = {2024} } ```