Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
148 changes: 148 additions & 0 deletions content/guides/grafana-mcp-server-gemini.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,148 @@
---
description: Integrate Gemini CLI with Grafana via Docker MCP Toolkit for natural language observability.
keywords: mcp, grafana, docker, gemini, devops
title: Connect Gemini to Grafana via MCP
summary: |
Learn how to leverage the Model Context Protocol (MCP) to interact with Grafana dashboards and datasources directly from your terminal.
levels: [intermediate]
subjects: [devops]
aliases:
- /guides/use-case/devops/
params:
time: 15 minutes
---

# Integrating Gemini CLI with Grafana via Docker MCP Toolkit

This guide shows how to connect Gemini CLI to a Grafana instance using the **Docker MCP Toolkit**.

## Prerequisites

* **Gemini CLI** installed and authenticated.
* **Docker Desktop** with the **MCP Toolkit** extension enabled.
* An active **Grafana** instance.


## 1. Provisioning Grafana Access

The MCP server requires a **Service Account Token** to interact with the Grafana API. Service Account Tokens are preferred over personal API keys because they can be revoked independently without affecting user access, and permissions can be scoped more narrowly.

1. Navigate to **Administration > Users and access > Service accounts** in your Grafana dashboard.
2. Create a new Service Account (e.g., `gemini-mcp-connector`).
3. Assign the **Viewer** role (or **Editor** if you require alert management capabilities).
4. Generate a new token. Copy the token immediately—you won't be able to view it again.

![Create a service account in Grafana](images/create-sa-grafana.webp)



## 2. MCP Server Configuration

The Docker MCP Toolkit provides a pre-configured Grafana catalog item. This connects the LLM to the Grafana API.

1. Open the **MCP Toolkit** in Docker Desktop.
2. Locate **Grafana** in the Catalog and add it to your active servers.
3. In the **Configuration** view, define the following:
* **Grafana URL:** The endpoint or URL of your instance.
* **Service Account Token:** The token generated in the previous step.

![Configure mcp grafana in docker](images/configure-mcp-grafana.webp)

Check failure on line 49 in content/guides/grafana-mcp-server-gemini.md

View workflow job for this annotation

GitHub Actions / validate (vale)

[vale] reported by reviewdog 🐶 [Vale.Terms] Use 'Grafana' instead of 'grafana'. Raw Output: {"message": "[Vale.Terms] Use 'Grafana' instead of 'grafana'.", "location": {"path": "content/guides/grafana-mcp-server-gemini.md", "range": {"start": {"line": 49, "column": 17}}}, "severity": "ERROR"}

Check failure on line 49 in content/guides/grafana-mcp-server-gemini.md

View workflow job for this annotation

GitHub Actions / validate (vale)

[vale] reported by reviewdog 🐶 [Vale.Spelling] Did you really mean 'mcp'? Raw Output: {"message": "[Vale.Spelling] Did you really mean 'mcp'?", "location": {"path": "content/guides/grafana-mcp-server-gemini.md", "range": {"start": {"line": 49, "column": 13}}}, "severity": "ERROR"}



## 3. Gemini CLI Integration

To register the Docker MCP gateway within Gemini, update your global configuration file located at `~/.gemini/settings.json`.

Ensure the `mcpServers` object includes the following entry:

```json
{
"mcpServers": {
"MCP_DOCKER": {
"command": "docker",
"args": [
"mcp",
"gateway",
"run"
]
}
}
}

```


## 4. Operational Validation

Restart your Gemini CLI session to load the new configuration. Verify the status of the MCP tools by running:

```bash
> /mcp list

```

![MCP Docker added to gemini cli](images/mcp-docker-gemini.webp)

Check failure on line 85 in content/guides/grafana-mcp-server-gemini.md

View workflow job for this annotation

GitHub Actions / validate (vale)

[vale] reported by reviewdog 🐶 [Vale.Terms] Use 'CLI' instead of 'cli'. Raw Output: {"message": "[Vale.Terms] Use 'CLI' instead of 'cli'.", "location": {"path": "content/guides/grafana-mcp-server-gemini.md", "range": {"start": {"line": 85, "column": 30}}}, "severity": "ERROR"}

Check failure on line 85 in content/guides/grafana-mcp-server-gemini.md

View workflow job for this annotation

GitHub Actions / validate (vale)

[vale] reported by reviewdog 🐶 [Vale.Spelling] Did you really mean 'gemini'? Raw Output: {"message": "[Vale.Spelling] Did you really mean 'gemini'?", "location": {"path": "content/guides/grafana-mcp-server-gemini.md", "range": {"start": {"line": 85, "column": 23}}}, "severity": "ERROR"}

A successful connection will show `MCP_DOCKER` as **Ready**, exposing over 61 tools for data fetching, dashboard searching, and alert inspection.

## Use Cases

### Datasource Discovery

Check failure on line 91 in content/guides/grafana-mcp-server-gemini.md

View workflow job for this annotation

GitHub Actions / validate (vale)

[vale] reported by reviewdog 🐶 [Vale.Spelling] Did you really mean 'Datasource'? Raw Output: {"message": "[Vale.Spelling] Did you really mean 'Datasource'?", "location": {"path": "content/guides/grafana-mcp-server-gemini.md", "range": {"start": {"line": 91, "column": 5}}}, "severity": "ERROR"}

_List all Prometheus and Loki datasources._

![List datasources](images/gemini-grafana-list-datasources.webp)



![List datasources](images/list-datasources-result.webp)


### Logs Inspection

Gemini performs intent parsing and translates the request into a LogQL query: `{device_name="edge-device-01"} |= "nginx"`. This query targets specific logs, extracting raw OpenTelemetry (OTel) data that includes container metadata and system labels, which Gemini then uses to identify the source of the issue.

![Filter logs based on loki labels](images/mcp-docker-grafana-loki-1.webp)

Check failure on line 106 in content/guides/grafana-mcp-server-gemini.md

View workflow job for this annotation

GitHub Actions / validate (vale)

[vale] reported by reviewdog 🐶 [Vale.Spelling] Did you really mean 'loki'? Raw Output: {"message": "[Vale.Spelling] Did you really mean 'loki'?", "location": {"path": "content/guides/grafana-mcp-server-gemini.md", "range": {"start": {"line": 106, "column": 24}}}, "severity": "ERROR"}

Once the system identifies Loki as the active datasource, it translates the human intent into a precise technical command. The AI autonomously constructs a LogQL `query: {device_name="edge-device-01"} |= "nginx"`. This query targets the specific Kubernetes pod logs, extracting raw OpenTelemetry (OTel) data that includes pod UIDs and container metadata. Instead of the user writing complex syntax, the prompt acts as the bridge to pull structured data from the containerized environment

Check failure on line 108 in content/guides/grafana-mcp-server-gemini.md

View workflow job for this annotation

GitHub Actions / validate (vale)

[vale] reported by reviewdog 🐶 [Vale.Spelling] Did you really mean 'datasource'? Raw Output: {"message": "[Vale.Spelling] Did you really mean 'datasource'?", "location": {"path": "content/guides/grafana-mcp-server-gemini.md", "range": {"start": {"line": 108, "column": 47}}}, "severity": "ERROR"}


![Gemini gets the Grafana's logs from MCP docker](images/mcp-docker-grafana-loki-2.webp)

In the final step, Gemini performs reasoning over the raw telemetry. After filtering through hundreds of lines to confirm the existence of Nginx logs, Gemini extracts a specific node_filesystem_device_error buried within the stream. By surfacing this critical event, it alerts the DevOps engineer to a volume mounting issue on the edge node, transforming raw data into an actionable incident report.

Check failure on line 113 in content/guides/grafana-mcp-server-gemini.md

View workflow job for this annotation

GitHub Actions / validate (vale)

[vale] reported by reviewdog 🐶 [Vale.Spelling] Did you really mean 'node_filesystem_device_error'? Raw Output: {"message": "[Vale.Spelling] Did you really mean 'node_filesystem_device_error'?", "location": {"path": "content/guides/grafana-mcp-server-gemini.md", "range": {"start": {"line": 113, "column": 179}}}, "severity": "ERROR"}

![Gemini gives an overall about the findings](images/mcp-docker-grafana-loki-3.webp)



### Dashboard Navigation

_How many dashboards we have?_

Check warning on line 121 in content/guides/grafana-mcp-server-gemini.md

View workflow job for this annotation

GitHub Actions / validate (vale)

[vale] reported by reviewdog 🐶 [Docker.We] Avoid using first-person plural like 'we'. Raw Output: {"message": "[Docker.We] Avoid using first-person plural like 'we'.", "location": {"path": "content/guides/grafana-mcp-server-gemini.md", "range": {"start": {"line": 121, "column": 22}}}, "severity": "WARNING"}

![List datasources](images/mcp-grafana-dashboards.webp)


_Tell me the summary of X dashboard_


![List datasources](images/mcp-grafana-summary-dashboard.webp)

### Other scenarios

Imagine you get a page that an application is slow. You could:

1. Use list_alert_rules to see which alert is firing.

Check failure on line 135 in content/guides/grafana-mcp-server-gemini.md

View workflow job for this annotation

GitHub Actions / validate (vale)

[vale] reported by reviewdog 🐶 [Vale.Spelling] Did you really mean 'list_alert_rules'? Raw Output: {"message": "[Vale.Spelling] Did you really mean 'list_alert_rules'?", "location": {"path": "content/guides/grafana-mcp-server-gemini.md", "range": {"start": {"line": 135, "column": 11}}}, "severity": "ERROR"}
2. Use search_dashboards to find the relevant application dashboard.
3. Use get_panel_image on a key panel to see the performance spike visually.
4. Use query_loki_logs to search for "error" or "timeout" messages during the time of the spike.
5. If you find the root cause, use create_incident to start the formal response and add_activity_to_incident to log your findings.

## Next steps

- Learn about [Advanced LogQL queries](https://grafana.com/docs/loki/latest/query/log_queries/)
- Set up [Team-wide MCP configurations](https://modelcontextprotocol.io/docs/develop/connect-local-servers)
- Explore [Grafana alerting with MCP](https://github.com/grafana/mcp-grafana)
- Get help in the [Docker Community Forums](https://forums.docker.com)

Need help setting up your Docker MCP environment or customizing your Gemini prompts? Visit the [Docker Community Forums](https://forums.docker.com) or see the [MCP Troubleshooting Guide](https://docs.docker.com/guides/grafana-mcp-server-gemini).
Binary file added content/guides/images/configure-mcp-grafana.webp
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file added content/guides/images/create-sa-grafana.webp
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file added content/guides/images/mcp-docker-gemini.webp
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file added content/guides/images/mcp-grafana-dashboards.webp
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading