by runjivu
Provides AWS CLI functionality via a Model Context Protocol server, enabling portable use across AI tools and editors.
Provides a standalone MCP server that replicates the use_aws
tool from the Amazon Q Developer CLI, exposing AWS CLI commands through a standardized JSON‑RPC interface.
cargo install use_aws_mcp
mcp.json
):
{
"mcpServers": {
"use_aws_mcp": {
"name": "use_aws_mcp",
"command": "use_aws_mcp",
"timeout": 300,
"env": {},
"disabled": false
}
}
}
./target/release/use_aws_mcp
use_aws
tool from any MCP‑compatible client (Avante, Cursor, etc.) supplying the required JSON arguments.use_aws
CLI each timeQ: Do I need the AWS CLI installed? A: Yes, the AWS CLI must be installed and configured with credentials.
Q: Can I use multiple AWS profiles?
A: Specify profile_name
in the request arguments, or set AWS_DEFAULT_PROFILE
in the environment for shell‑based MCP clients.
Q: What happens with large command outputs? A: Responses larger than 100 KB are truncated to protect memory usage.
Q: How does the server detect write operations?
A: Operations not starting with get
, describe
, list
, ls
, search
, or batch_get
are treated as write and require explicit user acceptance.
Q: Is the server cross‑platform? A: Built with Rust, it runs on Linux, macOS, and Windows (provided the AWS CLI is available).
🌟 amazon-q-cli is great, and it is great because it has use_aws
MCP tool to interact with AWS API.
💡 Wouldn't it be greater if this use_aws
was portable, and use it across different AI tools, whichever you're currently using?
⚡ use_aws_mcp
is a standalone Model Context Protocol (MCP) server that provides AWS CLI functionality through a standardized interface.
This server replicates the functionality of the use_aws
tool from the Amazon Q Developer CLI.
Usage with Avante, MCPHub in nvim
Usage with Cursor
curl https://sh.rustup.rs -sSf | sh
cargo build --release
The binary will be available at target/release/use_aws
.
To use this server with an MCP client, first install it using Cargo:
cargo install use_aws_mcp
Then configure your MCP client with:
{
"mcpServers": {
"use_aws_mcp": {
"name": "use_aws_mcp",
"command": "use_aws_mcp",
"timeout": 300,
"env": {},
"disabled": false
}
}
}
With q cli, mcp clients are shell process, so credentials env like AWS_DEFAULT_PROFILE
are automatically transfered to mcp server.
However, non shell mcp clients like cursor cannot take advantage of this, so it is best advised to require mcp clients directly to use specific aws profile.
📋 User Flow:
aws sso login
./target/release/use_aws_mcp
The server communicates via stdin/stdout using JSON-RPC protocol.
The server provides human-readable descriptions of AWS CLI commands. You can see this in action by running the example:
cargo run --example description_demo
This will output something like:
Running aws cli command:
Service name: s3
Operation name: list-buckets
Parameters:
- max-items: "10"
- query: "Buckets[].Name"
Profile name: development
Region: us-west-2
Label: List S3 buckets with query
✅ This command is read-only (no acceptance required)
The server provides a single tool called use_aws
with the following schema:
{
"name": "use_aws",
"description": "Execute AWS CLI commands with proper parameter handling and safety checks",
"inputSchema": {
"type": "object",
"properties": {
"service_name": {
"type": "string",
"description": "AWS service name (e.g., s3, ec2, lambda)"
},
"operation_name": {
"type": "string",
"description": "AWS CLI operation name (e.g., list-buckets, describe-instances)"
},
"parameters": {
"type": "object",
"description": "Optional parameters for the AWS CLI command",
"additionalProperties": true
},
"region": {
"type": "string",
"description": "AWS region (e.g., us-west-2, eu-west-1)"
},
"profile_name": {
"type": "string",
"description": "Optional AWS profile name"
},
"label": {
"type": "string",
"description": "Optional label for the operation"
}
},
"required": ["service_name", "operation_name", "region"]
}
}
{
"name": "use_aws",
"arguments": {
"service_name": "s3",
"operation_name": "ls",
"region": "us-west-2"
}
}
{
"name": "use_aws",
"arguments": {
"service_name": "ec2",
"operation_name": "describe-instances",
"region": "us-west-2",
"parameters": {
"instance-ids": "i-1234567890abcdef0"
}
}
}
{
"name": "use_aws",
"arguments": {
"service_name": "lambda",
"operation_name": "list-functions",
"region": "us-west-2",
"profile_name": "development"
}
}
The server automatically detects read-only operations based on the operation name prefix:
get
, describe
, list
, ls
, search
, batch_get
Large outputs are automatically truncated to prevent memory issues, with a maximum response size of 100KB.
cargo test
cargo build
RUST_LOG=use_aws=debug cargo run
# Run the description demo
cargo run --example description_demo
The project is structured as follows:
src/lib.rs
: Core library with types and constantssrc/error.rs
: Error handling typessrc/use_aws.rs
: Core AWS CLI functionality (replicated from original)src/mcp_server.rs
: MCP server implementationsrc/main.rs
: Binary entry pointexamples/description_demo.rs
: Example demonstrating command descriptionsIf you do not have Cargo (the Rust package manager) installed, you can get it by installing Rust using rustup:
curl --proto '=https' --tlsv1.2 -sSf https://sh.rustup.rs | sh
Follow the on-screen instructions to complete the installation. After installation, restart your terminal and ensure Cargo is available by running:
cargo --version
You should see the installed Cargo version printed.
This project is distributed as a Rust crate. The following dependencies are managed automatically by Cargo:
tokio
serde
serde_json
eyre
bstr
convert_case
async-trait
thiserror
tracing
tracing-subscriber
crossterm
test/dev dependencies:
tokio-test
You do not need to install these manually; Cargo will handle them during installation.
MIT, Apache-2.0
This server executes AWS CLI commands, which may have security implications:
Run with debug logging to see detailed information:
RUST_LOG=use_aws=debug ./target/release/use_aws
Please log in to share your review and rating for this MCP.
Explore related MCPs that share similar capabilities and solve comparable challenges
by zed-industries
A high‑performance, multiplayer code editor designed for speed and collaboration.
by modelcontextprotocol
Model Context Protocol Servers
by modelcontextprotocol
A Model Context Protocol server for Git repository interaction and automation.
by modelcontextprotocol
A Model Context Protocol server that provides time and timezone conversion capabilities.
by cline
An autonomous coding assistant that can create and edit files, execute terminal commands, and interact with a browser directly from your IDE, operating step‑by‑step with explicit user permission.
by continuedev
Enables faster shipping of code by integrating continuous AI agents across IDEs, terminals, and CI pipelines, offering chat, edit, autocomplete, and customizable agent workflows.
by upstash
Provides up-to-date, version‑specific library documentation and code examples directly inside LLM prompts, eliminating outdated information and hallucinated APIs.
by github
Connects AI tools directly to GitHub, enabling natural‑language interactions for repository browsing, issue and pull‑request management, CI/CD monitoring, code‑security analysis, and team collaboration.
by daytonaio
Provides a secure, elastic infrastructure that creates isolated sandboxes for running AI‑generated code with sub‑90 ms startup, unlimited persistence, and OCI/Docker compatibility.