ask-llm

mcp
Guvenlik Denetimi
Uyari
Health Uyari
  • License — License: MIT
  • Description — Repository has a description
  • Active repo — Last push 0 days ago
  • Low visibility — Only 7 GitHub stars
Code Gecti
  • Code scan — Scanned 12 files during light audit, no dangerous patterns found
Permissions Gecti
  • Permissions — No dangerous permissions requested
Purpose
This tool is an MCP server designed to facilitate AI-to-AI collaboration. It bridges your primary AI client (like Claude) with other large language models such as Gemini, Codex, and Ollama to provide second opinions, review code, and critique architecture plans.

Security Assessment
The overall risk is rated as Medium. The automated code scan checked 12 files and found no dangerous patterns, hardcoded secrets, or requests for unsafe system permissions. However, the core functionality requires making outbound network requests to external LLM provider APIs. This means your data, prompts, and code diffs will be transmitted over the internet to third-party services. If you choose to use the Ollama integration, this network risk is mitigated by keeping data entirely local.

Quality Assessment
The project is actively maintained, with its most recent push happening today. It uses the standard, permissive MIT license, and the repository includes clear documentation, CI pipelines, and a structured quick-start guide. The only notable drawback is its low visibility; it currently has only 7 GitHub stars. While the technical setup looks highly professional, the small community means the tool has not been widely battle-tested or heavily vetted by a broad audience of security researchers.

Verdict
Use with caution: the codebase appears clean and well-structured, but you should be mindful of your data privacy when sending proprietary code to third-party LLM APIs.
SUMMARY

MCP server for AI-to-AI collaboration — bridge Claude with Gemini, Codex, and other LLMs for code review, second opinions, and plan debate

README.md

Ask LLM

CI
GitHub Release
License: MIT

Package Type Version Downloads
ask-gemini-mcp MCP Server npm downloads
ask-codex-mcp MCP Server npm downloads
ask-ollama-mcp MCP Server npm downloads
ask-llm-mcp MCP Server npm downloads
@ask-llm/plugin Claude Code Plugin GitHub /plugin install

MCP servers + Claude Code plugin for AI-to-AI collaboration

MCP servers that bridge your AI client with multiple LLM providers for AI-to-AI collaboration. Works with Claude Code, Claude Desktop, Cursor, Warp, Copilot, and 40+ other MCP clients. Leverage Gemini's 1M+ token context, Codex's GPT-5.5, or local Ollama models — all via standard MCP.

Why?

  • Get a second opinion — Ask another AI to review your coding approach before committing
  • Debate plans — Send architecture proposals for critique and alternative suggestions
  • Review changes — Have multiple AIs analyze diffs to catch issues your primary AI might miss
  • Massive context — Gemini reads entire codebases (1M+ tokens) that would overflow other models
  • Local & private — Use Ollama for reviews where no data leaves your machine

Quick Start

Claude Code

# All-in-one — auto-detects installed providers
claude mcp add --scope user ask-llm -- npx -y ask-llm-mcp
Or install providers individually
claude mcp add --scope user gemini -- npx -y ask-gemini-mcp
claude mcp add --scope user codex -- npx -y ask-codex-mcp
claude mcp add --scope user ollama -- npx -y ask-ollama-mcp

Claude Desktop

Add to claude_desktop_config.json:

{
  "mcpServers": {
    "ask-llm": {
      "command": "npx",
      "args": ["-y", "ask-llm-mcp"]
    }
  }
}
Or install providers individually
{
  "mcpServers": {
    "gemini": {
      "command": "npx",
      "args": ["-y", "ask-gemini-mcp"]
    },
    "codex": {
      "command": "npx",
      "args": ["-y", "ask-codex-mcp"]
    },
    "ollama": {
      "command": "npx",
      "args": ["-y", "ask-ollama-mcp"]
    }
  }
}
Cursor, Codex CLI, OpenCode, and other clients

Cursor (.cursor/mcp.json):

{
  "mcpServers": {
    "ask-llm": { "command": "npx", "args": ["-y", "ask-llm-mcp"] }
  }
}

Codex CLI (~/.codex/config.toml):

[mcp_servers.ask-llm]
command = "npx"
args = ["-y", "ask-llm-mcp"]

Any MCP Client (STDIO transport):

{ "command": "npx", "args": ["-y", "ask-llm-mcp"] }

Replace ask-llm-mcp with ask-gemini-mcp, ask-codex-mcp, or ask-ollama-mcp for a single provider.

Claude Code Plugin

The Ask LLM plugin adds multi-provider code review, brainstorming, and automated hooks directly into Claude Code:

/plugin marketplace add Lykhoyda/ask-llm
/plugin install ask-llm@ask-llm-plugins

What You Get

Feature Description
/multi-review Parallel Gemini + Codex review with 4-phase validation pipeline and consensus highlighting
/gemini-review Gemini-only review with confidence filtering
/codex-review Codex-only review with confidence filtering
/ollama-review Local review — no data leaves your machine
/brainstorm Multi-LLM brainstorm: Claude Opus researches the topic against real files in parallel with external providers (Gemini/Codex/Ollama), then synthesizes all findings with verified findings weighted higher
/compare Side-by-side raw responses from multiple providers, no synthesis — for when you want to see how each provider phrases the same answer
Pre-commit hook Reviews staged changes before git commit, warns about critical issues

The review agents use a 4-phase pipeline inspired by Anthropic's code-review plugin: context gathering, prompt construction with explicit false-positive exclusions, synthesis, and source-level validation of each finding.

See the plugin docs for details.

Prerequisites

  • Node.js v20.0.0 or higher (LTS)
  • At least one provider:
    • Gemini CLInpm install -g @google/gemini-cli && gemini login
    • Codex CLI — installed and authenticated
    • Ollama — running locally with a model pulled (ollama pull qwen2.5-coder:7b)

MCP Tools

Tool Package Purpose
ask-gemini ask-gemini-mcp Send prompts to Gemini CLI with @ file syntax. 1M+ token context. Live progressive output via stream-json
ask-gemini-edit ask-gemini-mcp Get structured OLD/NEW code edit blocks from Gemini
fetch-chunk ask-gemini-mcp Retrieve chunks from cached large responses
ask-codex ask-codex-mcp Send prompts to Codex CLI. GPT-5.5 with mini fallback. Native session resume via sessionId
ask-ollama ask-ollama-mcp Send prompts to local Ollama. Fully private, zero cost. Server-side conversation replay via sessionId
ask-llm ask-llm-mcp Unified orchestrator — pick provider per call. Fan out to all installed providers
multi-llm ask-llm-mcp Dispatch the same prompt to multiple providers in parallel; returns per-provider responses + usage in one call
get-usage-stats all Per-session token totals, fallback counts, breakdowns by provider/model — all in-memory, no persistence
diagnose ask-llm-mcp Self-diagnosis: Node version, PATH resolution, provider CLI presence + versions. Read-only
ping all Connection test — verify MCP setup

All ask-* tools accept an optional sessionId parameter for multi-turn conversations and now return a structured AskResponse (provider, response, model, sessionId, usage) via MCP outputSchema alongside the human-readable text. The orchestrator (ask-llm-mcp) also exposes usage://current-session as an MCP Resource for live JSON snapshots.

Usage Examples

ask gemini to review the changes in @src/auth.ts for security issues
ask codex to suggest a better algorithm for @src/sort.ts
ask ollama to explain @src/config.ts (runs locally, no data sent anywhere)
use gemini to summarize @. the current directory
use multi-llm to compare what gemini and codex think about this approach

CLI Subcommands

The orchestrator binary (ask-llm-mcp) supports two CLI modes alongside the default MCP server:

# Interactive multi-provider REPL — switch providers, persist sessions, see usage live
npx ask-llm-mcp repl

# Diagnose your setup — Node version, PATH, provider CLI versions, env vars
npx ask-llm-mcp doctor          # human-readable
npx ask-llm-mcp doctor --json   # machine-readable, exit 1 on error

The REPL ships sessions per provider (/provider gemini, /provider codex, /new, /sessions, /usage) and inherits all the executor behavior (quota fallback, stream-json output for Gemini, native session resume).

Models

Provider Default Fallback
Gemini gemini-3.1-pro-preview gemini-3-flash-preview (on quota)
Codex gpt-5.5 gpt-5.5-mini (on quota)
Ollama qwen2.5-coder:7b qwen2.5-coder:1.5b (if not found)

All providers automatically fall back to a lighter model on errors.

Documentation

Contributing

Contributions are welcome! See open issues for things to work on.

License

MIT License. See LICENSE for details.

Disclaimer: This is an unofficial, third-party tool and is not affiliated with, endorsed, or sponsored by Google or OpenAI.

Yorumlar (0)

Sonuc bulunamadi