stitch-kit
Health Pass
- License — License: Apache-2.0
- Description — Repository has a description
- Active repo — Last push 0 days ago
- Community trust — 22 GitHub stars
Code Fail
- execSync — Synchronous shell command execution in bin/stitch-kit.mjs
Permissions Pass
- Permissions — No dangerous permissions requested
This MCP server and plugin integrates Google Stitch into AI coding agents like Claude Code and Codex CLI. It acts as a design intelligence layer, helping agents generate UI mockups, iterate on designs, and convert them into framework-specific code.
Security Assessment
The tool requires a Google API key to connect to the remote Stitch server, which means you must handle and store this sensitive credential safely on your machine. The automated scan flagged the use of synchronous shell command execution (`execSync`) in the main script. While this is common in developer tooling for running setup commands, it still poses a code execution risk if the inputs are ever manipulated. The repository does not request broadly dangerous system permissions, and no hardcoded secrets were detected. Overall risk: Medium.
Quality Assessment
The project is very new and lacks widespread community validation, currently sitting at 22 GitHub stars. However, it is actively maintained, with the most recent code push happening today. The codebase is properly protected under the permissive Apache-2.0 license, and the developer has provided a thorough, well-documented README with clear installation and usage instructions.
Verdict
Use with caution. The tool is active, licensed, and useful, but its small user base and reliance on synchronous shell execution warrant a careful review of the code before entrusting it with your Google API key.
Design superpowers for AI coding agents. 35 skills, 5 layers, 7 framework targets. Ideation, batch generation, design systems, iteration loops, and production conversion via Google Stitch MCP. Claude Code plugin + Codex CLI agent. The design intelligence layer your agent was missing.
stitch-kit — Design superpowers for AI coding agents | Google Stitch MCP plugin
I built this because I got tired of watching Claude Code generate UIs that look like a government form from 2004. Gray boxes, blue buttons, zero taste. Meanwhile, Google's Stitch generates genuinely beautiful, pixel-perfect screens from text — but it's just a raw MCP tool sitting there, and no coding agent knows how to use it properly.
So I taught them.
stitch-kit is 35 skills that give AI coding agents actual design intelligence. Not "here's a prompt, good luck" — real ideation, visual research, prompt engineering that produces quality output, multi-screen batch generation, design systems, iteration loops, and conversion to production framework components. The whole pipeline from "I have a vague idea" to shippable code.
Your agent goes from design-blind to design-competent. You describe what you want (or just brainstorm out loud), and the agent researches trends, proposes 3 design directions with real color palettes and typography, generates a full PRD, creates up to 10 screens in Stitch in one shot, iterates until it looks right, and converts to Next.js, Svelte, React, React Native, SwiftUI, or HTML.
Works in Claude Code, Codex CLI, and any agent that speaks MCP.
Install
Quick install (recommended)
npx @booplex/stitch-kit
One command. It auto-detects your setup (Claude Code, Codex CLI, or both), installs the agent + skills, configures Stitch MCP, and tells you if anything's missing.
npx @booplex/stitch-kit update # update to latest
npx @booplex/stitch-kit status # check what's installed
After installing, sign in at stitch.withgoogle.com to complete Google auth.
Claude Code plugin (optional, adds skills)
The NPX installer sets up the agent and MCP. For the full skill set (ideation, prompt engineering, design systems, iteration, conversion), also install the plugin inside Claude Code:
/plugin marketplace add https://github.com/gabelul/stitch-kit.git
/plugin install stitch-kit@stitch-kit
The agent works standalone with MCP tools, but skills add structured workflows that make the output noticeably better.
Manual setup (if you prefer)
Claude Code — manual steps# 1. Add Stitch MCP (remote HTTP server — needs API key from stitch.withgoogle.com/settings)
claude mcp add stitch --transport http https://stitch.googleapis.com/mcp \
--header "X-Goog-Api-Key: YOUR-API-KEY" -s user
# 2. Install the plugin (inside Claude Code)
/plugin marketplace add https://github.com/gabelul/stitch-kit.git
/plugin install stitch-kit@stitch-kit
Codex CLI — manual steps
git clone https://github.com/gabelul/stitch-kit.git
cd stitch-kit && bash install-codex.sh
Then add Stitch MCP to ~/.codex/config.toml:
[mcp_servers.stitch]
url = "https://stitch.googleapis.com/mcp"
[mcp_servers.stitch.headers]
X-Goog-Api-Key = "YOUR-API-KEY"
Get your API key at stitch.withgoogle.com/settings.
Use $stitch-kit to activate the agent or $stitch-orchestrator for the full pipeline.
What actually happens when you use it
Without stitch-kit, your agent sends Stitch a half-baked prompt, gets confused by ID formats, generates one screen, hands you raw HTML, and calls it a day. With stitch-kit:
Think first —
stitch-ideatedoes something your coding agent literally cannot: design research. It fetches trends, analyzes competitor UIs, and proposes 3 distinct design directions with real hex colors, typography pairings, and mood descriptions. It's the design buddy your agent was missing.Generate smart — The orchestrator detects whether your request needs ideation or can go straight to generation (there's a specificity scoring system — hex colors and layout details skip ideation, vague requests trigger it). It engineers a structured prompt, sends the full PRD to Stitch, and batch-generates up to 10 screens in one call. If there are more, it auto-continues.
Iterate without starting over — Edit screens with text prompts, generate design variants, apply design systems across multiple screens for visual consistency. The agent knows which Stitch tool to reach for (and more importantly, which ID format each one expects — because apparently Google couldn't pick one).
Ship real code — Convert to production components with dark mode, TypeScript, design tokens, and ARIA. Not "here's some JSX, figure it out" — proper Server/Client component split, theme integration, accessibility baked in.
There's also an agent definition (agents/stitch-kit.md) for both Claude Code and Codex — a Stitch-aware agent that knows the full skill set and doesn't hallucinate MCP tool names (which happens more than you'd think).
Architecture
Five layers. I didn't build 35 skills for fun — each layer exists because agents consistently fail at something specific with Stitch.
The ID format thing deserves its own callout. generate_screen_from_text wants "3780309359108792857". list_screens wants "projects/3780309359108792857". Pass the wrong one and you get a cryptic error. This is the #1 reason agents fail with raw Stitch MCP, and it's the reason I built 14 wrapper skills instead of letting agents call the API directly.
Details → docs/architecture.md
Skill anatomy
skills/[skill-name]/
├── SKILL.md — what it does, when to use it, step-by-step workflow
├── examples/ — real examples so the agent copies patterns instead of guessing
├── references/ — design contracts, checklists (loaded only when needed)
└── scripts/ — fetch helpers, validators, code generators
The examples folder is the secret weapon. Agents produce dramatically better output when they have real patterns to copy instead of hallucinating boilerplate from training data. (I learned this the hard way after watching Claude generate the same broken Tailwind config 15 times in a row.)
vs. the official Google Stitch Skills
Google published 6 skills. They cover the basics — enhance a prompt, convert to React, make a video walkthrough. Useful starting point.
stitch-kit has 35. It covers the full pipeline: conversational ideation with web research, multi-screen batch generation, design system management, iteration loops, and conversion to 7 frameworks. Every official skill has a local equivalent that does more:
| Official | stitch-kit | What's different |
|---|---|---|
design-md |
stitch-design-md |
Adds Section 6 — design system notes that feed back into Stitch prompts for consistent multi-screen output |
enhance-prompt |
stitch-ui-prompt-architect |
Two modes: (A) vague → enhanced, same as official; (B) Design Spec → structured [Context][Layout][Components] prompt. Mode B produces significantly better results. |
stitch-loop |
stitch-loop |
Visual verification step, explicit MCP naming, DESIGN.md integration |
react-components |
stitch-react-components |
MCP-native retrieval, optional DESIGN.md alignment |
remotion |
stitch-remotion |
Common patterns (slideshow, feature highlight, user flow), voiceover, dynamic text |
shadcn-ui |
stitch-shadcn-ui |
Init styles support, custom registries, validation checklist |
What's entirely new (doesn't exist in the official repo):
stitch-ideate— conversational design agent that researches trends, proposes directions, produces PRDs, and batch-generates all screensstitch-orchestrator— end-to-end coordinator with smart ideation routingstitch-mcp-*wrappers — all 14 Stitch API tools wrapped with ID format safety (this alone saves hours of debugging)stitch-mcp-edit-screens— iterate on existing designs without regenerating from scratchstitch-mcp-generate-variants— native variant generation with creativity controlsstitch-mcp-upload-screens-from-images— import screenshots for redesign workflowsstitch-mcp-create/update/list/apply-design-system— full Stitch Design System lifecyclestitch-ui-design-spec-generator— structured spec before prompt (better output than pure prompt enhancement)- Mobile targets:
stitch-react-native-components+stitch-swiftui-components stitch-design-system— token extraction → CSS custom properties with dark modestitch-a11y— WCAG 2.1 AA audit + auto-fixesstitch-animate— purposeful motion withprefers-reduced-motionhandledstitch-skill-creator— meta-skill for extending the plugin
Ship to any framework
Once your design is ready, pick your target:
| Target | Skill | What you get |
|---|---|---|
| Next.js 15 App Router | stitch-nextjs-components |
Server/Client split, next-themes, TypeScript strict |
| Svelte 5 / SvelteKit | stitch-svelte-components |
Runes API, scoped CSS, built-in transitions |
| Vite + React | stitch-react-components |
useTheme() hook, Tailwind, no App Router overhead |
| HTML5 + CSS | stitch-html-components |
No build step — PWA, WebView, Capacitor ready |
| shadcn/ui | stitch-shadcn-ui |
Radix primitives, token alignment with Stitch palette |
| React Native / Expo | stitch-react-native-components |
iOS + Android, useColorScheme, safe areas |
| SwiftUI | stitch-swiftui-components |
iOS 16+, @Environment(\.colorScheme), 44pt targets |
Full skill reference
All 35 skills with descriptions, layers, and the ID format table → docs/skills-index.md
MCP API schemas (JSON Schema for all 14 Stitch tools) → docs/mcp-schemas/
Prerequisites
- Stitch MCP — setup guide (Google account + API key required)
- Node.js for web framework conversions
- Xcode 15+ for SwiftUI, Expo CLI for React Native
Related
Other tools for agents that care about quality:
- slopbuster — AI text humanizer. 100+ patterns, two-pass audit, three-tier scoring. Makes AI-generated prose, code comments, and academic writing sound human.
- pixelslop — Design quality scanner. Opens real pages in Playwright, measures actual pixels, catches visual AI slop. The visual counterpart to slopbuster.
- pixeltamer — Image generation skill for AI coding agents. Two backends (OpenAI API or codex CLI signed in to ChatGPT subscription), three modes (generate, edit, compose up to 16 references), prompting doctrine for gpt-image-2 plus six production recipes.
- claude-code-skill-activator — Skill auto-detection for Claude Code. AI extracts keywords once, then fast offline matching suggests skills as you type.
Built by Gabi @ Booplex.com — because watching AI agents produce ugly UIs when beautiful tools exist was driving me nuts. Apache 2.0.
Reviews (0)
Sign in to leave a review.
Leave a reviewNo results found