story-ui

mcp
Guvenlik Denetimi
Basarisiz
Health Gecti
  • License — License: MIT
  • Description — Repository has a description
  • Active repo — Last push 0 days ago
  • Community trust — 172 GitHub stars
Code Basarisiz
  • execSync — Synchronous shell command execution in cli/deploy.ts
  • process.env — Environment variable access in cli/index.ts
  • execSync — Synchronous shell command execution in cli/setup.ts
  • fs.rmSync — Destructive file system operation in cli/setup.ts
Permissions Gecti
  • Permissions — No dangerous permissions requested
Purpose
This MCP server acts as an AI-powered Storybook story generator. It takes natural language descriptions, screenshots, or voice inputs and automatically generates working UI component stories for various JavaScript frameworks and design systems.

Security Assessment
The overall risk is Medium. The tool uses environment variables to handle necessary credentials securely, with no hardcoded secrets detected. However, the audit flagged multiple significant code-level concerns. The codebase relies on synchronous shell command execution (`execSync`) in multiple deployment and setup files. Additionally, it performs destructive file system operations (`fs.rmSync`) during setup. While typical for an installer, running unknown or unsandboxed shell commands and deleting files directly on your host machine introduces inherent risks. The tool also integrates with external LLM providers (OpenAI, Claude, Gemini), which requires sending your prompts and potentially proprietary component data over the network. No dangerous system permissions are requested beyond local file manipulation.

Quality Assessment
The project is in excellent shape regarding community health and trust. It is actively maintained, with the most recent code push occurring today, and features a solid community backing of 172 GitHub stars. Furthermore, the code is fully compliant with standard open-source practices, distributed under the standard MIT license.

Verdict
Use with caution — the project is legitimate, active, and highly functional, but developers should inspect the setup scripts before running them locally to ensure they are comfortable with the underlying shell command executions and file deletions.
SUMMARY

AI-powered Storybook story generator — works with any framework, any design system, any LLM provider

README.md

Story UI

AI-powered Storybook story generator for any JavaScript framework and design system

npm version
License: MIT

Story UI generates working Storybook stories from natural language. Describe what you want, and the AI writes production-ready story code using your design system's actual components. Works with any framework and any component library.

Why Story UI?

  • Design-System Agnostic: React, Vue, Angular, Svelte, Web Components — bring your own component library
  • Multi-Provider AI: Claude, OpenAI, or Gemini with automatic model selection
  • Self-Healing Generation: Validates generated code with TypeScript AST parsing and auto-corrects errors through an LLM retry loop
  • Voice Canvas: Speak your component ideas and see them rendered live in Storybook
  • Storybook MCP Integration: Fetches component docs and story patterns from Storybook's MCP addon for higher-quality output
  • Zero Lock-in: Use Mantine, Vuetify, Angular Material, Shoelace, shadcn/ui, or your own components

Quick Start

# Install Story UI
npm install -D @tpitre/story-ui

# Initialize in your project
npx story-ui init

# Start generating stories
npm run story-ui

Story UI will guide you through:

  1. Selecting your JavaScript framework
  2. Choosing a design system (or using your own)
  3. Configuring your preferred AI provider
  4. Setting up component discovery

Features

AI-Powered Story Generation

Describe what you want in natural language. Story UI generates complete, working Storybook stories using your design system's components with proper imports, props, and TypeScript types.

Self-Healing Code Generation

When generated code has syntax errors, invalid imports, or forbidden patterns, Story UI automatically:

  1. Validates with TypeScript AST parsing and pattern checking
  2. Sends errors back to the LLM with correction context
  3. Retries up to 3 times, tracking error history to detect stuck loops
  4. Selects the best attempt if all retries fail

Voice Canvas

A live playground mode where you speak component ideas and see them rendered instantly in Storybook. Uses browser speech recognition with auto-submit, pauses during generation, and renders output through an iframe with react-live.

Intelligent Iteration

Continue the conversation to refine generated stories. Story UI preserves context and modifies only what you request.

Vision Support

Attach screenshots or mockups to your prompt. The AI uses them as reference when generating components.

Story Management

  • Edit, delete, and organize generated stories from the panel
  • Orphan detection for stories without chat history
  • File-based manifest system for tracking generated assets

Multi-Framework Support

Framework Design Systems Status
React Mantine, Chakra UI, Material UI, shadcn/ui, Custom Fully Supported
Vue Vuetify, Custom Fully Supported
Angular Angular Material, Custom Fully Supported
Svelte Flowbite-Svelte, Custom Fully Supported
Web Components Shoelace, Custom Fully Supported

Multi-Provider LLM Support

Provider Models Default
Claude (Anthropic) Claude Opus 4.6, Claude Sonnet 4.6, Claude Haiku 4.5 claude-sonnet-4-6
GPT (OpenAI) GPT-5.4, GPT-5.4 Mini, o4 Mini gpt-5.4
Gemini (Google) Gemini 3.1 Pro Preview, Gemini 3 Flash Preview, Gemini 2.5 Flash gemini-3.1-pro-preview

Installation

Interactive Setup (Recommended)

npx story-ui init

The installer prompts for:

  1. Framework — React, Vue, Angular, Svelte, or Web Components
  2. Design System — framework-specific options (Mantine, Vuetify, Angular Material, etc.)
  3. AI Provider — Claude (recommended), OpenAI, or Gemini
  4. Configuration — paths, ports, and API keys

Manual Configuration

Create story-ui.config.js in your project root:

module.exports = {
  // Framework: 'react' | 'vue' | 'angular' | 'svelte' | 'web-components'
  componentFramework: 'react',

  // Component library import path
  importPath: '@mantine/core',

  // Generated stories location
  generatedStoriesPath: './src/stories/generated/',

  // LLM provider: 'claude' | 'openai' | 'gemini'
  llmProvider: 'claude',

  // Import style: 'barrel' (default) or 'individual'
  // Use 'individual' for libraries without barrel exports (shadcn/ui, Radix Vue, Angular Material)
  // Use 'barrel' for libraries with index.ts barrel exports (Mantine, Chakra, Vuetify)
  importStyle: 'barrel',

  // Story configuration
  storyPrefix: 'Generated/',
  defaultAuthor: 'Story UI AI',

  // Layout rules for multi-component compositions
  layoutRules: {
    multiColumnWrapper: 'SimpleGrid',
    columnComponent: 'div',
    containerComponent: 'Container',
  },
};

Import Examples (Web Components / Custom Libraries)

For component libraries with non-standard import paths, use importExamples to teach the AI your patterns:

module.exports = {
  framework: 'web-components',
  importPath: '../../../components',

  importExamples: [
    "import '../../../components/button/button'; // For <my-button>",
    "import '../../../components/card/card'; // For <my-card>",
  ],
};

Usage

Generating Stories

Start the Story UI server:

npm run story-ui

Open Storybook and navigate to the Story UI panel. Describe what you want:

Create a product card with image, title, price, and add to cart button

Story UI generates a complete .stories.tsx file using your design system's components, writes it to your generated stories directory, and Storybook picks it up automatically via file watcher.

Iterating

Continue the conversation to refine:

Make the button green and add a quantity selector

Story UI modifies only what you requested, preserving the rest of the story.

Voice Canvas

Switch to Voice Canvas mode in the panel header. Speak your component ideas and see them rendered live. The canvas uses react-live for instant rendering without page reloads.


Storybook MCP Integration

Story UI can connect to Storybook MCP (@storybook/addon-mcp) to fetch component documentation, UI building guidelines, and existing story patterns. This enhances generation quality by ensuring output matches your codebase.

Setup

  1. Install @storybook/addon-mcp and enable experimentalComponentsManifest in .storybook/main.js:
export default {
  features: {
    experimentalComponentsManifest: true,
  },
  addons: [
    '@storybook/addon-mcp',
    // ... other addons
  ],
};
  1. Add the Storybook URL to story-ui.config.js:
module.exports = {
  // ... other config
  storybookMcpUrl: 'http://localhost:6006',
  storybookMcpTimeout: 5000,
};

When both Story UI and Storybook are running, context is fetched automatically during generation, resulting in more accurate component usage and consistent code style.


MCP Server Integration

Story UI includes a Model Context Protocol (MCP) server for direct integration with AI clients like Claude Desktop and Claude Code.

Claude Desktop

  1. Open Claude Desktop > Settings > Connectors
  2. Click Add custom connector
  3. Enter your deployment URL + /mcp-remote/mcp
  4. Restart Claude Desktop

Claude Code

# Production deployment
claude mcp add --transport http story-ui https://your-app.up.railway.app/mcp-remote/mcp

# Local development
claude mcp add --transport http story-ui-local http://localhost:4001/mcp-remote/mcp

Available MCP Tools

Once connected, these tools are available in Claude conversations:

  • generate-story — Generate Storybook stories from natural language
  • list-components — Discover available components
  • list-stories — View existing generated stories
  • get-story / update-story / delete-story — Manage stories
  • get-component-props — Get component property information
  • test-connection — Verify MCP connection

Local STDIO Server

For Claude Desktop local integration without a deployed server:

{
  "mcpServers": {
    "story-ui": {
      "command": "npx",
      "args": ["@tpitre/story-ui", "mcp"]
    }
  }
}

This requires the HTTP server running on port 4001 (npm run story-ui).


Production Deployment

Story UI can be deployed as a standalone web application. Railway is recommended for its simplicity.

Deploy to Railway

npm install -g @railway/cli
railway login
railway init
railway up

Environment Variables:

  • ANTHROPIC_API_KEY — Required for Claude models
  • OPENAI_API_KEY — Optional, for OpenAI models
  • GEMINI_API_KEY — Optional, for Gemini models
  • STORYBOOK_PROXY_ENABLED — Enable Storybook proxy mode for live demos
  • STORYBOOK_PROXY_PORT — Internal Storybook port (default: 6006)

See DEPLOYMENT.md for detailed instructions and troubleshooting.


Design System Documentation

Story UI reads your design system guidelines before every generation to produce better output.

Directory-Based (Recommended)

Create a story-ui-docs/ directory with guidelines, tokens, and component documentation:

story-ui-docs/
├── guidelines/
│   ├── accessibility.md
│   └── responsive-design.md
├── tokens/
│   ├── colors.json
│   └── spacing.md
└── components/
    ├── button.md
    └── forms.md

Single-File

For simpler setups, create story-ui-considerations.md in your project root with design system rules, color usage, and component preferences.


CLI Reference

npx story-ui init          # Initialize Story UI in your project
npx story-ui start         # Start the MCP server (default port: 4001)
npx story-ui mcp           # Start STDIO MCP server for Claude Desktop
npx story-ui status        # Check installation status and version
npx story-ui update        # Update Story UI files to latest version

Environment Variables

# .env
LLM_PROVIDER=claude
ANTHROPIC_API_KEY=sk-ant-...
OPENAI_API_KEY=sk-...         # optional
GEMINI_API_KEY=...            # optional
VITE_STORY_UI_PORT=4001

API Reference

Story Generation

Method Endpoint Description
POST /mcp/generate-story Generate story from prompt
POST /mcp/generate-story-stream Streaming story generation
POST /mcp/canvas-generate Generate + write voice canvas story
POST /mcp/canvas-save Save canvas to named story file

Component Discovery

Method Endpoint Description
GET /mcp/components List discovered components
GET /mcp/props Get component props
GET /mcp/frameworks List supported frameworks
GET /mcp/frameworks/detect Detect current project framework

Story Management

Method Endpoint Description
GET /mcp/stories List generated stories
GET /mcp/stories/:storyId Get a specific story
GET /mcp/stories/:storyId/content Get story file content
DELETE /mcp/stories/:storyId Delete a story

Provider Management

Method Endpoint Description
GET /mcp/providers List available LLM providers
GET /mcp/providers/models List models per provider
POST /mcp/providers/configure Configure a provider
POST /mcp/providers/validate Validate an API key
POST /mcp/providers/model Set active model

Manifest

Method Endpoint Description
GET /story-ui/manifest Get story manifest
GET /story-ui/manifest/poll Poll for manifest changes
POST /story-ui/manifest/reconcile Reconcile manifest with filesystem
DELETE /story-ui/manifest/:fileName Remove entry from manifest

All endpoints are also available under the /story-ui/ prefix.

Request Format

{
  prompt: string;           // User's request
  provider?: string;        // 'claude' | 'openai' | 'gemini'
  model?: string;           // Specific model ID
  previousCode?: string;    // For iterations
  history?: Message[];      // Conversation history
  imageData?: string;       // Base64 image for vision
}

Contributing

We welcome contributions! See our Contributing Guide.

Development Setup

git clone https://github.com/southleft/story-ui.git
cd story-ui
npm install
npm run build

# Test in a consumer project (important: run from the consumer directory)
cd /path/to/your-storybook-project
PORT=4001 node /path/to/story-ui/dist/mcp-server/index.js

License

MIT - Story UI Contributors


Links

Yorumlar (0)

Sonuc bulunamadi