funnelenvy-skills

skill
Guvenlik Denetimi
Uyari
Health Uyari
  • License — License: MIT
  • Description — Repository has a description
  • Active repo — Last push 0 days ago
  • Low visibility — Only 6 GitHub stars
Code Gecti
  • Code scan — Scanned 7 files during light audit, no dangerous patterns found
Permissions Gecti
  • Permissions — No dangerous permissions requested
Purpose
This skill provides a suite of AI-powered tools for Claude Code focused on positioning, competitive research, and conversion rate optimization (CRO). It enables autonomous web research, analytics audits, experiment planning, and landing page generation.

Security Assessment
Overall Risk: Low. The light code audit scanned 7 files and found no dangerous patterns, hardcoded secrets, or requests for dangerous permissions. Because the tool performs "autonomous web research" and can interface with analytics platforms (like GA4 and Adobe Analytics), it inherently makes external network requests. Depending on the user's local setup and the data fed into the commands, it could process proprietary business data or analytics configurations. However, the tool itself does not appear to execute malicious shell commands or exfiltrate data blindly.

Quality Assessment
The project is actively maintained, with its most recent push occurring today. It is properly licensed under the permissive MIT license and includes clear, comprehensive documentation with straightforward quick-start instructions. The main drawback is its low community visibility. With only 6 GitHub stars, it has not been widely tested or vetted by the broader developer community, so any hidden edge cases are unlikely to be publicly documented yet.

Verdict
Safe to use, though professionals should be mindful of their proprietary analytics data when running these commands locally.
SUMMARY

AI-powered positioning, competitive research, and CRO skills for Claude Code. Autonomous web research, analytics audits, experiment planning, and landing page generation.

README.md

FunnelEnvy Skills

AI-powered positioning and competitive research skills for Claude Code. Built by FunnelEnvy.

Works standalone. Works better with FunnelEnvy's private data layer.

Skills

Skill Version Description
positioning-framework 1.0.0 Autonomous positioning and messaging framework from web research
ga4-audit 2.2.0 GA4 analytics audit with page grouping, opportunity sizing, element interactions, and trend analysis
aa-audit 1.0.0 Adobe Analytics audit with the same output schema as ga4-audit
hypothesis-generator 1.4.0 CRO experiment engine with 32 patterns, ICE scoring, test feasibility, contrarian filtering, and LIFT sequencing
landing-page-generator 1.0.0 B2B paid landing page generator with brief, copy, design, and QA phases
positioning-update 1.0.0 Apply client feedback and corrections to positioning context files
voice-inference 1.0.0 Brand voice analysis from website content with scored tone spectrum, vocabulary fingerprint, and actionable voice rules
experiment-mockup 1.0.0 Visual mockup generator for experiment hypotheses (in active development)
render-default-deliverables 1.0.0 Generates client-ready deliverables from positioning context

Quick Start

Clone the repo and work from inside it. Claude Code discovers skills automatically.

git clone https://github.com/FunnelEnvy/funnelenvy-skills.git
cd funnelenvy-skills

Run a skill:

/positioning-framework https://example.com
/positioning-update
/ga4-audit properties/123456789
/aa-audit --config path/to/config.json
/hypothesis-generator
/landing-page-generator example-co campaign-slug --stage all
/landing-page-generator example-co campaign-slug --stage brief
/voice-inference https://example.com
/voice-inference https://example.com --mode compare
/experiment-mockup 1

The --stage flag on landing-page-generator controls which phases run: brief, copy, design, qa, or all (default).

Research output goes to .claude/context/. Deliverables go to .claude/deliverables/.

Positioning Framework Depth Levels

The positioning-framework skill supports three depth levels. Other skills do not use the --depth flag.

Depth What It Does Time Tokens
--depth quick Fast triage. Company identity + inline health check. ~5-8 min ~70-90K
--depth standard (default) Full framework. Competitive analysis, messaging, scorecard, deliverables. ~30-35 min ~450-500K
--depth deep Extended competitive. 6+ competitors, deeper sources, deliverables. ~40-50 min ~550-650K

Each depth builds on prior work. Running quick then standard then deep is incremental, not redundant.

/positioning-framework https://example.com --depth quick
/positioning-framework https://example.com --depth deep
/positioning-framework https://example.com --competitive-focus "Acme Corp"
/positioning-framework https://example.com --property properties/123456789

Output Files

Context (.claude/context/)

File Description
company-identity.md Company facts, services, differentiators, proof points, constraints
competitive-landscape.md Market overview, competitor profiles, claim overlap, white space
audience-messaging.md Personas, messaging hierarchy, language bank, voice rules
positioning-scorecard.md Positioning health check, messaging gaps, confidence scores
performance-profile.md Page performance, conversion funnels, channel/device breakdown, data quality
brand-voice.md Scored tone spectrum, vocabulary fingerprint, categorized examples, voice rules
_fetch-registry.md Internal coordination file logging all URLs fetched by each agent

Deliverables (.claude/deliverables/)

File Description
executive-summary.md Positioning assessment for executives
messaging-guide.md Persona-by-persona messaging with voice rules
experiment-roadmap.md Prioritized experiment plan (produced by hypothesis-generator)
competitive-comparison-matrix.md Structured comparison grid across competitors
battle-cards/[competitor].md One-page competitor reference cards
campaigns/[slug]/brief.md Campaign brief for a landing page
campaigns/[slug]/copy.md Section-by-section landing page copy
campaigns/[slug]/page.html Single-file HTML landing page
campaigns/[slug]/qa-report.md QA validation report
experiments/[slug]/mockup.html Standalone HTML mockup of proposed experiment change
experiments/[slug]/placement.md CRO placement rationale and implementation notes

How It Works

Skills build on each other. Each one reads from and writes to .claude/context/, creating a shared knowledge layer that downstream skills consume.

positioning-framework researches a company and its competitors, then produces structured context files with evidence-backed analysis. It runs autonomous web research across multiple source tiers (website, reviews, Reddit, SEC filings, job postings) depending on depth level. At standard and deep depth, render-default-deliverables runs automatically after it completes.

ga4-audit pulls 10-15 reports from a GA4 property via direct API (preferred) or analytics-mcp fallback. Classifies conversion events, groups pages by type, sizes opportunities, discovers element-level interactions (CTA clicks, link text, custom parameters), and detects failure modes. Produces performance-profile.md. Standalone -- works with or without positioning context, though it can optionally enrich its output with product-line mappings from company-identity.md if one exists. Requires GA4 credentials (see skills/ga4-audit/.env.example) or analytics-mcp as fallback.

aa-audit is the Adobe Analytics equivalent of ga4-audit. Runs a Python script against the AA 2.0 Reporting API and produces the same performance-profile.md schema, so all downstream skills (hypothesis-generator, render-default-deliverables) work identically regardless of analytics platform. Requires Python 3 with requests, Adobe Analytics API credentials (env vars), and a client config JSON file.

hypothesis-generator reads everything the other skills produced and generates a prioritized experiment roadmap. Without GA4 data, it works from positioning gaps alone (confidence capped at 4/5). With GA4 data, it unlocks 19 performance-driven triggers, calibrates ICE scores using real traffic numbers, adds baseline metrics and test feasibility estimates to each hypothesis, and routes infeasible experiments (insufficient traffic) to "What's Not Here" with alternative approaches.

voice-inference analyzes how a company communicates by extracting 12-15 pages across content types (homepage, product, blog, case studies, about) and building an evidence-backed voice profile. Scores tone dimensions, identifies vocabulary patterns and sentence architecture, catalogs 33+ categorized examples, and derives actionable voice rules. Two modes: observe (infer from content alone) and compare (compare inferred voice against customer-provided brand docs). Does not require positioning-framework to have been run first. Produces brand-voice.md.

experiment-mockup (in active development) takes a hypothesis from the experiment roadmap and builds a visual mockup showing the proposed change in the context of the real target page. In live mode (requires Chrome DevTools MCP), it injects the change into the user's browser, matches the site's design system using computed styles, and iterates on placement and styling in real time. In static mode (automatic fallback), it extracts page HTML and builds a standalone mockup file. Both modes produce a CRO placement rationale explaining why the element is positioned where it is, what visual hierarchy strategy it uses, and how the dev team should implement it.

positioning-update applies client feedback, stakeholder corrections, and new intelligence to existing context files. Paste an email, Slack thread, or meeting notes and it classifies each piece of information, shows you a structured change plan, and executes surgical edits after approval. No web research. Triggers deliverable re-render automatically.

render-default-deliverables converts context files into polished, shareable documents. No research, no analysis. Pure synthesis and formatting. Run it manually with /render-default-deliverables any time after editing context files.

Recommended Order

# 1. Build positioning context (who you are, your market, your competitors)
# Add --property to use GA4 data for page selection (optional)
/positioning-framework https://example.com --property properties/123456789

# 2. Apply client feedback to correct and enrich context (optional)
/positioning-update

# 3. Pull analytics data (what's actually happening on your site)
/ga4-audit properties/123456789

# 4. Generate data-informed experiment ideas
/hypothesis-generator

# 5. Analyze brand voice (standalone, works without positioning context)
/voice-inference https://example.com

# 6. Visualize specific experiment changes as mockups
/experiment-mockup 1

# 7. Re-render deliverables any time context changes
/render-default-deliverables

Each step is optional and independent, but they compound. Positioning context makes the analytics audit smarter (product-line page grouping). Analytics data makes hypothesis-generator smarter (traffic-calibrated scores, performance-driven triggers). Run what you have access to.

Prerequisites

Most skills are pure markdown with no external dependencies. These are the exceptions:

Skill Requirement Why
ga4-audit GA4 credentials (see .env.example) OR analytics-mcp Queries GA4 via direct API (preferred) or MCP fallback. Python 3 + requests for direct API.
aa-audit Python 3 + requests package, Adobe Analytics API credentials (env vars), client config JSON Runs a Python script against the AA 2.0 Reporting API
experiment-mockup (live mode) Chrome DevTools MCP Injects changes into live browser DOM. Falls back to static mode without it.

License

MIT

Yorumlar (0)

Sonuc bulunamadi