Quickstart
Get from zero to a full growth plan in five commands.
Prerequisites
The 5-step workflow
Step 1: Create a config file
uvx skene-growth config --init
This creates .skene-growth.config in your current directory with sensible defaults. The file is created with restrictive permissions (0600 on Unix) since it will hold your API key.
Step 2: Configure your LLM provider
uvx skene-growth config
This shows your current configuration, then asks if you want to edit it. If you choose to edit, an interactive setup walks you through:
- Provider -- choose from openai, gemini, anthropic, lmstudio, ollama, or generic (any OpenAI-compatible endpoint)
- Model -- pick from a curated list per provider, or enter a custom model name
- Base URL -- only prompted if you select the
genericprovider - API key -- entered as a password field (hidden input)
The configuration is saved back to your .skene-growth.config file.
Tip: You can skip this step entirely by passing
--api-keyand--providerflags directly to each command, or by setting theSKENE_API_KEYandSKENE_PROVIDERenvironment variables.
Step 3: Analyze your codebase
uvx skene-growth analyze .
This scans your codebase and generates two files in ./skene-context/:
growth-manifest.json-- structured data about your tech stack, existing growth features, and growth opportunitiesgrowth-template.json-- a business-type-aware growth template with prioritized recommendations
The analysis uses your configured LLM to understand your codebase structure, detect the technology stack (framework, language, database, hosting), identify existing growth features, and surface new growth opportunities.
You can pass a different path instead of . to analyze a project elsewhere on disk:
uvx skene-growth analyze /path/to/your/project
Step 4: Generate a growth plan
uvx skene-growth plan
This reads growth-manifest.json and growth-template.json from ./skene-context/ (auto-detected) and generates a growth-plan.md file in the same directory.
The plan is produced by a "Council of Growth Engineers" analysis -- multiple specialized perspectives evaluate your codebase and converge on a prioritized growth strategy. The output includes:
- An executive summary
- Prioritized growth opportunities with implementation details
- A technical execution section with the recommended "next build"
- An implementation todo list
For onboarding-focused analysis instead of general growth, add the --onboarding flag:
uvx skene-growth plan --onboarding
Step 5: Build an implementation prompt
uvx skene-growth build
This command:
- Reads
growth-plan.mdfrom./skene-context/(auto-detected) - Extracts the Technical Execution section (the recommended next build, exact logic, data triggers, sequence)
- Uses your LLM to generate a focused implementation prompt
- Asks where you want to send it:
- Cursor -- opens via deep link
- Claude -- launches in terminal
- Show -- prints the full prompt to the terminal
The prompt is also saved to a file in ./skene-context/ for later use.
What you get
After running all five steps, your ./skene-context/ directory contains:
| File | Description |
|---|---|
growth-manifest.json | Structured analysis of your codebase: tech stack, current growth features, opportunities |
growth-template.json | Business-type-aware growth template with prioritized recommendations |
growth-plan.md | Full growth plan with executive summary, priorities, and technical execution details |
implementation-prompt.md | Ready-to-use prompt for your AI coding assistant |
Alternative: Quick one-liner
If you want to try the analysis without setting up a config file first, pass your API key inline:
uvx skene-growth analyze . --api-key "your-key"
This uses the default provider (openai) and model (gpt-4o). To use a different provider:
uvx skene-growth analyze . --api-key "your-key" --provider gemini --model gemini-3-flash-preview
Alternative: Free audit (no API key)
If you want to see what skene-growth does before configuring an LLM, the audit command runs a local preview with no API key required:
uvx skene-growth audit .
This shows a sample growth analysis report demonstrating the kind of strategic insights available with full API-powered analysis. It does not require any API key or LLM configuration.
Next steps
- Analyze command in depth -- all flags, output customization, excluding folders
- Plan command in depth -- context directories, onboarding mode, custom manifest paths
- Build command in depth -- prompt generation, Cursor/Claude integration
- Configuration reference -- config files, environment variables, precedence rules
- LLM providers -- setup for OpenAI, Gemini, Anthropic, LM Studio, Ollama, and generic endpoints