Data Analyst Intelligence for AI IDEs
An AI skill that provides structured, professional data analysis workflows with built-in validation - helping AI coding assistants perform data analysis like a careful human analyst.
CrushData AI provides:
- 10 Analysis Workflows - EDA, Dashboard, A/B Test, Cohort, Funnel, Time Series, Segmentation, Data Cleaning, Ad-hoc, KPI Reporting
- 400+ Searchable Patterns - Metrics, SQL, Python, Charts, Database Tips, Common Mistakes
- Context-Building Protocol - Forces AI to ask questions and validate before delivering results
- 4 Industry Modules - SaaS, E-commerce, Finance, Marketing specific metrics
npm install -g crushdataaiThe -g flag means Global Install:
Local Install (npm install) |
Global Install (npm install -g) |
|
|---|---|---|
| Location | ./node_modules/ in current folder |
System-wide (e.g., %APPDATA%\npm\) |
| Scope | Only available in that project | Available everywhere on your computer |
| Use Case | Libraries for your project | CLI tools you want to run anywhere |
Then in any project:
cd your-project
crushdataai init --ai all # All AI IDEs
crushdataai init --ai claude # Claude Code onlyWhen you run crushdataai init --ai all, the CLI:
-
Creates
.shared/data-analyst/- Contains the BM25 search engine and 13 CSV knowledge databases (~400 rows of data analyst patterns) -
Creates AI IDE config files based on
--aiflag:Flag Creates --ai claude.claude/skills/data-analyst/SKILL.md--ai cursor.cursor/commands/data-analyst.md--ai windsurf.windsurf/workflows/data-analyst.md--ai antigravity.agent/workflows/data-analyst.md--ai copilot.github/prompts/data-analyst.prompt.md--ai kiro.kiro/steering/data-analyst.md--ai allAll of the above -
Your AI IDE automatically detects the config files and enables the
/data-analystcommand
To update the CLI and refresh your project's AI skill files:
npm install -g crushdataai@latest
# Update specific IDE (recommended):
crushdataai init --ai cursor --force
# Or update everything:
crushdataai init --forceCrushData AI now features a Connection Manager to securely handle your data credentials.
Run the connect command to open the management UI:
crushdataai connect- Supported Types: CSV, MySQL, PostgreSQL, Shopify, BigQuery, Snowflake
- Private & Secure: Credentials are stored locally on your machine (
~/.crushdataai/connections.json). They are never uploaded to any server or included in the npm package.
Note
Persistence: Once you add a connection, you can close the UI (Ctrl+C). The AI IDE reads the saved connection details directly from your local config file, so the server does NOT need to keep running.
crushdataai connectionscrushdataai init --ai allThe skill activates automatically (Claude) or via slash command (others).
Example Workflow:
- User Request: "Analyze the sales trends in
my-shop-data" - AI Action: The AI checks your saved connections.
- AI Action: The AI runs:
npx crushdataai snippet my-shop-data --lang python
- Result: The AI receives the secure code to connect to your data (read-only) and proceeds with analysis.
The skill activates automatically when you request data analysis work. Just chat naturally:
Analyze customer churn for my SaaS product
Use the slash command to invoke the skill:
/data-analyst Analyze customer churn for my SaaS product
Type / in chat to see available commands, then select data-analyst:
/data-analyst Analyze customer churn for my SaaS product
In VS Code with Copilot, type / in chat to see available prompts, then select data-analyst:
/data-analyst Analyze customer churn for my SaaS product
Analyze customer churn for my SaaS product
Create a dashboard for e-commerce analytics
Calculate MRR and ARR from subscription data
Build a cohort retention analysis
Perform A/B test analysis on conversion rates
# Search workflows
python3 .shared/data-analyst/scripts/search.py "EDA" --domain workflow
# Search metrics
python3 .shared/data-analyst/scripts/search.py "churn" --domain metric
# Search SQL patterns
python3 .shared/data-analyst/scripts/search.py "cohort" --domain sql
# Industry-specific
python3 .shared/data-analyst/scripts/search.py "MRR" --industry saas| Domain | Content |
|---|---|
workflow |
Step-by-step analysis processes |
metric |
Metric definitions with formulas |
chart |
Visualization recommendations |
cleaning |
Data quality patterns |
sql |
SQL patterns (window functions, cohorts) |
python |
pandas/polars code snippets |
database |
PostgreSQL, BigQuery, Snowflake tips |
report |
Dashboard UX guidelines |
validation |
Common mistakes to avoid |
| Industry | Key Metrics |
|---|---|
saas |
MRR, ARR, Churn, CAC, LTV, NRR |
ecommerce |
Conversion, AOV, Cart Abandonment |
finance |
Margins, ROI, Cash Flow, Ratios |
marketing |
CTR, CPA, ROAS, Lead Conversion |
- Discovery - AI asks about business context before coding
- Data Profiling - Mandatory checks before analysis
- Data Cleaning (ETL) - Handle missing values/duplicates in
etl/folder - Validation - Verify JOINs, aggregations, and totals
- Sanity Checks - Compare to benchmarks before delivery
To prevent global conflicts, the AI is instructed to:
- Check: Look for existing
venvor.venv. - Create: If missing, run
python3 -m venv venv. - Reports: Save all validation/profiling outputs to
reports/folder. Create if missing.
This prevents the common AI mistakes:
- β Wrong metric definitions
- β Duplicate row inflation
- β Incorrect JOIN types
- β Unreasonable totals
- β Cluttered workspaces (scripts are organized in
analysis/andetl/)
Apache 2.0

