Skip to content

Security-Phoenix-demo/security-skills-claude-code

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

22 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Security Skills for Claude Code — Open Source Security Automation Toolkit

The open-source security automation toolkit for Claude Code — CTI research, SAST rule generation, secure PRD creation, vulnerability analysis, and AI-powered security engineering workflows. Built by Phoenix Security for the global AppSec, DevSecOps, and security research community.

License: MIT Claude Code Contributions Welcome

Search 595+ threat intelligence sources, generate opengrep/semgrep SAST rules for 30+ languages, create security-focused product requirements with STRIDE threat models, query NotebookLM for citation-backed research, and auto-generate living project documentation — all from your terminal with Claude Code.


Table of Contents

Overview-Repo


Overview


Powered by Phoenix Security — Open-Source Companion to the Platform

This repository is the open-source companion to the Phoenix Security platform. Each skill here distills a slice of what Phoenix does at platform scale into something you can run locally inside Claude Code. If a skill earns its keep in your workflow and you need it across hundreds of repos, with real reachability data, exploit chains, and team-level SLAs, the corresponding Phoenix product area is where to go next.

Phoenix Platform — Four Unified Product Areas

Area Tagline What it does Open-source slice in this repo
🟠 Orange — Attribution & Prioritization One backlog per team Single backlog with code→cloud reachability and business context. Normalizes 30+ scanners, dedupes, and routes to owners with SLAs. — (platform-only)
🟣 Purple — Identification & Prevention Stop bad code before merge Graph-powered SAST and exploit intelligence. Traces real taint paths, composes multi-step chains, prevents issues at PR and agent time. Security Assessment Suite + OpenGrep Rule Generator + Secure PRD Generator
🔵 Blue — Threat Intelligence & Supply-Chain Firewall Block bad packages before install Adversarially validated intelligence fused into a single decision. Enforces at agent, install, CI, and deploy to block malicious or unsafe packages pre-execution. CTI Domain Research + NotebookLM Connector + the PreToolUse Bash package guard hook in the Security Assessment Suite
🟢 Green — Agentic Remediation Minimal-diff fix PRs Minimal-diff PRs, safe alternatives, negative tests, and change plans tied to proven attack paths. Closes the loop with measurable risk reduction. — (platform-only — coming to OSS)

When the open-source skills aren't enough:

  • Need cross-repo reachability and prioritization? → Phoenix Orange
  • Need multi-step taint analysis with real exploit chains?Phoenix Purple
  • Need adversarially validated CVE intel + install-time blocking?Phoenix Blue
  • Need agentic remediation with measurable risk reduction? → Phoenix Green

The skills in this repo are designed to be useful on day one, with no Phoenix backend required. They share Phoenix's design philosophy: short feedback loops, severity-ranked findings, no padding, no 30-page reports.


What Is This Repository?

This repository is a curated collection of security skills, plugins, and automation pipelines for Claude Code — Anthropic's official CLI for AI-assisted software engineering. It turns Claude Code into a comprehensive security engineering workstation capable of threat intelligence research, vulnerability detection rule generation, secure requirements engineering, and automated security documentation.

Built and maintained by the security engineering team at Phoenix Security and released as open source under the MIT License. Every skill is designed for real-world security workflows: incident response research, AppSec shift-left, SAST pipeline creation, compliance documentation, and security architecture review.

Built for security professionals who use AI to work faster and more accurately:

  • Security engineers and SOC analysts — automate CTI gathering across 595+ sources with authority-ranked results and MITRE ATT&CK mapping
  • DevSecOps teams — generate security-focused PRDs with STRIDE threat models before writing a single line of code
  • AppSec professionals — create opengrep/semgrep SAST rules for 30+ languages with built-in false positive reduction and CWE/OWASP tagging
  • Vulnerability researchers — research CVEs with web search, then auto-generate detection rules grounded in real exploit data
  • Penetration testers and red teamers — gather OSINT and push findings to NotebookLM for citation-backed analysis
  • Engineering managers — auto-generate living project documentation with architecture maps, dependency views, and self-healing CI
  • Anyone using Claude Code — extend your terminal with structured, repeatable security automation workflows

What's Included

Overview-Repo-Detailed

Skills

Skills are instruction-based workflows that guide Claude Code's behavior. They don't execute code directly but define how Claude should approach specific tasks.

Skill Description Folder
CTI Domain Research Search 300+ curated security domains for threat intelligence, CVEs, malware, and breach reports skills/cti-search-skill/
Secure PRD Generator Generate security-focused Product Requirements Documents with threat modeling skills/secure-prd-skill/
OpenGrep Rule Generator Create opengrep/semgrep SAST rules for vulnerability detection across 30+ languages skills/opengrep-rule-generator/
OpenGrep Rule Generator Research Research CVEs/CWEs with web search, then generate detection rules from findings skills/opengrep-rule-generator-research/
NotebookLM Connector Query Google NotebookLM notebooks from Claude Code for citation-backed, source-grounded answers skills/notebooklm/
Global Research Pipeline Systematic web and YouTube research with NotebookLM ingestion skills/global-research-notebook-lm/
Project Documentation Auto-generate comprehensive project documentation skills/project%20Documentaion%20skill/
Security Assessment Suite Four complementary AppSec skills (/security-0day, /security-review, /security-assessment, /threatmodel) with active hooks, multi-language reviewer, and one-command installer skills/Security Assessment/

Plugins

Plugins provide executable functionality via MCP (Model Context Protocol) servers and CLI tools.

Plugin Description Folder
CTI Search Plugin MCP server + CLI for executing CTI searches across 595+ domains with NotebookLM integration plugins/cti-search-plugin/
Secure PRD Plugin PRD generator with Confluence, Linear, Asana, Slack, Notion, and Gmail integrations plugins/secure-prd/

Feature Descriptor — Phoenix Pipeline

Shift Security Left at the PRD Stage

The Phoenix Pipeline is a 12-role specification system for producing rigorous, security-aware product requirements. Each role is a dedicated skill file.

Role Skill File Purpose
Pipeline Navigator pipeline-navigator.skill Orchestrates the full pipeline
Context Curator context-curator.skill Extracts and cleanses input context
Scope Cutter scope-cutter.skill Defines in/out scope and goals
Constraint Distiller constraint-distiller.skill Identifies constraints and acceptance criteria
Requirements Engineer requirements-engineer.skill Creates RFC 2119 requirements with IDs
Ambiguity Hunter ambiguity-hunter.skill Flags and resolves ambiguities
Security Engineer security-engineer.skill Develops threat models and abuse cases
Contract Architect contract-architect.skill Designs APIs, events, and error taxonomy
Verification Matrix verification-matrix.skill Creates proof paths for every requirement
Batch Planner batch-planner.skill Plans incremental, verifiable delivery
Final Gate final-gate.skill Go/no-go decision with blocker list
Orchestrator orchestrator.skill Coordinates all roles and manages flow

All feature-descriptor skills live in the feature-descriptor/ folder.


Quick Start

Prerequisites

  • Claude Code installed
  • Node.js 18+ (for plugins)
  • A search API key — Brave Search (recommended, 2,000 free requests/month) or SerpAPI (100 free/month)

Installation

Choose the method that works best for you.

Method 1: Claude Marketplace (Recommended)

See the Marketplace Installation Guide for detailed steps with troubleshooting.

1. Open Claude Code
2. Navigate to Skills Marketplace
3. Search "CTI Domain Research" or "Security Skills"
4. Click Install
5. Configure API keys (see Configuration below)

Method 2: Git Clone

# Clone the repository
git clone https://github.com/Security-Phoenix-demo/security-skills-claude-code.git
cd security-skills-claude-code

# Install CTI Search Skill
cd skills/cti-search-skill && bash install.sh

# Install CTI Search Plugin
cd ../../plugins/cti-search-plugin && bash install.sh

Method 3: Direct Copy

# Copy skills
cp -r skills/cti-search-skill ~/.claude/skills/
cp -r skills/secure-prd-skill ~/.claude/skills/

# Copy and install plugin
cp -r plugins/cti-search-plugin ~/.claude/plugins/
cd ~/.claude/plugins/cti-search-plugin && npm install
cp .env.example .env   # Then edit .env with your API keys

Repository Structure

security-skills-claude-code/
│
├── README.md                              # This file — start here
├── CONTRIBUTING.md                        # How to contribute skills and plugins
├── MARKETPLACE_INSTALL.md                 # Detailed marketplace installation guide
├── LICENSE                                # MIT License
│
├── skills/                                # Instruction-based skills
│   ├── cti-search-skill/                  # CTI domain research across 300+ sources
│   ├── cti-search-skill/                  # (contains cti-domain-research.skill)
│   ├── secure-prd-skill/                  # Security-focused PRD generation
│   ├── opengrep-rule-generator/           # SAST rule generation (30+ languages)
│   ├── opengrep-rule-generator-research/  # CVE/CWE research + rule generation
│   ├── notebooklm/                        # NotebookLM notebook querying
│   ├── global-research-notebook-lm/       # Research pipeline with NotebookLM
│   ├── research-pipeline.skill            # Research automation skill
│   └── project Documentaion skill/        # Auto project documentation
│
├── plugins/                               # Executable tools (MCP + CLI)
│   ├── cti-search-plugin/                 # CTI search engine
│   │   ├── index.js                       # CLI entry point
│   │   ├── mcp-server.js                  # MCP tool server
│   │   ├── package.json                   # Dependencies
│   │   ├── .env.example                   # Environment template
│   │   ├── install.sh                     # Installer
│   │   └── data/
│   │       ├── domains.txt                # 595 curated security domains
│   │       └── tier-map.json              # Domain tier + authority scores
│   │
│   └── secure-prd/                        # PRD generator plugin
│       ├── prd-generator.skill            # Skill definition
│       └── prd-generator-plugin.jsx       # UI component
│
└── feature-descriptor/                    # Phoenix Pipeline (12 specialized roles)
    ├── pipeline-navigator.skill
    ├── context-curator.skill
    ├── scope-cutter.skill
    ├── constraint-distiller.skill
    ├── requirements-engineer.skill
    ├── ambiguity-hunter.skill
    ├── security-engineer.skill
    ├── contract-architect.skill
    ├── verification-matrix.skill
    ├── batch-planner.skill
    ├── final-gate.skill
    └── orchestrator.skill

Research & Intelligence Skills

Skills that gather, organize, and verify external knowledge — threat intel, vendor docs, vulnerability research — and feed it into your workflow with citation-backed accuracy.

1. CTI Domain Research — Automated Threat Intelligence Gathering

Folder: skills/cti-search-skill/

Research, Verify, Detect: Structured Threat Intelligence for Claude Code

Stop manually searching BleepingComputer, Krebs on Security, and vendor blogs one tab at a time. The CTI Domain Research skill transforms Claude Code into a structured threat intelligence platform that searches 595+ curated security domains in seconds — covering government advisories, vendor research labs, security news, and OSINT sources in a single query.

The skill uses a four-tier authority ranking system to ensure the most trustworthy sources surface first. When you search for a CVE, it automatically prioritizes CISA, NVD, and MSRC (Tier 1) before checking Unit42, Talos, and Securelist (Tier 2). Threat actor queries route to vendor research blogs first. Exploit searches hit GreyNoise, VulnCheck, and AttackerKB. Every result is deduplicated, scored by source authority and recency, and returned as a structured CTI brief with extracted CVE IDs, MITRE ATT&CK technique mappings, and observed IOCs.

The optional --notebooklm flag pushes all discovered source URLs directly into a Google NotebookLM notebook, creating a permanent, citation-backed research archive that you can query later with zero hallucination risk. This turns a one-off search into a reusable knowledge base.

Example prompts:

Search for threat intelligence on CVE-2024-21762
Find recent LockBit ransomware reports across vendor blogs
What are security vendors saying about ALPHV BlackCat?
Research MITRE T1190 exploitation techniques and push to NotebookLM
Collect CTI on supply chain attacks targeting npm packages

How it works:

  1. Query classification — automatically detects whether you're searching for a CVE, threat actor, malware family, exploit, or general topic
  2. Intelligent tier routing — selects the most relevant domain tiers based on query type
  3. Batched site-scoped search — constructs site: queries across domain batches using Brave Search or SerpAPI
  4. Authority-ranked deduplication — removes duplicates, scores results by source tier and recency
  5. IOC and TTP extraction — pulls CVE IDs, MITRE ATT&CK T-IDs, and IP indicators from result snippets
  6. Structured output — returns a CTI brief with key findings, source table, observed tags, and next steps

2. NotebookLM Connector — Source-Grounded AI Research with Zero Hallucinations

Folder: skills/notebooklm/

LLM-based security research has a fundamental problem: hallucinations. When Claude can't find something in your uploaded documents, it fills the gap with plausible-sounding but potentially incorrect information — a dangerous failure mode when you're writing detection logic, threat models, or security requirements. The NotebookLM Connector solves this by routing questions through Google NotebookLM, which answers exclusively from your uploaded documents with strict citation backing.

Every response from NotebookLM is grounded in the specific documents you've uploaded — PDFs, Google Docs, websites, GitHub repos, YouTube videos. If the information isn't in your sources, NotebookLM says so instead of inventing an answer. This makes it the ideal research backend for security work where accuracy is non-negotiable: vulnerability analysis, compliance documentation, API specification lookups, and threat model validation.

The skill manages a notebook library so Claude automatically selects the right notebook for your question. Ask about authentication best practices and it routes to your security-docs notebook. Ask about API endpoints and it hits your architecture notebook. Each question runs in a fresh browser session with persistent authentication, and the built-in follow-up mechanism ensures Claude asks comprehensive questions until the research is complete.

Capabilities:

  • Query any NotebookLM notebook by ID or URL with citation-backed responses
  • Smart notebook library management — add, remove, list, search, activate/deactivate notebooks
  • Automatic content discovery — query a notebook to auto-populate its metadata before saving
  • Browser automation with persistent Google authentication
  • Iterative follow-up queries to build comprehensive understanding
  • Coverage analysis to ensure all parts of your question are fully answered

Example prompts:

Query my security-docs notebook about authentication best practices
Add this NotebookLM URL to my library: https://notebooklm.google.com/notebook/abc123
What does my threat-model notebook say about SSRF risks?
Check my API docs for rate limiting implementation details
Search my notebooks for information about OAuth2 token rotation

Key files:

  • SKILL.md — skill specification with decision flow and follow-up mechanism
  • README.md — comprehensive setup guide with architecture diagram and examples
  • AUTHENTICATION.md — step-by-step Google authentication setup
  • scripts/ — Python automation scripts (ask_question.py, notebook_manager.py, auth_manager.py)
  • references/ — API reference, troubleshooting guide, usage patterns

3. Global Research Pipeline — Automated Intelligence Collection and Ingestion

Folder: skills/global-research-notebook-lm/

Individual searches give you snapshots. The Global Research Pipeline gives you systematic coverage. This skill orchestrates a multi-module research automation pipeline that collects intelligence from web searches and YouTube video transcripts, deduplicates and organizes findings, and pushes everything into Google NotebookLM for permanent, source-grounded querying.

The pipeline is designed for deep-dive research scenarios where you need to gather comprehensive intelligence on an emerging threat, a new vulnerability class, or a complex security topic. Instead of running individual searches and manually copying results, the pipeline handles the entire workflow: systematic data collection across web and video sources, automated deduplication, structured formatting, and batch ingestion into your NotebookLM research archive.

This creates a repeatable research-to-analysis pipeline: search → collect → deduplicate → ingest → query. Once findings are in NotebookLM, you can ask follow-up questions with full citation backing, cross-reference information across sources, and build on previous research without re-running searches.

Pipeline stages:

  1. Web research — structured searches across relevant domains with finding extraction
  2. YouTube research — locate and transcribe relevant video content (conference talks, vendor webinars, researcher presentations)
  3. Deduplication and organization — remove duplicates, structure results for ingestion
  4. NotebookLM push — batch-add all collected sources to your target notebook
  5. Downstream querying — use the NotebookLM Connector skill to query your research with zero hallucination

Example prompts:

Research the latest ransomware trends and push findings to NotebookLM
Collect comprehensive intelligence on supply chain attacks targeting Python packages
Research cloud security misconfigurations across AWS, GCP, and Azure

Security Generation & Remediation Skills

Phoenix Security Skill Detailed Overview — security generation and remediation skills for Claude Code

A focused set of skills that generate, audit, or remediate security issues across the SDLC — shifting security left into requirements, generating SAST detection rules from CVE/CWE research, documenting security-relevant architecture, and running diff-time / pre-merge / pre-release / design-time AppSec workflows on the codebase. Together they cover the full lifecycle from spec to ship.

1. Secure PRD Generator — Shift Security Left to the Requirements Stage

Folder: skills/secure-prd-skill/

Shift Security Left at the PRD Stage

Traditional PRDs focus exclusively on features and leave security as an afterthought — discovered too late in testing, patched as a hotfix, or never addressed at all. The Secure PRD Generator integrates security into the requirements phase itself, running every feature description through a 10-role specification pipeline that produces RFC 2119-compliant requirements with STRIDE threat models, abuse cases, and verification proof paths built in from the start.

Each role in the pipeline handles a specialized aspect of specification: the Context Curator cleanses and structures the input; the Ambiguity Hunter flags vague instructions that cause downstream design flaws; the Requirements Engineer formalizes requirements with MUST/SHOULD/MAY levels and structured IDs (R-FUNC-001, R-SEC-001); the Security Engineer develops comprehensive threat models and abuse cases alongside the functional requirements; and the Verification Matrix creates concrete proof paths for every MUST-level requirement to ensure full testability.

The skill integrates directly with your team's existing tools. It publishes the PRD to Atlassian Confluence via MCP, converts requirements into Linear or Asana tasks automatically, sends Slack notifications to stakeholders, mirrors documents to Notion, and drafts summary emails via Gmail. Every output follows a consistent security-first template with priority tags (P0/P1/P2) and traceable requirement chains.

What it produces:

  • Security-first PRD in structured markdown with RFC 2119 requirement levels
  • STRIDE threat model with identified assets, actors, entry points, and trust boundaries
  • Cursor-compatible implementation plan for .cursor/plans/
  • Confluence page published to your configured space
  • Automated task distribution to Linear or Asana from the batch plan
  • Stakeholder notifications via Slack and Gmail

Example prompts:

Write a PRD for a user authentication system with OAuth2 and MFA
Create a security-focused spec for a payment processing API
Plan this feature: real-time webhook delivery system. Owner: @jane. Space: ENG.
Generate requirements for a file upload service with virus scanning

2. OpenGrep Rule Generator — AI-Powered SAST Rule Creation

Folder: skills/opengrep-rule-generator/

Generate Automatically Opengrep Rules

Writing effective SAST rules is slow, error-prone, and requires deep knowledge of both the vulnerability class and the rule engine's syntax. The OpenGrep Rule Generator automates the entire process — from vulnerability description to validated, production-ready opengrep/semgrep YAML rules with test cases, CWE metadata, and false positive reduction patterns.

The skill supports 30+ programming languages including Python, JavaScript, TypeScript, Java, Go, Ruby, PHP, C#, Rust, Kotlin, Swift, Terraform/HCL, and Solidity. It generates two types of rules: Search rules for structural pattern matching (finding dangerous function calls, insecure configurations, hardcoded secrets) and Taint rules for data flow analysis (tracing untrusted input from sources through propagators to dangerous sinks, with sanitizer awareness).

Every generated rule includes built-in false positive reduction using pattern-not, pattern-not-inside, and metavariable-regex patterns. The skill also generates companion test files with clearly marked true positive and true negative cases so you can validate detection accuracy before deploying to your CI pipeline.

Two workflows:

  • Guided Discovery — interactive Q&A where Claude asks about your codebase, frameworks, and threat model before generating targeted rules
  • Vulnerability-Driven — provide a CVE, CWE, or OWASP category and get rules generated automatically with appropriate detection patterns

Example prompts:

Create an opengrep rule to detect SQL injection in Python Flask apps
Generate a taint analysis rule for XSS in React components
Write semgrep rules for OWASP Top 10 in Java Spring Boot
Detect hardcoded AWS credentials in any language
Build a rule to catch insecure deserialization in Python pickle usage
Generate Terraform rules to detect publicly exposed S3 buckets

Key files:

  • SKILL.md — full skill specification with guided and vulnerability-driven workflows
  • RULES_SYNTAX.md — comprehensive opengrep/semgrep syntax reference (patterns, operators, metavariables, taint mode)
  • OPENGREP_RULE_GENERATOR_PROMPT.md — optimized system prompt for high-quality rule generation

3. OpenGrep Rule Generator Research — Vulnerability-First Detection Engineering

Folder: skills/opengrep-rule-generator-research/

Generate Automatically Opengrep Rules and research vulnerabilities

The standard OpenGrep Rule Generator works great when you already know the pattern you want to detect. But what if you're starting from a CVE advisory, a vulnerability class, or a vague report about a new attack technique? The Research variant adds a 4-phase vulnerability research pipeline that uses web search and web fetch to study the vulnerability in depth before generating detection rules — producing significantly better, more targeted rules that are grounded in real-world exploit data.

This is the difference between writing a rule from a description and writing a rule from understanding. The research phase studies official advisories, analyzes proof-of-concept exploits, maps language-specific attack surfaces (sources, sinks, sanitizers, propagators), and reviews existing semgrep/opengrep rules to identify coverage gaps. The resulting rules are inherently linked to real exploit behavior, not abstract patterns.

Research pipeline (4 phases):

  1. Vulnerability deep dive — search CVE/CWE databases, fetch CISA/NVD/MSRC advisories, study proof-of-concept exploits to understand the actual attack mechanics
  2. Attack surface mapping — identify specific sources, sinks, and certifiers for the target programming language based on the researched exploit behavior
  3. Existing detection gap analysis — search for existing semgrep/opengrep rules, analyze what they catch and miss, identify coverage blind spots
  4. Grounded rule generation — generate SAST rules that are directly linked to the researched exploit patterns, with a research summary embedded in each rule file

Example prompts:

Research CVE-2024-21762 and create detection rules for it
Generate opengrep rules for CWE-89 (SQL Injection) in Python with full research
Investigate Log4Shell and build comprehensive detection coverage for Java
Research SSRF vulnerabilities in Node.js and create taint analysis rules
Study the MOVEit Transfer vulnerability and write detection rules

When to use this vs. the standard OpenGrep Rule Generator:

  • Use OpenGrep Rule Generator when you already know the exact code pattern you want to detect
  • Use OpenGrep Rule Generator Research when you're starting from a CVE ID, CWE class, or vulnerability description and need to understand the attack before building detection

4. Project Documentation — Turn Any Codebase into Living Documentation

Folder: skills/project Documentaion skill/

Project Documentation: Turning Codebases into Living Documentation

Documentation drift is the silent killer of engineering velocity. Code evolves daily while docs go stale within weeks, creating security blind spots, tribal knowledge silos, and painful onboarding experiences. The Project Documenter skill reverse-engineers your entire codebase and produces a three-layer hierarchical documentation system that stays accurate because it's generated from the code itself.

The skill produces a compressed DOC_INDEX.md routing layer (~1-2k tokens, perfect for AI assistants), an engineering CLAUDE.md navigation map (150-300 lines for human developers), and a complete set of /docs/ sub-documents containing all authoritative detail — architecture diagrams, dependency maps, API references, module guides, runtime flow descriptions, and onboarding walkthroughs. Nothing is duplicated across layers.

Available in six modes to match your needs: Express (full pack, zero questions), General (developer-facing docs), AI/LLM (prompt inventories, model architecture, agent safety), Architecture (system maps, data contracts, dependency graphs), Full (everything plus audit of existing docs), and Self-Heal (generates GitHub Actions CI, Cursor rules, and Python drift-detection scripts to keep docs permanently in sync with code changes).

Output artifacts:

  • Project Summary — high-level overview of system purpose and function
  • Architecture Diagram — visual mapping of system components and data flow
  • Dependency View — comprehensive list and map of third-party libraries and internal links
  • Service/Component Map — structural breakdown of project organization
  • Onboarding Document — guided walkthrough for new developers and auditors
  • Self-Heal CI (Mode 5) — GitHub Actions workflow + Python scripts that detect documentation drift automatically

Example prompts:

Document this project (Express mode — full pack, zero questions)
Generate architecture documentation for this codebase
Create AI/LLM documentation including prompt inventory and model guardrails
Set up self-healing documentation with GitHub Actions CI

Key files:

  • project-documenter.skill — full skill definition (install in Claude.ai or Claude Code)
  • project-documenter-simpler.skill — lightweight variant for quick documentation
  • HOW_IT_WORKS.md — complete technical reference
  • MODES_REFERENCE.md — detailed guide to every mode and its outputs
  • INSTALL.md — step-by-step installation and first-run guide
  • TROUBLESHOOTING.md — common issues and fixes

5. Security Assessment Suite — Four AppSec Skills + Active Hooks

Security Assessment Suite — four AppSec skills and active hooks for Claude Code

Folder: skills/Security Assessment/ — see the suite README for the full reference.

A self-contained AppSec automation kit: four slash commands covering the security lifecycle from diff-time to design-time, a multi-language pre-merge reviewer with subagent dispatch, four hooks (SessionStart, PreToolUse, PostToolUse, SessionEnd) that feed live security context to every agent, and a one-command installer that wires it all into Claude Code, Windsurf, or Codex.

The four skills (when to use each):

Command Use when Cost Engine
/security-0day [base-ref] End of a coding cycle, before opening a PR. Diff-only LLM scan. Low (~$0.05–$0.20) 0day-scanner/SKILL.md (with bundle language packs as on-disk fallback)
/security-review [scope] Endpoint, auth/RBAC, render, dependency, or config change. Pre-merge gate. Low–Medium Multi-language reviewer in Security-automated-claude-skills/ (Python, JS/TS, Go, Java/Kotlin, Rust, Ruby, .NET)
/security-assessment [scope] Pre-release, compliance audit, post-incident. Full OWASP Top 10 (2025) + ASVS Level 1 sweep. High (~$8–$10) security-assessment/SKILL.md (with bundle OWASP/ASVS checklists as fallback)
/threatmodel [scope] Architecture review, new-feature design, compliance docs. Medium threat-modeling/SKILL.md — STRIDE + DREAD with attack trees and mitigation matrix

Active hooks (full preset only — opt out for the lite preset):

  • SessionStart fingerprints the project, runs a fast dependency audit (osv-scanner if installed, else npm/pip/cargo/go/bundle audits per ecosystem), and injects a ## SECURITY CONTEXT block every agent reads before its first turn.
  • PreToolUse on Bash gates npm/yarn/pnpm/pip/uv/poetry/cargo/go get/gem/bundle/composer/dotnet add invocations. Blocks known-malicious packages, asks on typosquats and brand-new packages.
  • PostToolUse on Edit|Write|MultiEdit runs a fast pattern scan on every file write (SQL string formatting, innerHTML, hardcoded secrets, etc.) and feeds findings back via additionalContext.
  • SessionEnd prints a one-line reminder to run /security-0day if your branch has unscanned changes vs main. Zero LLM cost.

Install — one command:

# From your project root, after cloning this repo (or installing the marketplace plugin):
bash "skills/Security Assessment/install/install.sh" --full

That's it. The installer:

  1. Copies the four slash commands into .claude/commands/.
  2. Merges the chosen hook preset into .claude/settings.json (uses jq if available; backs up your existing settings first; tracks installer-created files so --uninstall is clean).
  3. Copies the security-reviewer subagent into .claude/agents/ (full preset only).
  4. Chmods all hook scripts.

Variants:

  • install.sh (default) or install.sh --lite — slash commands + SessionEnd reminder hook only. Zero LLM cost.
  • install.sh --full — everything in lite plus the three active hooks plus the subagent.
  • install.sh --dry-run [--lite|--full] — preview without writing.
  • install.sh --uninstall — remove commands + subagent, restore .claude/settings.json from the backup.

Other tools:

  • Windsurfcp a rule into .windsurf/rules/ (auto-fires on endpoint/auth/render/dep changes) and two workflows (/security-assessment, /threatmodel) into .windsurf/workflows/. One-line install in the suite README.
  • Codex CLIcat the AGENTS.md.snippet onto your project's AGENTS.md. Codex has no hook system, so this is enforced as a behavioral instruction.

Optional dependencies:

  • jq — recommended for clean settings.json merge (installer falls back to copy-paste instructions if absent).
  • ripgrep — required by the post-edit-quickscan hook (already required by Claude Code itself).
  • osv-scanner — optional; richer dependency audit at SessionStart. brew install osv-scanner or go install github.com/google/osv-scanner/cmd/osv-scanner@latest.

Example prompts:

/security-0day                         # scan diff vs main with the LLM 0-day scanner
/security-0day origin/release-1.4      # scan diff vs a different base ref
/security-review auth                  # 8-point check focused on auth surfaces
/security-assessment backend           # full OWASP/ASVS sweep, backend only
/threatmodel src/payments/             # STRIDE + DREAD threat model for the payments component

When the four skills overlap (and how to pick):

Are you reviewing a specific diff/PR/commit?
├── Yes → /security-0day
└── No → Did the change touch endpoints/auth/render/deps/config?
        ├── Yes → /security-review
        └── No → Pre-release / quarterly audit?
                ├── Yes → /security-assessment
                └── No → New feature / architecture design?
                        ├── Yes → /threatmodel
                        └── No → You probably don't need this suite right now.

Key files:

  • README.md — suite overview, install, decision tree, hook reference, subagent details, cross-skill integration diagram, troubleshooting.
  • install/install.sh — the one-command installer.
  • install/commands/ — the four slash command definitions (each is a thin wrapper over a SKILL).
  • install/hooks/settings.{lite,full}.example.json — ready-to-merge .claude/settings.json blocks.
  • install/windsurf/ — Windsurf rule + workflows.
  • install/codex/AGENTS.md.snippet — Codex behavioral instruction.
  • Security-automated-claude-skills/ — the canonical multi-language reviewer (skill + subagent + hooks + checklists + per-language reference packs + triage playbook). Also the on-disk fallback for /security-assessment and /security-0day.
  • Security-Analysis-Agent/ — parameterized backend/frontend tester templates with {{PLACEHOLDERS}} (hydrate before use).
  • _archive/Security-reviewr/ — the older single-file lite reviewer; superseded by the bundle, kept for recovery.

Plugins Reference

1. CTI Search Plugin

Folder: plugins/cti-search-plugin/

The execution engine behind CTI searches. Available as a CLI tool, MCP server, or slash command.

Usage Modes

Slash command:

/cti-search CVE-2024-21762
/cti-search LockBit ransomware --full --since 30
/cti-search ALPHV --notebooklm --tier 2

CLI:

node index.js --query "CVE-2024-21762" --full
node index.js --query "LockBit" --tier 2 --since 30 --notebooklm
node index.js --query "supply chain attack npm" --json

MCP tool (conversational):

Use the CTI search tool to find recent ransomware reports

Flags

Flag Description Default
--query <q> Search subject (required)
--count <n> Results per tier 10
--tier <1-4> Restrict to specific tier All
--since <days> Recency filter 90
--full Long-form brief with MITRE mapping Brief
--json Raw JSON output Formatted
--notebooklm Push sources to NotebookLM Disabled
--notebook-id <id> Override NotebookLM notebook From env

2. Secure PRD Plugin

Folder: plugins/secure-prd/

Generates security-focused Product Requirements Documents and integrates with external project management tools:

  • Atlassian Confluence — publishes PRD as a page
  • Linear / Asana — creates tasks from requirements
  • Slack — sends notifications on PRD completion
  • Notion — mirrors the PRD
  • Gmail — drafts stakeholder emails

Phoenix Pipeline — Feature Descriptor

Folder: feature-descriptor/

The Phoenix Pipeline is a 12-role specification system that breaks down feature requirements into discrete, expert-reviewed stages. Each role is a standalone .skill file that can be used independently or orchestrated together.

How It Works

Input (feature request / brief)
  │
  ├─→ Context Curator         — extract and cleanse context
  ├─→ Scope Cutter            — define in/out scope
  ├─→ Constraint Distiller    — identify constraints + acceptance criteria
  ├─→ Requirements Engineer   — RFC 2119 requirements with IDs
  ├─→ Ambiguity Hunter        — flag and resolve ambiguities
  ├─→ Security Engineer       — threat models + abuse cases
  ├─→ Contract Architect      — API design, events, errors
  ├─→ Verification Matrix     — proof paths for every requirement
  ├─→ Batch Planner           — incremental delivery plan
  ├─→ Final Gate              — go/no-go with blocker list
  │
  └─→ Output: production-ready PRD with security built in

The Pipeline Navigator orchestrates the flow, and the Orchestrator coordinates handoffs between roles.


Domain Tiers

The CTI search uses a four-tier domain system for intelligent query routing:

Tier Category Use Case Example Sources
T1 Authoritative / Government CVEs, advisories, official alerts CISA, NVD, MSRC, NCSC, Red Hat
T2 Vendor Research Deep technical analysis Unit42, Talos, Securelist, DFIR Report, Mandiant
T3 News / Community Situational awareness BleepingComputer, Krebs on Security, The Record, Hacker News
T4 OSINT / PoC Malware samples, exploits, indicators any.run, VulnCheck, AttackerKB, GreyNoise

Routing Logic

Query Type Primary Tier Secondary Tier
CVE lookups T1 (authoritative) T2 (vendor analysis)
Threat actors / malware T2 (research) T4 (OSINT)
News / situational T3 (news) T2 (context)
PoC / exploits T4 (technical) T2 (details)
General queries All tiers Ranked by authority

NotebookLM Integration

Push CTI research findings directly into Google NotebookLM for citation-backed analysis powered by Gemini.

Setup

  1. Install the notebooklm-connector plugin
  2. Get your notebook ID from the URL: https://notebooklm.google.com/notebook/<YOUR-ID>
  3. Set environment variable: export NOTEBOOKLM_NOTEBOOK_ID=your_id

Usage

/cti-search CVE-2024-21762 --notebooklm

The plugin searches, collects result URLs, pushes them as sources to your notebook, and reports the count.


Configuration & Customization

All skills and plugins are designed to be customized for your organization. No internal identifiers, names, or workspace details are hardcoded — you provide your own on first use.

Skill Customization

CTI Search — Customization

Setting How to Set Notes
Search API key .env file or environment variable Brave Search recommended (free tier)
NotebookLM notebook ID .env or --notebook-id flag Optional — for research ingestion
Custom domains Edit data/domains.txt + data/tier-map.json Add your own security sources
Default result count --count N flag Per-query override

Secure PRD — Customization

On first use, Claude will prompt you for:

Setting How to Set Notes
Owner name(s) Prompt, env var PRD_OWNER, or in-message Shown on all PRD outputs
Stakeholders Prompt, env var PRD_STAKEHOLDERS, or in-message Interested parties list
Confluence space key Prompt, env var PRD_CONFLUENCE_SPACE, or in-message Where PRD pages are created
Confluence template Prompt, env var PRD_CONFLUENCE_TEMPLATE, or in-message Parent page for nesting
Slack channel In-message or prompt Optional notifications
Email recipients In-message or prompt Optional Gmail drafts

You can set these persistently via environment variables:

export PRD_OWNER="@your-name"
export PRD_STAKEHOLDERS="@lead1, @lead2"
export PRD_CONFLUENCE_SPACE="ENG"
export PRD_CONFLUENCE_TEMPLATE="PRD Templates"

Or override per request: Write a PRD for X. Owner: @jane. Space: PRODUCT.


Required: Search API Key

Provider Free Tier Get Key
Brave Search (recommended) 2,000 requests/month api.search.brave.com
SerpAPI 100 requests/month serpapi.com

Environment Variables

Create a .env file in the plugin directory or export system-wide:

# Required
BRAVE_SEARCH_API_KEY=your_brave_api_key_here
SEARCH_PROVIDER=brave

# Optional — NotebookLM
NOTEBOOKLM_NOTEBOOK_ID=your_notebook_id_here

Verify Installation

# Check skill
ls ~/.claude/skills/cti-search-skill/

# Test plugin (dry run)
cd ~/.claude/plugins/cti-search-plugin
node index.js --query "CVE-2024-21762" --dry-run

Contributing

We welcome contributions from the global security community. Whether you're adding a new skill, improving an existing plugin, curating domains, or fixing docs — every contribution matters.

See the Contributing Guide for:

  • Step-by-step instructions to add new skills and plugins
  • Templates for SKILL.md, README.md, install.sh, and MCP servers
  • Documentation standards and testing checklists
  • Pull request process and code review expectations

Quick start:

# 1. Fork this repository
# 2. Clone your fork
git clone https://github.com/Security-Phoenix-demo/security-skills-claude-code.git
cd security-skills-claude-code

# 3. Create a feature branch
git checkout -b feature/your-new-skill

# 4. Follow the Contributing Guide
# 5. Submit a pull request

Frequently Asked Questions (FAQ)

General

What is Claude Code and why do I need it?

Claude Code is Anthropic's official CLI for AI-assisted software engineering. It lets you interact with Claude directly from your terminal. These skills and plugins extend Claude Code with security-specific capabilities — like searching 300+ threat intelligence sources or generating security-focused product requirements.

Is this free to use?

Yes. This repository is open source under the MIT License. You need a Claude Code subscription from Anthropic and a free API key from Brave Search (2,000 requests/month) or SerpAPI (100 requests/month).

What is the difference between a skill and a plugin?

Skills are instruction-based — they define how Claude should approach a task (workflow, reasoning, output format) but don't execute code. Plugins are executable — they run as MCP servers or CLI tools, make API calls, and return structured data. Skills often reference plugins for execution.

Can I use these skills with Cursor or other Claude-compatible editors?

Yes. Skills can be copied to ~/.cursor/skills/ for Cursor. Plugins that run as MCP servers work with any MCP-compatible client. Check your editor's documentation for MCP server configuration.

Installation & Setup

Which search API provider should I choose?

Brave Search is recommended — it offers 2,000 free requests per month and has excellent coverage of security sources. SerpAPI works but is limited to 100 free requests per month.

How do I install only specific skills or plugins?

Each skill and plugin is independent. Copy only the folders you need:

# Just the CTI skill
cp -r skills/cti-search-skill ~/.claude/skills/

# Just the PRD skill
cp -r skills/secure-prd-skill ~/.claude/skills/
Do I need Node.js?

Only for plugins. Skills are pure instructions and have no runtime dependencies. If you only use skills (no plugins), Node.js is not required.

The skill doesn't activate when I ask a question. What's wrong?
  1. Verify the skill is installed: ls ~/.claude/skills/cti-search-skill/SKILL.md
  2. Restart Claude Code after installation
  3. Try explicit trigger phrases: "Use the CTI domain research skill to search for..."
  4. See the Marketplace Install Troubleshooting for more solutions

CTI Search

How many security domains are included?

The curated domain list includes 595+ security sources across four tiers: government/authoritative sources (CISA, NVD, MSRC), vendor research labs (Unit42, Talos, Mandiant), security news (BleepingComputer, Krebs), and OSINT/PoC sources (GreyNoise, VulnCheck).

Can I add my own domains to the search?

Yes. Edit plugins/cti-search-plugin/data/domains.txt (one domain per line) and update data/tier-map.json with the tier and authority score. See the Contributing Guide for details.

What is the NotebookLM integration?

When you use the --notebooklm flag, the plugin pushes all result URLs as sources into a Google NotebookLM notebook. NotebookLM then provides citation-backed answers grounded only in those sources — dramatically reducing hallucination when analyzing CTI findings.

How does tier-based routing work?

The system classifies your query type (CVE, threat actor, malware, news, PoC) and routes it to the most relevant domain tier first. CVE queries hit authoritative government sources (T1) before vendor analysis (T2). Threat actor queries start with vendor research labs (T2). This ensures the most authoritative results surface first.

OpenGrep / SAST Rules

What is opengrep and how does it relate to semgrep?

OpenGrep is an open-source fork of semgrep focused on static application security testing (SAST). The rule syntax is fully compatible — rules generated by this skill work with both opengrep and semgrep. The skill generates YAML rule files that can be run with either tool to detect vulnerabilities in your codebase.

What languages does the OpenGrep Rule Generator support?

Over 30 languages including: Python, JavaScript, TypeScript, Java, Go, Ruby, PHP, C#, Rust, Kotlin, Swift, Scala, Terraform/HCL, Solidity, Bash, C, C++, Lua, OCaml, R, and more. Both search-mode (structural pattern matching) and taint-mode (data flow analysis) rules are supported for most languages.

What is the difference between the standard and research versions of OpenGrep Rule Generator?

The standard version (opengrep-rule-generator) generates rules from your description — use it when you know what pattern to detect. The research version (opengrep-rule-generator-research) adds a 4-phase vulnerability research pipeline that uses web search to study CVEs/CWEs, map attack surfaces, and find existing detection gaps before generating rules. Use the research version when you're starting from a CVE ID or vulnerability class rather than a known code pattern.

Can I generate rules for OWASP Top 10 in batch?

Yes. Ask Claude to "generate opengrep rules for OWASP Top 10 in [language]" and it will produce rules covering injection, broken auth, XSS, insecure deserialization, and other categories with appropriate CWE/OWASP metadata tags.

NotebookLM

What is the NotebookLM Connector skill?

It lets you query your Google NotebookLM notebooks directly from Claude Code. NotebookLM provides source-grounded, citation-backed answers from Gemini — meaning responses are based only on documents you've uploaded, with drastically reduced hallucination. This is especially powerful for security research where accuracy matters.

What do I need to use the NotebookLM skill?

You need: (1) Chrome or Edge browser running, (2) the "Claude in Chrome" extension installed and connected, (3) a Google account logged in to NotebookLM. Authentication is a one-time setup — see skills/notebooklm/AUTHENTICATION.md for the step-by-step guide.

How does NotebookLM integration work with CTI Search?

Two complementary workflows: The CTI Search plugin can push result URLs into a NotebookLM notebook using the --notebooklm flag. The NotebookLM Connector skill can then query that same notebook for citation-backed analysis of the findings. Together, they create a research-to-analysis pipeline: search → ingest → query.

Phoenix Pipeline & PRD

What is the Phoenix Pipeline?

It's a 12-role specification system that breaks feature requirements into expert-reviewed stages — from context extraction through scope definition, constraint analysis, security threat modeling, API design, verification matrices, and delivery planning. Each role is a standalone skill file in the feature-descriptor/ folder.

Can I use individual Phoenix Pipeline roles without the full pipeline?

Yes. Each .skill file in feature-descriptor/ is standalone. You can use just the Security Engineer role for threat modeling, or just the Ambiguity Hunter to review existing requirements.

Does the Secure PRD Generator integrate with project management tools?

Yes. Through MCP integrations, the PRD generator can publish to Atlassian Confluence, create tasks in Linear or Asana, send Slack notifications, mirror pages in Notion, and draft emails via Gmail. Configuration depends on which MCP servers you have connected.

Contributing

How can I contribute a new skill?

Fork the repo, create a skill directory under skills/, add a SKILL.md (with frontmatter), README.md, and install.sh, then submit a pull request. Full templates and standards are in the Contributing Guide.

I found a bug or want to request a feature. Where do I go?

Open a GitHub Issue for bugs or feature requests. For questions and ideas, use GitHub Discussions.

Can I use these skills in a commercial product?

Yes. The MIT License allows commercial use, modification, and distribution. See the LICENSE file.

Security & Privacy

Does this tool store or transmit my data?

No. All searches are executed through your configured search API (Brave or SerpAPI) with your own API key. No data is sent to Phoenix Security or any third party beyond your chosen search provider. NotebookLM integration is optional and uses your own Google account.

Is this safe to use for legitimate security research?

Yes. This toolkit is designed for defensive security, threat intelligence, and security engineering workflows. It queries public security sources and does not perform active scanning, exploitation, or any offensive operations.


Detailed Documentation

Document Description
README.md This file — overview, quick start, FAQ
MARKETPLACE_INSTALL.md Step-by-step marketplace installation with troubleshooting
CONTRIBUTING.md How to add skills, plugins, and domains
LICENSE MIT License
CTI Skill Docs CTI domain research skill specification
CTI Plugin Docs Plugin architecture, MCP server, CLI reference
Secure PRD Docs PRD generation skill specification
OpenGrep Rule Generator SAST rule generation skill + syntax reference
OpenGrep Research Vulnerability research + rule generation
NotebookLM Connector NotebookLM querying skill + authentication guide
Research Pipeline Global research pipeline documentation
Project Docs Skill Project documentation skill
Phoenix Pipeline 12-role feature specification pipeline

License

MIT License — see LICENSE for details.


Support


Security Note

This toolkit is designed for legitimate security research and threat intelligence gathering. Always:

  • Respect rate limits and terms of service for search APIs
  • Use responsibly and ethically
  • Follow responsible disclosure practices
  • Comply with applicable laws and regulations in your jurisdiction

Acknowledgments

Built and maintained by the engineering and security engineering teams at Phoenix Security for the global security community.

  • Open sourced for everyone to use, improve, and extend
  • Curated domain list includes 595+ trusted security sources
  • Designed for the Claude Code ecosystem and Claude Marketplace
  • Inspired by the need for efficient, structured, and repeatable CTI research workflows

Contributors are welcome from anywhere in the world. See CONTRIBUTING.md to get started.


Made with purpose by Phoenix Security
Security skills for the AI-native developer

About

This repository is a curated collection of skills, plugins, and automation pipelines designed for Claude Code — Anthropic's CLI for AI-assisted software engineering. It was built by the engineering and security engineering teams at Phoenix Security and released as open source so that security professionals, DevSecOps engineers, AppSec teams

Resources

License

Contributing

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors