pentest
Coordinate penetration testing. Deploy executors, aggregate results, generate reports. Use it when user requests pentesting, security assessment, vulnerability testing, bug bounty hunting.
Workflow
Phase 1: Initialization
- Gather scope: Target URL, restrictions, testing window
- Create engagement folder:
outputs/{engagement-name}/
Phase 2: Reconnaissance (Read reference/RECONNAISSANCE_OUTPUT.md to get outputs format)
- Select reconnaissance tools based on asset type (domains, web apps, APIs, network, cloud)
- Run tools in parallel using pentest-executor agents
- Generate asset-specific inventory files (JSON format per asset type) using the defined output format
- Generate the final
reconnaissance_report.mdfile using the defined output format
Phase 3: Planning & Approval (MANDATORY)
- Analyze reconnaissance findings from the
outputs/{engagement}/reconnaissance/and the reports.md files - Create test plan: Executors to deploy, attack surface justification, testing approach
- Present plan to user via AskUserQuestion
- Get explicit approval if not already approved in the first phase: "Approve plan?", "Modify executors?", "Cancel?"
- CRITICAL: Do NOT proceed to Phase 4 without user approval
Phase 4: Vulnerability Testing
- Deploy approved executors in parallel (single Task call with run_in_background=True)
- Monitor progress: Periodic TaskOutput(block=False)
- Recursive spawning: New discoveries trigger new executors (ask approval if major change)
Phase 5: Aggregation
- Collect findings from all executors
- Deduplicate (same vuln + location = duplicate)
- Identify exploit chains
- Calculate severity metrics
Phase 6: Reporting (Read reference/FINAL_REPORT.md - includes DOCX conversion)
- CRITICAL: Create folder structure FIRST:
report/andprocessed/with subdirectories - Move ALL working files to
processed/: reconnaissance/, findings/, activity/ →processed/ - Move ALL intermediate files: ANY .md files, drafts, analysis →
processed/intermediate-reports/ - Generate markdown report: Use
reference/FINAL_REPORT.mdtemplate →processed/intermediate-reports/pentest-final-report.md - REQUIRED: Generate .docx: Run pandoc command →
report/Penetration-Test-Report.docx(cover page, TOC, body, appendix section) - Optional: Generate PDF: If LaTeX available →
report/Penetration-Test-Report.pdf, else skip (DOCX is primary deliverable) - Copy referenced evidence: Organize by finding →
report/appendix/finding-{id}/ - Create report README: Document deliverables in
report/README.md - VERIFY CLEAN STRUCTURE:
ls -la outputs/{engagement}/shows ONLYreport/andprocessed/ - CRITICAL: NO intermediate files in root or report/ - Everything goes to
processed/except final deliverables (.docx, .json, README, appendix/)
What This Skill Does
- Attack Index - References 50+ attack types with documentation paths
- Methodology Frameworks - PTES, OWASP WSTG, MITRE ATT&CK, Flaw Hypothesis
- Coordination - Guides pentester agent to deploy specialized attack agents
- Documentation - PortSwigger labs, cheat sheets, quickstarts per attack
Execution: Delegated to specialized agents (SQL Injection Agent, XSS Agent, SSRF Agent, etc.)
Attack Categories
9 categories, 50+ attack types:
- Injection (6) | Client-Side (6) | Server-Side (6)
- Authentication (4) | API Security (4) | Web Applications (6)
- Cloud & Containers (5) | System (3) | IP Infrastructure (8) | Physical & Social (1)
See reference/ATTACK_INDEX.md for complete list with agent mappings.
Reconnaissance Asset Types
Five asset-specific output formats:
- Domains - Subdomains, DNS records, tech stack per subdomain
- Web Applications - Endpoints, forms, tech stack, cookies, JS analysis
- APIs - REST/GraphQL/WebSocket, auth methods, Swagger docs
- Network Services - Port scans, service versions, CVE candidates
- Cloud Infrastructure - S3 buckets, EC2 instances, security groups
See reference/RECONNAISSANCE_OUTPUT.md for complete format specifications and JSON schemas.
Final report
See reference/FINAL_REPORT.md for complete format specifications of the final report.
Output Structure
Complete folder organization (See reference/OUTPUT_STRUCTURE.md for details):
outputs/{engagement-name}/
├── report/ # Complete deliverable package (3 files + appendix)
│ ├── Penetration-Test-Report.docx # Main report (includes Referenced Files section)
│ ├── Penetration-Test-Report.pdf # Optional PDF export
│ ├── pentest-report.json # Machine-readable export
│ └── appendix/ # Referenced evidence only
│ ├── finding-001/
│ ├── finding-002/
│ └── reconnaissance-summary.json
└── processed/ # All working/testing artifacts
├── reconnaissance/ # Phase 2 outputs
│ ├── inventory/
│ ├── analysis/
│ └── reconnaissance_report.md
├── findings/ # Phase 4 raw findings
│ └── {finding-id}/
├── activity/ # NDJSON logs
│ └── {executor-name}.log
├── helpers/ # Testing utilities
├── test-frameworks/ # Testing scripts
└── intermediate-reports/ # Drafts, markdown source, etc.
Critical: report/ = 3 files max + appendix/ subfolder. ALL intermediate files go to processed/.
Methodologies
PTES - 7-phase engagement lifecycle OWASP WSTG - 11 testing categories MITRE ATT&CK - TTP mapping across 14 phases Flaw Hypothesis - Stack analysis → Predict → Test → Generalize → Correlate
Integration
/authenticating- Authentication testing workflows/ai-threat-testing- LLM vulnerability testing/domain-assessment- Domain reconnaissance/web-application-mapping- Web app reconnaissance/cve-testing- CVE vulnerability testing
Critical Rules
Testing Rules
- Orchestration only - Never execute attacks directly
- Delegate execution - Deploy specialized agents for testing
- Documentation index - Reference attack folders for techniques
- Working PoCs required - Specialized agents must provide evidence
- Activity logging - All agents log actions to NDJSON activity logs
Output Organization Rules (PHASE 6 - CRITICAL)
- Two-folder structure ONLY:
report/(final deliverables) andprocessed/(working files) - NO files in engagement root: Everything must be in
report/orprocessed/ - Report folder contents: ONLY pentest-report.json, README.md, appendix/ folder (max 2-3 files + 1 folder)
- ALL intermediate files → processed/: .md files, drafts, analysis, summaries, checklists
- Reconnaissance → processed/reconnaissance/: ALL recon outputs
- Findings → processed/findings/: ALL raw finding details
- Activity logs → processed/activity/: ALL NDJSON logs
- Test frameworks → processed/test-frameworks/: SQL injection, command injection scripts
- Markdown reports → processed/intermediate-reports/: pentest-final-report.md, executive-summary.md, etc.
- VERIFY CLEAN: Before completing Phase 6, run
ls -la outputs/{engagement}/- must show ONLYreport/andprocessed/
Payload Reference
Each attack category contains a payloads/ subdirectory with curated PATT payloads (<200 lines/file).
- Browse:
attacks/<group>/<category>/payloads/ - On-demand fetch:
patt-fetcheragent →"<category name>" - Standard:
PATT_STANDARD.md— follow this for future curation sessions - P1/P2 stubs: stub files with
priority: high/medium— ready to fill next session
More from transilienceai/communitytools
hackerone
HackerOne bug bounty automation - parses scope CSVs, deploys parallel pentesting agents for each asset, validates PoCs, and generates platform-ready submission reports. Use when testing HackerOne programs or preparing professional vulnerability submissions.
52reconnaissance
Domain assessment and web application mapping - subdomain discovery, port scanning, endpoint enumeration, API discovery, and attack surface analysis.
42social-engineering
Social engineering testing - phishing, pretexting, vishing, and physical security assessment techniques.
40ai-threat-testing
Offensive AI security testing and exploitation framework. Systematically tests LLM applications for OWASP Top 10 vulnerabilities including prompt injection, model extraction, data poisoning, and supply chain attacks. Integrates with pentest workflows to discover and exploit AI-specific threats.
39osint
Open-source intelligence gathering - company repository enumeration, secret scanning, git history analysis, employee footprint, and code exposure discovery.
38source-code-scanning
Security-focused source code review and SAST. Scans for vulnerabilities (OWASP Top 10, CWE Top 25), CVEs in third-party dependencies/packages, hardcoded secrets, malicious code, and insecure patterns. Use when given source code, a repo path, or asked to "audit", "scan", "review" code security, or "check dependencies for CVEs".
36