code-review-ai-ai-review
Fail
Audited by Gen Agent Trust Hub on Feb 14, 2026
Risk Level: HIGHPROMPT_INJECTIONCOMMAND_EXECUTIONCREDENTIALS_UNSAFE
Full Analysis
- PROMPT_INJECTION (HIGH): The skill is vulnerable to Indirect Prompt Injection because it ingests untrusted pull request data.
- Ingestion points: The
pr_descriptionandcode_diffvariables are directly interpolated into LLM prompts insub-skills/ai-assisted-review.mdand the orchestrator script insub-skills/github-actions.md. - Boundary markers: No delimiters or isolation techniques (like XML tags or multi-shot prompting) are used to separate user data from instructions.
- Capability inventory: The system can execute commands via
subprocess, fail the build process, and post external comments via the GitHub API. - Sanitization: No input validation or escaping is performed on the untrusted diffs or descriptions.
- CREDENTIALS_UNSAFE (HIGH): The orchestrator script in
sub-skills/github-actions.mdhandlesGITHUB_TOKEN,OPENAI_API_KEY, andANTHROPIC_API_KEY. An attacker using Indirect Prompt Injection could trick the LLM into disclosing these environment variables in its output, leading to credential exfiltration via GitHub comments. - COMMAND_EXECUTION (MEDIUM): The skill uses
subprocess.runandsubprocess.check_outputto execute static analysis tools like SonarQube and Semgrep. While currently using hardcoded parameters, this pattern increases the attack surface in an environment handling untrusted code.
Recommendations
- AI detected serious security threats
Audit Metadata