high-performance-inference
Fail
Audited by Gen Agent Trust Hub on Feb 16, 2026
Risk Level: HIGHREMOTE_CODE_EXECUTIONCOMMAND_EXECUTIONEXTERNAL_DOWNLOADSPROMPT_INJECTION
Full Analysis
- Unverifiable Dependencies & Remote Code Execution (HIGH): The script
scripts/vllm-server.py(line 120) configures thevllm.LLMengine withtrust_remote_code=True. This setting permits arbitrary Python code bundled with HuggingFace models to execute on the host system. This is an RCE vector if a user specifies a malicious model identifier via theMODEL_NAMEenvironment variable. - Privilege Escalation (HIGH): The file
references/edge-deployment.mdprovides instructions for runningsudocommands (sudo nvpmodel,sudo jetson_clocks) to manage hardware performance. Promoting system-level modifications with elevated privileges is a high-risk practice. - Unverifiable Dependencies & Remote Code Execution (MEDIUM): The documentation in
references/edge-deployment.mdencourages downloading software from unverified sources usinggit clone https://github.com/ggerganov/llama.cpp, which is not on the list of trusted repositories. - Indirect Prompt Injection (HIGH): The skill presents a high-risk attack surface for indirect prompt injection.
- Ingestion points: Model identifiers and configurations provided through environment variables in
scripts/vllm-server.py, and external datasets used inreferences/quantization-guide.md. - Boundary markers: No boundary markers or instructions to ignore embedded commands are present in the processing logic.
- Capability inventory: Includes arbitrary command execution via
subprocess.Popen, file system modifications through model saving, and remote code execution via thetrust_remote_codeflag. - Sanitization: There is no evidence of sanitization or origin validation for the models or datasets ingested by the skill.
Recommendations
- AI detected serious security threats
Audit Metadata