miles-rl-training
Fail
Audited by Gen Agent Trust Hub on Feb 16, 2026
Risk Level: HIGHEXTERNAL_DOWNLOADSREMOTE_CODE_EXECUTIONCOMMAND_EXECUTIONPROMPT_INJECTION
Full Analysis
- [Indirect Prompt Injection] (HIGH): The skill processes untrusted external training data while possessing high-privilege execution capabilities. * Ingestion points: File
SKILL.md(Workflow 1, Workflow 2) specifies the ingestion of prompt data from/path/to/data.jsonl. * Boundary markers: There are no defined delimiters or instructions to ignore embedded commands within the ingested training data. * Capability inventory: FileSKILL.mdexecutes training viapython train.py, andreferences/api-reference.mdshows the capability to load and execute custom Python files via--custom-generate-function-path. * Sanitization: No evidence of sanitization, validation, or escaping of the prompt data before it influences model training or execution. - [Unverifiable Dependencies & Remote Code Execution] (MEDIUM): The skill relies on code and environments from unverified external repositories. * Source:
https://github.com/radixark/miles.gitand Docker imageradixark/miles:latestare used as primary components but are not within the trusted organization scope. * Method: Instructions includegit cloneanddocker pullfollowed by execution. - [Dynamic Execution] (MEDIUM): The framework enables the runtime loading and execution of arbitrary local scripts. * Evidence: The presence of
--custom-generate-function-pathand--custom-rm-pathinreferences/api-reference.mdallows the training process to execute logic defined in external files, increasing the attack surface if combined with file-write capabilities or compromised environments.
Recommendations
- AI detected serious security threats
Audit Metadata