adk-eval-guide
Warn
Audited by Gen Agent Trust Hub on Mar 5, 2026
Risk Level: MEDIUMPROMPT_INJECTIONCOMMAND_EXECUTION
Full Analysis
- [PROMPT_INJECTION]: The skill metadata contains misleading information regarding its origin. The frontmatter in
SKILL.mdidentifies the author as 'Google', which contradicts the actual skill authorship ('eliasecchig'). This is a form of metadata poisoning that can lead users to over-trust the skill's instructions and security posture. - [PROMPT_INJECTION]: A code example in
references/multimodal-eval.mdfor a custom evaluation metric creates a significant surface for indirect prompt injection. - Ingestion points: Untrusted agent output is collected from
actual.final_response.partsand interpolated directly into a prompt for a judge model. - Boundary markers: The implementation uses only basic newline delimiters (
\n\nAgent response:) which do not reliably isolate the untrusted content from the judge's instructions. - Capability inventory: The script utilizes the
google-genaiSDK (client.aio.models.generate_content) to execute model calls based on the poisoned prompt. - Sanitization: The code example lacks any form of input sanitization, escaping, or validation for the agent-generated text before it is processed by the judge.
- [COMMAND_EXECUTION]: The skill instructions and examples promote the execution of various CLI commands and the dynamic loading of Python code.
- Evidence:
SKILL.mdandreferences/criteria-guide.mddescribe running evaluations viaadk evalandmake eval, and explain how to configurecustom_metricsintest_config.jsonby providing Python module paths that are loaded and executed at runtime.
Audit Metadata