using-llm-specialist
Pass
Audited by Gen Agent Trust Hub on Mar 1, 2026
Risk Level: SAFE
Full Analysis
- [SAFE]: The skill serves as a documentation router and educational resource for Large Language Model (LLM) engineering, providing high-quality technical guidance.
- [SAFE]: All code blocks (Python, Shell, JSON, CSV) are provided as illustrative examples for users and developers; the skill itself does not contain executable scripts that would perform unauthorized actions.
- [SAFE]: The inclusion of adversarial patterns, such as jailbreak prompts and prompt injection examples, is strictly for defensive education and the development of safety filters, aligning with the skill's stated purpose.
- [SAFE]: No hardcoded credentials, sensitive file path access, or unauthorized network operations were identified. The external libraries referenced (e.g., openai, tiktoken, transformers) are standard industry tools for AI development.
- [SAFE]: The skill provides robust guidance on data privacy and security, including PII (Personally Identifiable Information) redaction and content moderation strategies.
Audit Metadata