llm-application-dev-langchain-agent
Pass
Audited by Gen Agent Trust Hub on Feb 17, 2026
Risk Level: SAFE
Full Analysis
- SAFE (SAFE): No malicious patterns or security risks detected. The skill functions as a domain-specific instruction set for an LLM to act as a coding assistant.
- The skill consists entirely of Markdown documentation and Python code templates.
- It explicitly recommends security best practices, such as using environment variables for secrets instead of hardcoding them.
- All mentioned dependencies are standard, reputable libraries in the AI development ecosystem (e.g., LangChain, FastAPI, Pydantic).
Audit Metadata