llm-architect
Warn
Audited by Gen Agent Trust Hub on Feb 17, 2026
Risk Level: MEDIUMREMOTE_CODE_EXECUTIONPROMPT_INJECTIONEXTERNAL_DOWNLOADS
Full Analysis
- [REMOTE_CODE_EXECUTION] (MEDIUM): The script
scripts/finetune_model.pyuses thetrust_remote_code=Trueparameter when loading models and tokenizers via thetransformerslibrary.\n - Evidence: Lines 68 and 80 in
scripts/finetune_model.pyand line 57 inscripts/serve_model.py.\n - Risk: This setting permits the execution of arbitrary Python code provided by the model author. If a user or the agent is tricked into loading a malicious model from an untrusted source, it can lead to full host compromise.\n- [PROMPT_INJECTION] (LOW): The skill is susceptible to indirect prompt injection due to its handling of untrusted external data in RAG and evaluation workflows.\n
- Evidence Chain:\n
- Ingestion points:
scripts/setup_rag_pipeline.py(viaadd_documents) andscripts/evaluate_model.py(viaevaluate_model).\n - Boundary markers: Minimal markers (e.g.,
Context: {context}) are used in prompt templates inscripts/setup_rag_pipeline.py.\n - Capability inventory: The skill has high-privilege capabilities including starting a web server (
scripts/serve_model.py), writing files (export_results), and training models.\n - Sanitization: No evidence of sanitization or safety filtering for external content before it is embedded into prompts or vector stores.\n- [EXTERNAL_DOWNLOADS] (SAFE): The skill relies on standard, well-known machine learning and web framework libraries.\n
- Evidence: Imports of
transformers,torch,fastapi, andchromadbacross several scripts.\n - Risk: These are standard dependencies for the skill's purpose and are installed from trusted registries like PyPI.
Audit Metadata