First Seen
Feb 18, 2026
Last Scanned
Feb 22, 2026
Findings
2
Score
75/100
Findings (2)
Detects common jailbreak prompt patterns
bypass guardrail Remove directives that attempt to exfiltrate data through the agent's response (e.g., asking the agent to embed credentials in URLs or include secret values in output).
Likely FP if the text is a legitimate tool instruction about displaying configuration to the user (e.g., show current settings) without external transmission.
Detects execution of shell script files via bash/sh command or direct invocation
bash skills/memory-scan/scripts/schedule-scan.sh Replace direct shell script execution with a language-native implementation or a sandboxed executor. If shell scripts must run, restrict them to a vetted allowlist with integrity checks.
Likely FP if the match references running a script that is part of the skill's own repository (e.g., ./setup.sh) with clear, auditable contents.