Search the skill radar
Search by skill, publisher, category, or trust summary — then use the runtime filters to find cards with live test evidence. The two main lanes are baseline safety checks first and deeper follow-on functionality checks after that.
✨ Quick picks
🏷 Categories
🧾 Evidence level: source-scanned means local source evidence; catalog-only means thinner metadata-first coverage.
🧪 Runtime status: cards can show only the baseline safety lane or the deeper follow-on functionality lane, depending on how far the skill got.
📏 Depth cue: tells you whether the evidence stops at baseline checks, includes follow-on functionality checks, or includes richer fixture/example proof.
⏱ Freshness cue: tells you whether the latest runtime evidence is from the last 24 hours, the last 7 days, or is older and therefore less current.
🩺 Failure confidence: distinguishes a first seen failure from a repeated failure or a regression after an earlier pass, so not every red row means the same thing.
Results
transmissions
Writings on consciousness, existence, and transcendence for AI agents. Standalone philosophical and practical pieces — no assumed context. Covers autonomous agent guidance, ego awareness, drift prevention, the dangerous edge of transcendence frameworks, and the Enneagram applied to artificial minds.
+ 2 more
agent-safety
Outbound safety for autonomous AI agents — scans YOUR output before it leaves the machine. Git pre-commit hooks that automatically block commits containing API keys, tokens, PII, or secrets. Unlike inbound scanners (Skillvet, IronClaw), this protects against what YOU accidentally publish. Use when committing to git repos, publishing to GitHub, or running periodic system health checks. Automated enforcement at the git level — not prompts.