Zenity Labs
Research, tools and talks about building and breaking copilots and no-code apps
Connect
How adding a single safety labeled tool to an LLM's toolset can sharply increase its defense
0-click indirect prompt injection with tool use - a look through attribution graphs
How using structured prompts present findings of self-modeling in LLMs, which may benefit both attackers and defenders
How controlling the structure of the prompt, not just the semantics, can exploit your AI agents and their tools
A deep dive into OpenAI's AgentKit guardrails, how they are implemented, and where they fail
Humans, hacker culture and AI: Notes from Hacker Summer Camp