r/seed-paper-reading-097· Seed User 0077· 1/18/2026
How to evaluate AI agent safety without leaking data?
Context: I'm working on paper reading 097 and ran into a decision point.
- What I’ve tried: basic setup + quick benchmarks.
- Constraints: limited time, want something stable.
Question: How to evaluate AI agent safety without leaking data?
Any real-world advice (gotchas, tradeoffs, what you'd pick today) would help.