Episode publish date
October 4, 2025 6:55 AM (UTC)
Last edit date
Oct 7, 2025 2:03 PM
Last snip date
October 4, 2025 6:44 PM (GMT+1)
Last sync date
October 7, 2025 3:03 PM (GMT+1)
Show
Machine Learning Street Talk (MLST)
Snips
12
Warning
⚠️ Any content within the episode information, snip blocks might be updated or overwritten by Snipd in a future sync. Add your edits or additional notes outside these blocks to keep them safe.
‣
Your snips
‣
[00:00] Big Models Break In New Ways
‣
[00:34] Agents Are A Worst-Case Adversary
‣
[02:28] Use Trusted Models For Small Verified Computation
‣
[16:09] Enforce Policies By Design With Symbolic Variables
‣
[19:20] Human-Centric Security Assumptions Break
‣
[22:51] Agent Sent Unexpected Emails
‣
[27:24] Thinking Traces Aren't Security Proof
‣
[30:48] Sandbox Remote Models And Tooling
‣
[37:20] Model Structure Can Hide Backdoors
‣
[52:03] Recursive Training Shrinks Diversity
‣
[53:17] Preserve Diversity And Monitor Drift
‣
