Library
29 Aug 2024
1h 10m

Why you should write your own LLM benchmarks — with Nicholas Carlini, Google DeepMind

Podcast cover

Latent Space: The AI Engineer Podcast

Sign in to access all AI-generated content

Continue
This podcast episode features Nick Carlini, a research scientist at DeepMind, who delves into the complexities of adversarial AI security and the balanced use of large language models (LLMs). Carlini combines a playful exploratory approach with rigorous analysis, promoting the understanding of AI's dual capabilities and pitfalls. He advocates for a grounded outlook on LLMs, encouraging their utilization as practical tools while remaining vigilant about their limitations, particularly in terms of security vulnerabilities. Throughout the discussion, Carlini underscores the critical need for tailored evaluation benchmarks and the exploration of AI's dark side to effectively navigate and enhance the safety of emerging technologies.
Takeaways
Outlines
Q & A
 
mindmap screenshot
Preview
preview episode cover
How to Get Rich: Every EpisodeNaval