Artwork

Indhold leveret af IVANCAST PODCAST. Alt podcastindhold inklusive episoder, grafik og podcastbeskrivelser uploades og leveres direkte af IVANCAST PODCAST eller deres podcastplatformspartner. Hvis du mener, at nogen bruger dit ophavsretligt beskyttede værk uden din tilladelse, kan du følge processen beskrevet her https://da.player.fm/legal.
Player FM - Podcast-app
Gå offline med appen Player FM !

Peering Into the Black Box: The Rise of Representation Engineering

37:27
 
Del
 

Manage episode 448992995 series 3351512
Indhold leveret af IVANCAST PODCAST. Alt podcastindhold inklusive episoder, grafik og podcastbeskrivelser uploades og leveres direkte af IVANCAST PODCAST eller deres podcastplatformspartner. Hvis du mener, at nogen bruger dit ophavsretligt beskyttede værk uden din tilladelse, kan du følge processen beskrevet her https://da.player.fm/legal.

Join us in SHIFTERLABS’ latest experimental podcast series powered by Notebook LM, where we bridge research and conversation to illuminate groundbreaking ideas in AI. In this episode, we dive into “Representation Engineering: A Top-Down Approach to AI Transparency,” an insightful paper from the Center for AI Safety, Carnegie Mellon University, Stanford, and other leading institutions. This research redefines how we view transparency in deep learning by shifting the focus from neurons and circuits to high-level representations.

Discover how Representation Engineering (RepE) introduces new methods for reading and controlling cognitive processes in AI models, offering innovative solutions to challenges like honesty, hallucination detection, and fairness. We explore its applications across essential safety domains, including model control and ethical behavior. Tune in to learn how these advances could shape a future of AI that is more transparent, accountable, and aligned with human values.

This series is part of SHIFTERLABS’ ongoing commitment to pushing the boundaries of educational technology and fostering discussions at the intersection of research, technology, and responsible innovation.

  continue reading

100 episoder

Artwork
iconDel
 
Manage episode 448992995 series 3351512
Indhold leveret af IVANCAST PODCAST. Alt podcastindhold inklusive episoder, grafik og podcastbeskrivelser uploades og leveres direkte af IVANCAST PODCAST eller deres podcastplatformspartner. Hvis du mener, at nogen bruger dit ophavsretligt beskyttede værk uden din tilladelse, kan du følge processen beskrevet her https://da.player.fm/legal.

Join us in SHIFTERLABS’ latest experimental podcast series powered by Notebook LM, where we bridge research and conversation to illuminate groundbreaking ideas in AI. In this episode, we dive into “Representation Engineering: A Top-Down Approach to AI Transparency,” an insightful paper from the Center for AI Safety, Carnegie Mellon University, Stanford, and other leading institutions. This research redefines how we view transparency in deep learning by shifting the focus from neurons and circuits to high-level representations.

Discover how Representation Engineering (RepE) introduces new methods for reading and controlling cognitive processes in AI models, offering innovative solutions to challenges like honesty, hallucination detection, and fairness. We explore its applications across essential safety domains, including model control and ethical behavior. Tune in to learn how these advances could shape a future of AI that is more transparent, accountable, and aligned with human values.

This series is part of SHIFTERLABS’ ongoing commitment to pushing the boundaries of educational technology and fostering discussions at the intersection of research, technology, and responsible innovation.

  continue reading

100 episoder

모든 에피소드

×
 
Loading …

Velkommen til Player FM!

Player FM is scanning the web for high-quality podcasts for you to enjoy right now. It's the best podcast app and works on Android, iPhone, and the web. Signup to sync subscriptions across devices.

 

Hurtig referencevejledning

Lyt til dette show, mens du udforsker
Afspil