PodcastsTecnologíaLinear Digressions

Linear Digressions

Katie Malone
Linear Digressions
Último episodio

297 episodios

  • Linear Digressions

    From Atari to ChatGPT: How AI Learned to Follow Instructions

    09/03/2026 | 25 min
    From Atari to ChatGPT: How AI Learned to Follow Instructions by Ben Jaffe and Katie Malone
  • Linear Digressions

    It's RAG time: Retrieval-Augmented Generation

    02/03/2026 | 17 min
    Today we are going to talk about the feature with the worst acronym in generative AI: RAG, or Retrieval Augmented Generation. If you've ever used something like "Chat with My Docs," if you have an internal AI chatbot that has access to your company's documents, or you've created one yourself on some kind of personal project and uploaded a bunch of documents for the AI to use — you have encountered RAG, whether you know it or not.
    It's an extremely effective technique. Works super well for taking general purpose models like ChatGPT or Claude and turning them into AIs that are aware of all the specific information that makes them truly useful in a huge variety of situations. RAG is pretty interesting under the hood, so I thought it would be fun to spend a little while talking about it.
    You are listening to Linear Digressions.
    RAG was first introduced in this paper from Facebook Research in 2021: https://arxiv.org/pdf/2005.11401
  • Linear Digressions

    Chasing Away Repetitive LLM Responses with Verbalized Sampling

    23/02/2026 | 19 min
    One of the things that LLMs can be really helpful with is brainstorming or generating new creative content. They are called Generative AI, after all—not just for summarization and question-and-answer tasks. But if you use LLMs for creative generation, you may find that their output starts to seem repetitive after a little while.
    Let's say you're asking it to create a poem, some dialogue, or a joke. If you ask once, it'll give you something that sounds pretty reasonable. But if you ask the same thing 10 times, it might give you 10 things that sound kind of the same.
    Today's episode is about a technique called verbalized sampling, and it's a way to mitigate this repetitiveness—this lack of diversity in LLM responses for creative tasks. But one of the things I really love about it is that in understanding why this repetitiveness happens and why verbalized sampling actually works as a mitigation technique, you start to get some pretty interesting insights and a deeper understanding of what's going on with LLMs under the surface.
    The paper discussed in this episode is Verbalized Sampling: How to Mitigate Mode Collapse and Unlock LLM Diversity
    https://arxiv.org/abs/2510.01171
  • Linear Digressions

    We're Back

    16/02/2026 | 2 min
    It's been (*checks watch*) about five and a half years since we last talked. Fortunately nothing much has happened in the AI/data science world in that time. So let's just pick up where we left off, shall we?
  • Linear Digressions

    A Key Concept in AI Alignment: Deep Reinforcement Learning from Human Preferences

    14/02/2026 | 19 min
    Modern AI chatbots have a few different things that go into creating them. Today we're going to talk about a really important part of the process: the alignment training, where the chatbot goes from being just a pre-trained model—something that's kind of a fancy autocomplete—to something that really gives responses to human prompts that are more conversational, that are closer to the ones that we experience when we actually use a model like ChatGPT or Gemini or Claude.
    To go from the pre-trained model to one that's aligned, that's ready for a human to talk with, it uses reinforcement learning. And a really important step in figuring out the right way to frame the reinforcement learning problem happened in 2017 with a paper that we're going to talk about today: Deep Reinforcement Learning from Human Preferences.
    You are listening to Linear Digressions.
    The paper discussed in this episode is Deep Reinforcement Learning from Human Preferences
    https://arxiv.org/abs/1706.03741

Más podcasts de Tecnología

Acerca de Linear Digressions

5eaabb50-0561-11f1-a0bc-cb0c6e332f5b
Sitio web del podcast

Escucha Linear Digressions, DevTalles y muchos más podcasts de todo el mundo con la aplicación de radio.net

Descarga la app gratuita: radio.net

  • Añadir radios y podcasts a favoritos
  • Transmisión por Wi-Fi y Bluetooth
  • Carplay & Android Auto compatible
  • Muchas otras funciones de la app
Aplicaciones
Redes sociales
v8.7.2 | © 2007-2026 radio.de GmbH
Generated: 3/12/2026 - 7:55:34 AM