Daniel Filan's profile photo

Daniel Filan

Featured in:

Articles

  • Jan 24, 2025 | lesswrong.com | Daniel Filan

    YouTube linkTypically this podcast talks about how to avert destruction from AI. But what would it take to ensure AI promotes human flourishing as well as it can? Is alignment to individuals enough, and if not, where do we go form here? In this episode, I talk with Joel Lehman about these questions. Topics we discuss: Positive visions of AI Improving recommendation systemsDaniel Filan (00:09):Hello, everyone.

  • Jan 19, 2025 | lesswrong.com | Daniel Filan

    YouTube linkSuppose we’re worried about AIs engaging in long-term plans that they don’t tell us about. If we were to peek inside their brains, what should we look for to check whether this was happening? In this episode Adrià Garriga-Alonso talks about his work trying to answer this question. Topics we discuss: The Alignment Workshop Daniel Filan (00:09):Hello, everyone.

  • Jan 9, 2025 | lesswrong.com | Daniel Filan |Ryan Kidd

    MATS currently has more people interested in being mentors than we are able to support—for example, for the Winter 2024-25 Program, we received applications from 87 prospective mentors who cumulatively asked for 223 scholars (for a cohort where we expected to only accept 80 scholars). As a result, we need some process for how to choose which researchers to take on as mentors and how many scholars to allocate each.

  • Nov 27, 2024 | lesswrong.com | Daniel Filan

    YouTube linkYou may have heard of singular learning theory, and its “local learning coefficient”, or LLC - but have you heard of the refined LLC? In this episode, I chat with Jesse Hoogland about his work on SLT, and using the refined LLC to find a new circuit in language models. Topics we discuss: About Jesse The Alignment Workshop About Timaeus SLT that isn’t developmental interpretability The refined local learning coefficient Daniel Filan (00:09):Hello, everyone.

  • Nov 16, 2024 | lesswrong.com | Daniel Filan

    YouTube linkRoad lines, street lights, and licence plates are examples of infrastructure used to ensure that roads operate smoothly. In this episode, Alan Chan talks about using similar interventions to help avoid bad outcomes from the deployment of AI agents. Topics we discuss: How the Alignment Workshop is Agent infrastructure Why agent infrastructure A trichotomy of agent infrastructure Agent IDs Agent channels Relation to AI controlDaniel Filan (00:09):Hello, everyone.

Contact details

Socials & Sites

Try JournoFinder For Free

Search and contact over 1M+ journalist profiles, browse 100M+ articles, and unlock powerful PR tools.

Start Your 7-Day Free Trial →