Jonas Hallgren's profile photo

Jonas Hallgren

Featured in: Favicon bmj.com Favicon jech.bmj.com

Articles

  • Jan 10, 2025 | lesswrong.com | Alvin Ånestrand |Jonas Hallgren

    The Alignment Mapping Program: Forging Independent Thinkers in AI Safety - A Pilot RetrospectiveThe AI safety field faces a critical challenge: we need researchers who can not only implement existing solutions but also forge new, independent paths. In 2023, inspired by John Wentworth's work on agency and learning from researchers like Rohin Shah and Adam Shimi who have highlighted the limitations of standard AI safety education, we launched the Alignment Mapping Program (AMP).

  • Jan 7, 2025 | lesswrong.com | Jonas Hallgren

    Okay, so after engaging a lot with Steven Byrnes post on [Intuitive self-models] 6. Awakening / Enlightenment / PNSE I’ve tried to think about the most important framing shift that is different from me compared to what Steven has written. We will be going through a generative model argument from an agency and predictive processing perspective. This will then lead to an understanding of meditation insights and “awakening” as changes in the self-perceived information boundary.

  • Dec 5, 2024 | lesswrong.com | Jonas Hallgren |Seth Herd

    EVERYONE, CALM DOWN!Meaning Alignment Institute just dropped their first post in basically a year and it seems like they've been up to some cool stuff. Their perspective on value alignment really grabbed my attention because it reframes our usual technical alignment conversations around rules and reward functions into something more fundamental - what makes humans actually reliably good and cooperative?

  • Dec 5, 2024 | lesswrong.com | Jonas Hallgren

    EVERYONE, CALM DOWN! Meaning Alignment Institute just dropped their first post in basically a year and it seems like they've been up to some cool stuff. Their perspective on value alignment really grabbed my attention because it reframes our usual technical alignment conversations around rules and reward functions into something more fundamental - what makes humans actually reliably good and cooperative?

  • Apr 2, 2024 | lesswrong.com | Jonas Hallgren

    This post wouldn't have happened without this post, for which I'm forever gratefulIn order to be able to steer away asteroids from earth we first have to be able to model them properly. This is a post on how to do this safely. Coincidentally, this would be perfect technology for Open Asteroid Impact to improve its direction models, yet as it is on the EA forum, it is obviously meant for good and can't be misused in any way.

Contact details

Socials & Sites

Try JournoFinder For Free

Search and contact over 1M+ journalist profiles, browse 100M+ articles, and unlock powerful PR tools.

Start Your 7-Day Free Trial →