
Jonas Hallgren
Articles
-
Jan 10, 2025 |
lesswrong.com | Alvin Ånestrand |Jonas Hallgren
The Alignment Mapping Program: Forging Independent Thinkers in AI Safety - A Pilot RetrospectiveThe AI safety field faces a critical challenge: we need researchers who can not only implement existing solutions but also forge new, independent paths. In 2023, inspired by John Wentworth's work on agency and learning from researchers like Rohin Shah and Adam Shimi who have highlighted the limitations of standard AI safety education, we launched the Alignment Mapping Program (AMP).
-
Jan 7, 2025 |
lesswrong.com | Jonas Hallgren
Okay, so after engaging a lot with Steven Byrnes post on [Intuitive self-models] 6. Awakening / Enlightenment / PNSE I’ve tried to think about the most important framing shift that is different from me compared to what Steven has written. We will be going through a generative model argument from an agency and predictive processing perspective. This will then lead to an understanding of meditation insights and “awakening” as changes in the self-perceived information boundary.
-
Dec 5, 2024 |
lesswrong.com | Jonas Hallgren |Seth Herd
EVERYONE, CALM DOWN!Meaning Alignment Institute just dropped their first post in basically a year and it seems like they've been up to some cool stuff. Their perspective on value alignment really grabbed my attention because it reframes our usual technical alignment conversations around rules and reward functions into something more fundamental - what makes humans actually reliably good and cooperative?
-
Dec 5, 2024 |
lesswrong.com | Jonas Hallgren
EVERYONE, CALM DOWN! Meaning Alignment Institute just dropped their first post in basically a year and it seems like they've been up to some cool stuff. Their perspective on value alignment really grabbed my attention because it reframes our usual technical alignment conversations around rules and reward functions into something more fundamental - what makes humans actually reliably good and cooperative?
-
Apr 2, 2024 |
lesswrong.com | Jonas Hallgren
This post wouldn't have happened without this post, for which I'm forever gratefulIn order to be able to steer away asteroids from earth we first have to be able to model them properly. This is a post on how to do this safely. Coincidentally, this would be perfect technology for Open Asteroid Impact to improve its direction models, yet as it is on the EA forum, it is obviously meant for good and can't be misused in any way.
Try JournoFinder For Free
Search and contact over 1M+ journalist profiles, browse 100M+ articles, and unlock powerful PR tools.
Start Your 7-Day Free Trial →