
Robert Wiblin
Exploring the inviolate sphere of ideas one interview at a time: https://t.co/2YMw00bkIQ
Articles
-
6 days ago |
80000hours.org | Robert Wiblin
I do think that there is something to celebrate here. … But we’re far from done. … There’s an opportunity to actually really make this work — I think that’s an optimistic way of looking at it. Now that they have said the nonprofit will be in control, what we’re hoping for is to see details that will show us how the nonprofit will actually be in control. Like, the words are great… but what’s going to back that up?
-
1 week ago |
80000hours.org | Robert Wiblin
TranscriptIan Dunt: Britain can tend to be a bit lost in the opium dreams of its own past. Whenever you hear a prime minister talk about Downing Street, the thing they always talk about is the staircase with all of the portraits of the various prime ministers. And I get it: some of those prime ministers were people that were crucial in the history of liberal democracy, or some of them fought off Nazism. I get that that is obviously a powerful thing.
-
4 weeks ago |
80000hours.org | Robert Wiblin |Lukas Finnveden
Once we get to a world where it is technologically possible to replace those researchers with AI systems — which could just be fully obedient, instruction-following AI systems — then you could feasibly have a situation where there’s just one person at the top of the organisation that gives a command: “This is how I want the next AI system to be developed.” And then this army of loyal, obedient AIs will then do all of the technical work.
-
1 month ago |
80000hours.org | Robert Wiblin
Another thing you might do once you’ve caught your AI trying to escape is let the AI think it succeeded, and see what it does. Presumably when the AI got caught, it was in the midst of taking some actions that it thought would let it launch a rogue deployment. If your model knows a bunch of security vulnerabilities in your software, it might start deploying all those things. — Buck ShlegerisMost AI safety conversations centre on alignment: ensuring AI systems share our values and goals.
-
2 months ago |
80000hours.org | Robert Wiblin
You should really pause and reflect on the fact that many companies now are saying what we want to do is build AGI — AI that is as good as humans. OK, what does it look like? What does a good society look like when we have humans and we have trillions of AI beings going around that are functionally much more capable? What’s the vision like? How do we coexist in an ethical and morally respectable way? And it’s like… there’s nothing. We’re careening towards this vision that is just a void, essentially.
Try JournoFinder For Free
Search and contact over 1M+ journalist profiles, browse 100M+ articles, and unlock powerful PR tools.
Start Your 7-Day Free Trial →X (formerly Twitter)
- Followers
- 38K
- Tweets
- 683
- DMs Open
- Yes

RT @nabeelqu: So we're in an RL-scaling world, where reward hacking and deception by AI models is common; a capabilities scaling arms race;…

RT @labenz: OpenAI's o3/o4 models show *huge* gains toward "automating the job of an OpenAI research engineer" With that, the risk of AI-e…

RT @DKokotajlo: Internal deployment was always the main threat model IMO. IIRC I tried to get the Preparedness Framework to cover internal…