Giulio
Rome Italy—ACX Meetups Everywhere Fall 2024
Language for Goal Misgeneralization: Some Formalisms from my MSc Thesis
Due to the weather, we’re moving this from Saturday to Sunday. Same time same place.
Rome – ACX Meetups Everywhere Spring 2024
why not?
I think there can be some parallels to be made between debates on gun control and debates on recent AI regulations
Would be nice to have a website collating people’s public p(doom) statements
Given the name, I imagine this line of research is inspired by model organism research, although I wonder if a simpler “demonstrations of misalignment” could’ve been sufficient/better.
“don’t hate the player, hate the game”
Moloch is “the game”
Asteroid movies probably made people more receptive to x-risk from cosmic collisions
maybe we need a movie about x-risk from misaligned AI? something like Ex Machina and/or Her but with more focus on consequences and less robots
idk could be “counterproductive” too I guess
Agency, LLMs and AI Safety—A First Pass
It has come to my attention he’s on a sabbatical. That’s great, but his activity (tweets, podcasts) don’t suggest the level of detachment from engagement I was imagining
has EY considered taking a break? Like a really long (at least 1 year) vacation where he’s mostly disconnected from AI news and just the world im general. Maybe sail the world or something. Starting to seem like he has given up anyway. Maybe exiting the bubble a bit will allow for new hope (and ideas? motivation?) to form.
“Quote tweeting” this:
Wonder if it’s worth synchronizing my Twitter with LW shortform.
Probably not. I think I will just handpick which tweets I repost here. Plus some shortform exclusives maybe.
is shortform basically just twitter on LW? seems a little like it
- Mar 5, 2023, 1:36 AM; 1 point) 's comment on thesofakillers’s Shortform by (
thank you for the post! Are you able to share roughly what the interview/hiring process will be like? E.g. how many stages, duration, etc