I’d recommend the “AI safety via conditioning predictive models” doc my coauthors and I are working on right now—it’s not quite ready to be published publicly yet, but we have a full draft that we’re looking for comments on right now. Messaged to both of you privately; feel free to share with other HAIST members.
I’d recommend the “AI safety via conditioning predictive models” doc my coauthors and I are working on right now—it’s not quite ready to be published publicly yet, but we have a full draft that we’re looking for comments on right now. Messaged to both of you privately; feel free to share with other HAIST members.
Could I get a link to this as well?
Would also love to have a look.