It was mentioned that lacking a topic was a bit off-putting to people I have two topics to go over next meeting.
First off is the EA community in Houston at large. First off I cannot seem to find an active EA meetup in Houston other than that at UH which is new. I would be interested in discussing how to help the local community grow and persist. It seems like there would be a need for an EA community to exist outside the university structure.
Secondly, I am half way through Hanson’s Elephant in the Brain. I feel like signaling hypothesis would be fun to discuss in a rationalist meetup.
The crux of these types of arguments seems to be conflating the provable safety of an agent in a system with the expectation of absolute safety. In my experience, this is the norm, not the exception, and needs to be explicitly addressed.
In agreement with what you posted above, I think it is formally trivial to construct a scenario in which a pedestrian jumps in front of a car, making it provably impossible for a vehicle to stop in time to avoid a collision using high school physics.
Likewise, I have the intuition that AI safety, in general, should have various “no-go theorems” about unprovability outside a reasonable problem scope or that finding such proofs would be np-hard or worse. If you know of any specific results( outside of general computability theory) , could you please share them? It would be nice if the community could avoid falling into the trap of trying to prove too much.
(Sorry if this isn’t the correct location for this post.)