RSS

AI Safety Cases

TagLast edit: Nov 19, 2024, 10:17 PM by Rauno Arike

A safety case is a structured argument showing that a system is acceptably safe for a specific use in a specific environment. Safety cases typically include:

Near- and medium-term AI Con­trol Safety Cases

Martín SotoDec 23, 2024, 5:37 PM
9 points
0 comments6 min readLW link

A sketch of an AI con­trol safety case

Jan 30, 2025, 5:28 PM
57 points
0 comments5 min readLW link

Notes on con­trol eval­u­a­tions for safety cases

Feb 28, 2024, 4:15 PM
49 points
0 comments32 min readLW link

An­thropic: Three Sketches of ASL-4 Safety Case Components

Zach Stein-PerlmanNov 6, 2024, 4:00 PM
95 points
33 comments1 min readLW link
(alignment.anthropic.com)

AI com­pa­nies are un­likely to make high-as­surance safety cases if timelines are short

ryan_greenblattJan 23, 2025, 6:41 PM
145 points
5 comments13 min readLW link

New re­port: Safety Cases for AI

joshcMar 20, 2024, 4:45 PM
89 points
14 comments1 min readLW link
(twitter.com)

Toward Safety Cases For AI Scheming

Oct 31, 2024, 5:20 PM
60 points
1 comment2 min readLW link
No comments.