RSS

Sharp Left Turn

TagLast edit: Dec 30, 2024, 9:49 AM by Dakara

Sharp Left Turn is a scenario where, as an AI trains, its capabilities generalize across many domains while the alignment properties that held at earlier stages fail to generalize to the new domains.

See also: Threat Models, AI Takeoff, AI Risk

A cen­tral AI al­ign­ment prob­lem: ca­pa­bil­ities gen­er­al­iza­tion, and the sharp left turn

So8resJun 15, 2022, 1:10 PM
272 points
55 comments10 min readLW link1 review

Refin­ing the Sharp Left Turn threat model, part 1: claims and mechanisms

Aug 12, 2022, 3:17 PM
86 points
4 comments3 min readLW link1 review
(vkrakovna.wordpress.com)

We may be able to see sharp left turns coming

Sep 3, 2022, 2:55 AM
54 points
29 comments1 min readLW link

Refram­ing in­ner alignment

davidadDec 11, 2022, 1:53 PM
53 points
13 comments4 min readLW link

Su­per­in­tel­li­gence’s goals are likely to be random

Mikhail SaminMar 13, 2025, 10:41 PM
3 points
6 comments5 min readLW link

“Sharp Left Turn” dis­course: An opinionated review

Steven ByrnesJan 28, 2025, 6:47 PM
205 points
26 comments31 min readLW link

[In­ter­view w/​ Quintin Pope] Evolu­tion, val­ues, and AI Safety

fowlertmOct 24, 2023, 1:53 PM
11 points
0 comments1 min readLW link

The Sharp Right Turn: sud­den de­cep­tive al­ign­ment as a con­ver­gent goal

avturchinJun 6, 2023, 9:59 AM
38 points
5 comments1 min readLW link

[Question] A few Align­ment ques­tions: util­ity op­ti­miz­ers, SLT, sharp left turn and identifiability

Igor TimofeevSep 26, 2023, 12:27 AM
6 points
1 comment2 min readLW link

We don’t un­der­stand what hap­pened with cul­ture enough

Jan_KulveitOct 9, 2023, 9:54 AM
87 points
22 comments6 min readLW link1 review

Evolu­tion Solved Align­ment (what sharp left turn?)

jacob_cannellOct 12, 2023, 4:15 AM
23 points
89 comments4 min readLW link

Refin­ing the Sharp Left Turn threat model, part 2: ap­ply­ing al­ign­ment techniques

Nov 25, 2022, 2:36 PM
39 points
9 comments6 min readLW link
(vkrakovna.wordpress.com)

[Question] How is the “sharp left turn defined”?

Chris_LeongDec 9, 2022, 12:04 AM
14 points
4 comments1 min readLW link

Vic­to­ria Krakovna on AGI Ruin, The Sharp Left Turn and Paradigms of AI Alignment

Michaël TrazziJan 12, 2023, 5:09 PM
40 points
3 comments4 min readLW link
(www.theinsideview.ai)

Re­sponse to Quintin Pope’s Evolu­tion Pro­vides No Ev­i­dence For the Sharp Left Turn

ZviOct 5, 2023, 11:39 AM
129 points
29 comments9 min readLW link

Agency over­hang as a proxy for Sharp left turn

Nov 7, 2024, 12:14 PM
6 points
0 comments5 min readLW link

[Question] Has Eliezer pub­li­cly and satis­fac­to­rily re­sponded to at­tempted re­but­tals of the anal­ogy to evolu­tion?

kalerJul 28, 2024, 12:23 PM
10 points
14 comments1 min readLW link

Mo­ral gauge the­ory: A spec­u­la­tive sug­ges­tion for AI alignment

James DiacoumisFeb 23, 2025, 11:42 AM
4 points
2 comments8 min readLW link

Goal Align­ment Is Ro­bust To the Sharp Left Turn

Thane RuthenisJul 13, 2022, 8:23 PM
43 points
16 comments4 min readLW link

It mat­ters when the first sharp left turn happens

Adam JermynSep 29, 2022, 8:12 PM
45 points
9 comments4 min readLW link

A smart enough LLM might be deadly sim­ply if you run it for long enough

Mikhail SaminMay 5, 2023, 8:49 PM
19 points
16 comments8 min readLW link

Smoke with­out fire is scary

Adam JermynOct 4, 2022, 9:08 PM
52 points
22 comments4 min readLW link

Evolu­tion pro­vides no ev­i­dence for the sharp left turn

Quintin PopeApr 11, 2023, 6:43 PM
206 points
65 comments15 min readLW link1 review

Disen­tan­gling in­ner al­ign­ment failures

Erik JennerOct 10, 2022, 6:50 PM
23 points
5 comments4 min readLW link

A caveat to the Orthog­o­nal­ity Thesis

Wuschel SchulzNov 9, 2022, 3:06 PM
38 points
10 comments2 min readLW link
No comments.