RSS

Nicholas / Heather Kross

Karma: 1,985

Theoretical AI alignment (and relevant upskilling) in my free time. My current view of the field is here (part 1) and here (part 2).

Genderfluid (differs on hour/​day-ish timescale.). It’s not a multiple-personality thing.

/​nickai/​

Is prin­ci­pled mass-out­reach pos­si­ble, for AGI X-risk?

Nicholas / Heather Kross21 Jan 2024 17:45 UTC
9 points
5 comments3 min readLW link

Learn­ing Math in Time for Alignment

Nicholas / Heather Kross9 Jan 2024 1:02 UTC
32 points
3 comments3 min readLW link

Up­grad­ing the AI Safety Community

16 Dec 2023 15:34 UTC
42 points
9 comments42 min readLW link

[Question] In­tel­li­gence En­hance­ment (Monthly Thread) 13 Oct 2023

Nicholas / Heather Kross13 Oct 2023 17:28 UTC
52 points
40 comments1 min readLW link

How to Get Ra­tion­al­ist Feedback

Nicholas / Heather Kross5 Oct 2023 2:03 UTC
13 points
0 comments2 min readLW link

Musk, Star­link, and Crimea

Nicholas / Heather Kross23 Sep 2023 2:35 UTC
−13 points
0 comments5 min readLW link

[Question] In­cen­tives af­fect­ing al­ign­ment-re­searcher encouragement

Nicholas / Heather Kross29 Aug 2023 5:11 UTC
28 points
3 comments1 min readLW link

[Question] How nec­es­sary is in­tu­ition, for ad­vanced math?

Nicholas / Heather Kross20 Jul 2023 0:18 UTC
11 points
8 comments1 min readLW link

[Question] Build knowl­edge base first, or backchain?

Nicholas / Heather Kross17 Jul 2023 3:44 UTC
11 points
5 comments1 min readLW link

Ra­tion­al­ity, Ped­a­gogy, and “Vibes”: Quick Thoughts

Nicholas / Heather Kross15 Jul 2023 2:09 UTC
14 points
1 comment4 min readLW link

Align­ment Me­gapro­jects: You’re Not Even Try­ing to Have Ideas

Nicholas / Heather Kross12 Jul 2023 23:39 UTC
55 points
30 comments2 min readLW link

My Cen­tral Align­ment Pri­or­ity (2 July 2023)

Nicholas / Heather Kross3 Jul 2023 1:46 UTC
12 points
1 comment3 min readLW link

My Align­ment Timeline

Nicholas / Heather Kross3 Jul 2023 1:04 UTC
22 points
0 comments2 min readLW link

How to Search Mul­ti­ple Web­sites Quickly

Nicholas / Heather Kross22 Jun 2023 0:42 UTC
16 points
1 comment1 min readLW link

[Question] Does any­one’s full-time job in­clude read­ing and un­der­stand­ing all the most-promis­ing for­mal AI al­ign­ment work?

Nicholas / Heather Kross16 Jun 2023 2:24 UTC
15 points
2 comments1 min readLW link

Dreams of “Matho­pe­dia”

Nicholas / Heather Kross2 Jun 2023 1:30 UTC
40 points
16 comments2 min readLW link
(www.thinkingmuchbetter.com)

Ab­strac­tion is Big­ger than Nat­u­ral Abstraction

Nicholas / Heather Kross31 May 2023 0:00 UTC
18 points
0 comments5 min readLW link
(www.thinkingmuchbetter.com)

My AI Align­ment Re­search Agenda and Threat Model, right now (May 2023)

Nicholas / Heather Kross28 May 2023 3:23 UTC
25 points
0 comments6 min readLW link
(www.thinkingmuchbetter.com)

Why and When In­ter­pretabil­ity Work is Dangerous

Nicholas / Heather Kross28 May 2023 0:27 UTC
20 points
8 comments8 min readLW link
(www.thinkingmuchbetter.com)

Why I’m Not (Yet) A Full-Time Tech­ni­cal Align­ment Researcher

Nicholas / Heather Kross25 May 2023 1:26 UTC
39 points
21 comments4 min readLW link
(www.thinkingmuchbetter.com)