Archive
Sequences
About
Search
Log In
Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
RSS
AlexMennen
(Alex Mennen)
Karma:
4,451
All
Posts
Comments
New
Top
Old
Page
1
What is calibration?
AlexMennen
13 Mar 2023 6:30 UTC
27
points
1
comment
4
min read
LW
link
Searching for a model’s concepts by their shape – a theoretical framework
Kaarel
,
gekaklam
,
Walter Laurito
,
Kay Kozaronek
,
AlexMennen
and
June Ku
23 Feb 2023 20:14 UTC
51
points
0
comments
19
min read
LW
link
Event [Berkeley]: Alignment Collaborator Speed-Meeting
AlexMennen
and
Carson Jones
19 Dec 2022 2:24 UTC
18
points
2
comments
1
min read
LW
link
Why bet Kelly?
AlexMennen
15 Nov 2022 18:12 UTC
32
points
14
comments
5
min read
LW
link
Average probabilities, not log odds
AlexMennen
12 Nov 2021 21:39 UTC
27
points
20
comments
5
min read
LW
link
Mapping Out Alignment
Logan Riggs
,
adamShimi
,
Gurkenglas
,
AlexMennen
and
Gyrodiot
15 Aug 2020 1:02 UTC
43
points
0
comments
5
min read
LW
link
AlexMennen’s Shortform
AlexMennen
8 Dec 2019 4:51 UTC
7
points
1
comment
1
min read
LW
link
When wishful thinking works
AlexMennen
1 Sep 2018 23:43 UTC
41
points
1
comment
3
min read
LW
link
Safely and usefully spectating on AIs optimizing over toy worlds
AlexMennen
31 Jul 2018 18:30 UTC
24
points
16
comments
2
min read
LW
link
Computational efficiency reasons not to model VNM-rational preference relations with utility functions
AlexMennen
25 Jul 2018 2:11 UTC
16
points
5
comments
3
min read
LW
link
A comment on the IDA-AlphaGoZero metaphor; capabilities versus alignment
AlexMennen
11 Jul 2018 1:03 UTC
40
points
1
comment
1
min read
LW
link
Logical uncertainty and mathematical uncertainty
AlexMennen
1 Jul 2018 0:33 UTC
0
points
0
comments
1
min read
LW
link
(www.lesswrong.com)
Logical uncertainty and Mathematical uncertainty
AlexMennen
26 Jun 2018 1:08 UTC
35
points
6
comments
4
min read
LW
link
More on the Linear Utility Hypothesis and the Leverage Prior
AlexMennen
26 Feb 2018 23:53 UTC
16
points
4
comments
9
min read
LW
link
Value learning subproblem: learning goals of simple agents
AlexMennen
18 Dec 2017 2:05 UTC
0
points
0
comments
2
min read
LW
link
Against the Linear Utility Hypothesis and the Leverage Penalty
AlexMennen
14 Dec 2017 18:38 UTC
39
points
47
comments
11
min read
LW
link
Being legible to other agents by committing to using weaker reasoning systems
AlexMennen
3 Dec 2017 7:49 UTC
4
points
1
comment
3
min read
LW
link
Metamathematics and probability
AlexMennen
22 Sep 2017 4:04 UTC
1
point
0
comments
1
min read
LW
link
(alexmennen.com)
Metamathematics and Probability
AlexMennen
22 Sep 2017 3:07 UTC
1
point
0
comments
1
min read
LW
link
(alexmennen.com)
Density Zero Exploration
AlexMennen
17 Aug 2017 0:43 UTC
4
points
0
comments
2
min read
LW
link
Back to top
Next