ACX Brno meetup #2

adekcz11 Jun 2023 13:53 UTC
2 points
0 comments1 min readLW link

[Linkpost] Large Lan­guage Models Con­verge on Brain-Like Word Representations

Bogdan Ionut Cirstea11 Jun 2023 11:20 UTC
36 points
12 comments1 min readLW link

In­fer­ence-Time In­ter­ven­tion: Elic­it­ing Truth­ful An­swers from a Lan­guage Model

likenneth11 Jun 2023 5:38 UTC
195 points
4 comments1 min readLW link
(arxiv.org)

You Are a Com­puter, and No, That’s Not a Metaphor

jakej11 Jun 2023 5:38 UTC
12 points
1 comment22 min readLW link
(sigil.substack.com)

Snake Eyes Paradox

Martin Randall11 Jun 2023 4:10 UTC
22 points
25 comments6 min readLW link

[Question] [Mostly solved] I get dis­tracted while read­ing, but can eas­ily com­pre­hend au­dio text for 8+ hours per day. What are the best AI text-to-speech read­ers? Alter­na­tively, do you have other ideas for what I could do?

kuira11 Jun 2023 3:49 UTC
18 points
7 comments1 min readLW link

The Dic­ta­tor­ship Problem

alyssavance11 Jun 2023 2:45 UTC
34 points
145 comments11 min readLW link

Higher Di­men­sion Carte­sian Ob­jects and Align­ing ‘Tiling Si­mu­la­tors’

lukemarks11 Jun 2023 0:13 UTC
22 points
0 comments5 min readLW link

Us­ing Con­sen­sus Mechanisms as an ap­proach to Alignment

Prometheus10 Jun 2023 23:38 UTC
9 points
2 comments6 min readLW link

Hu­man­i­ties first math prob­lem, The shal­low gene pool.

archeon10 Jun 2023 23:09 UTC
−2 points
0 comments1 min readLW link

I can see how I am Dumb

Johannes C. Mayer10 Jun 2023 19:18 UTC
46 points
11 comments5 min readLW link

Etho­dy­nam­ics of Omelas

dr_s10 Jun 2023 16:24 UTC
79 points
17 comments9 min readLW link

Deal­ing with UFO claims

ChristianKl10 Jun 2023 15:45 UTC
3 points
32 comments1 min readLW link

A The­ory of Un­su­per­vised Trans­la­tion Mo­ti­vated by Un­der­stand­ing An­i­mal Communication

jsd10 Jun 2023 15:44 UTC
19 points
0 comments1 min readLW link
(arxiv.org)

[Question] What are brains?

Valentine10 Jun 2023 14:46 UTC
10 points
22 comments2 min readLW link

EY in the New York Times

Blueberry10 Jun 2023 12:21 UTC
6 points
14 comments1 min readLW link
(www.nytimes.com)

Goal-mis­gen­er­al­iza­tion is ELK-hard

rokosbasilisk10 Jun 2023 9:32 UTC
2 points
0 comments1 min readLW link

[Question] What do benefi­cial TDT trades for hu­man­ity con­cretely look like?

Stephen Fowler10 Jun 2023 6:50 UTC
4 points
0 comments1 min readLW link

cloud seed­ing doesn’t work

bhauth10 Jun 2023 5:14 UTC
7 points
2 comments1 min readLW link

[FICTION] Un­box­ing Ely­sium: An AI’S Escape

Super AGI10 Jun 2023 4:41 UTC
−16 points
4 comments14 min readLW link

[FICTION] Prometheus Ris­ing: The Emer­gence of an AI Consciousness

Super AGI10 Jun 2023 4:41 UTC
−14 points
0 comments9 min readLW link

for­mal­iz­ing the QACI al­ign­ment for­mal-goal

10 Jun 2023 3:28 UTC
54 points
6 comments13 min readLW link
(carado.moe)

Ex­pert trap: Why is it hap­pen­ing? (Part 2 of 3) – how hind­sight, hi­er­ar­chy, and con­fir­ma­tion bi­ases break con­duc­tivity and ac­cu­racy of knowledge

Paweł Sysiak9 Jun 2023 23:00 UTC
3 points
0 comments7 min readLW link

Ex­pert trap: What is it? (Part 1 of 3) – how hind­sight, hi­er­ar­chy, and con­fir­ma­tion bi­ases break con­duc­tivity and ac­cu­racy of knowledge

Paweł Sysiak9 Jun 2023 23:00 UTC
6 points
2 comments8 min readLW link

[Question] How ac­cu­rate is data about past earth tem­per­a­tures?

tailcalled9 Jun 2023 21:29 UTC
10 points
2 comments1 min readLW link

Proxi-An­tipodes: A Geo­met­ri­cal In­tu­ition For The Difficulty Of Align­ing AI With Mul­ti­tudi­nous Hu­man Values

Matthew_Opitz9 Jun 2023 21:21 UTC
7 points
0 comments5 min readLW link

Why AI may not save the World

Alberto Zannoni9 Jun 2023 17:42 UTC
0 points
0 comments4 min readLW link
(a16z.com)

You can now listen to the “AI Safety Fun­da­men­tals” courses

PeterH9 Jun 2023 16:45 UTC
6 points
0 comments1 min readLW link
(forum.effectivealtruism.org)

Ex­plor­ing Con­cept-Spe­cific Slices in Weight Ma­tri­ces for Net­work Interpretability

DuncanFowler9 Jun 2023 16:39 UTC
1 point
0 comments6 min readLW link

A plea for solu­tion­ism on AI safety

jasoncrawford9 Jun 2023 16:29 UTC
72 points
6 comments6 min readLW link
(rootsofprogress.org)

Michael Shel­len­berger: US Has 12 Or More Alien Space­craft, Say Mili­tary And In­tel­li­gence Contractors

lc9 Jun 2023 16:11 UTC
11 points
31 comments3 min readLW link
(public.substack.com)

Im­prove­ment on MIRI’s Corrigibility

9 Jun 2023 16:10 UTC
54 points
8 comments13 min readLW link

D&D.Sci 5E: Re­turn of the League of Defen­ders Eval­u­a­tion & Ruleset

aphyer9 Jun 2023 15:25 UTC
30 points
8 comments6 min readLW link

In­ternLM—China’s Best (Un­ver­ified)

Lao Mein9 Jun 2023 7:39 UTC
51 points
4 comments1 min readLW link

[Question] Mark for fol­low up?

JNS9 Jun 2023 5:59 UTC
5 points
4 comments2 min readLW link

Bring­ing Lit­tle Kids to Con­tra Dances

jefftk9 Jun 2023 2:20 UTC
22 points
0 comments2 min readLW link
(www.jefftk.com)

[Question] (solved) how do i find oth­ers’ short­form posts?

kuira9 Jun 2023 2:15 UTC
1 point
1 comment1 min readLW link

[Question] AI Rights: In your view, what would be re­quired for an AGI to gain rights and pro­tec­tions from the var­i­ous Govern­ments of the World?

Super AGI9 Jun 2023 1:24 UTC
10 points
26 comments1 min readLW link

A com­par­i­son of causal scrub­bing, causal ab­strac­tions, and re­lated methods

8 Jun 2023 23:40 UTC
73 points
3 comments22 min readLW link

Up­dates and Reflec­tions on Op­ti­mal Ex­er­cise af­ter Nearly a Decade

romeostevensit8 Jun 2023 23:02 UTC
210 points
56 comments2 min readLW link1 review

Take­aways from the Mechanis­tic In­ter­pretabil­ity Challenges

scasper8 Jun 2023 18:56 UTC
94 points
5 comments6 min readLW link

Leave an Emo­tional Line of Retreat

Johannes C. Mayer8 Jun 2023 18:36 UTC
23 points
1 comment1 min readLW link

Cur­rent AI harms are also sci-fi

Christopher King8 Jun 2023 17:49 UTC
26 points
3 comments1 min readLW link

Two Ways To Re­duce Un­hap­piness That Comes From Dis­torted Views of Reality

Anne Hsu8 Jun 2023 17:43 UTC
3 points
0 comments7 min readLW link

Col­lab­o­ra­tion in Science: Hap­pier Peo­ple ↔ Bet­ter Research

nadinespy8 Jun 2023 17:42 UTC
3 points
0 comments32 min readLW link

Biomimetic al­ign­ment: Align­ment be­tween an­i­mal genes and an­i­mal brains as a model for al­ign­ment be­tween hu­mans and AI sys­tems

geoffreymiller8 Jun 2023 16:05 UTC
10 points
1 comment16 min readLW link

A po­ten­tially high im­pact differ­en­tial tech­nolog­i­cal de­vel­op­ment area

Noosphere898 Jun 2023 14:33 UTC
5 points
2 comments2 min readLW link

[Question] Ques­tion for Pre­dic­tion Mar­ket peo­ple: where is the money sup­posed to come from?

Robert_AIZI8 Jun 2023 13:58 UTC
25 points
26 comments1 min readLW link

AI #15: The Prin­ci­ple of Charity

Zvi8 Jun 2023 12:10 UTC
73 points
16 comments44 min readLW link
(thezvi.wordpress.com)

if you’re read­ing this it’s too late (a new the­ory on what is caus­ing the Great Stag­na­tion)

rogersbacon8 Jun 2023 11:49 UTC
−10 points
2 comments13 min readLW link
(www.secretorum.life)