lol Paul is a very non-disparaging person. He always makes his criticism constructive, i don’t know if there’s any public evidence of him disparaging anyone regardless of NDAs
Tao Lin
I’ve recently gotten into partner dancing and I think it’s a pretty superior activity
One lesson you could take away from this is “pay attention to the data, not the process”—this happened because the data had longer successes than failures. If successes were more numerous than failures, many algorithms would have imitated those as well with null reward.
I think the “fraction of Training compute” going towards agency vs nkn agency will be lower in video models than llms, and llms will likely continue to be bigger, so video models will stay behind llms in overall agency
Helpfullness finetuning might make these models more capable when they’re on the correct side of the debate. Sometimes RLHF(like) models simply perform worse on tasks they’re finetuned to avoid even when they don’t refuse or give up. Would be nice to try base model debaters
A core advantage of bandwidth limiting over other cybersec interventions is its a simple system we can make stronger arguments about, implemented on a simple processor, without the complexity and uncertainty of modern processors and OSes
no clock speed stays the same, but clock cycle latency of communication between regions increases. Just like CPUs require more clock cycles to access memory than they used to.
do we have any reason to believe that particular election won’t be close
I’d expect artificial sweeteners are already very cheap, and most people want more tested chemicals.
There exists an Effective Altruism VR discord group. It used to have regular VRChat meetups in like 2021 but doesn’t have much activity now i think
I’d be interested in experiments with more diverse data. Maybe this only works because the passages are very short and simple and uniform, and are using very superposition-y information that wouldn’t exist in longer and more diverse text
i thought about this for a minute and landed on no counting for lorentz factor. Things hitting on the side have about the same relative velocity as things hitting from the front . Because they’re hitting the side they could either bounce off or dump all their tangent kinetic energy into each other. like because all the relative velocity is tangent, they could in principle interact without exchanging significant energy. But probably the side impacts are just as dangerous. Which might make them more dangerous because you have less armor on the side
probes probably want a very skinny aspect ratio. If cosmic dust travels at 20km/s, that’s 15k times slower than the probe is travelling, so maybe that means the probe should be eg 10cm wide and 1.5km long
important to note that gpt4 is more like 300x scale equivalent of gpt3, not 100x, based on gpt4 being trained with (rumored) 2e25 flops vs contemporary gpt3-level models (llama2-7b) being trained on 8e22 flops ( 250 times the compute for that particular pair)
Some months before release they had a RLHF-ed model, where the RLHF was significantly worse on most dimensions than the model they finally released. This early RLHF-ed model was mentioned in eg Sparks of AGI.
Send us example gnarly bugs
if AI does change the offence defence balance, it could be because defending an AI (that doesnt need to protect humans) is fundamentally different than defending humans, allowing the AI to spend much less on defence
video can get extremely expensive without specific architectural support. Eg a folder of images takes up >10x the space of the equivalent video, and using eg 1000 tokens per frame for 30 frames/second is a lot of compute
looks slightly behind gpt-4-base in benchmarks. On the tasks where gemini uses chain-of-thought best-of-32 with optimized prompts it beats gpt-4-base, but ones where it doesnt its same or behind
I don’t think slaughtering billions of people would be very useful. As a reference point, wars between countries almost never result in slaughtering that large a fraction of people