My own take is that there is small but non-negligible risk before Anthropic’s ASL-3. For my part I’d vote to move to a lower threshold, or to require more stringent protective measures when working with any system bigger than LLaMA.
I think we’re going to find in the future that even models of Llama or Llama2 size have significant dangers attached depending on the training data and methods applied. This is even setting aside the possibility of future algorithmic improvements which move the params/compute needed for a given capability level downwards. Llama2 is already quite scary in some respects, and could be made scarier by bad actors. Hopefully this point can be made convincingly clear to relevant decision-makers without needing there to be an actualized catastrophe first.
I think we’re going to find in the future that even models of Llama or Llama2 size have significant dangers attached depending on the training data and methods applied. This is even setting aside the possibility of future algorithmic improvements which move the params/compute needed for a given capability level downwards. Llama2 is already quite scary in some respects, and could be made scarier by bad actors. Hopefully this point can be made convincingly clear to relevant decision-makers without needing there to be an actualized catastrophe first.