Any idea what’s happening with the 34B model? Why might it be so much less “safe” than the bigger and smaller versions? And what about the base version of the 34B—are they not releasing that? But the base version isn’t supposed to be “safe” anyway...
Any idea what’s happening with the 34B model? Why might it be so much less “safe” than the bigger and smaller versions? And what about the base version of the 34B—are they not releasing that? But the base version isn’t supposed to be “safe” anyway...
Relevant rumors / comments:
Seems like we can continue to scale tokens and get returns model performance well after 2T tokens. : r/LocalLLaMA (reddit.com)
LLaMA 2 is here : r/LocalLLaMA (reddit.com)