Yes, I used it quite a bit. So, yes, all of us can compare to some extent.
But I’ve also read Janus enough (here and on twitter) to know that RLHF mutilates models quite a bit (both via “mode collapse” and via other multiple pathologies; the net result is drastic restrictions of the set of simulations the model can create).
So it potentially might be that base-GPT-4 is drastically more powerful than RLHF’d GPT-4 if one knows how to handle it right...
So, in fact, I particularly wanted Janus’ impressions to be recorded and shared. That’s because I really wanted to know how base-GPT-4 looks through the prism of their general insights, given their writings on the Simulator theory and on LLMs in general (and their ability to deal with potentially high non-triviality of dealing with non-RLHF’d GPT-4; in this sense, note their remark on how base-GPT-4 is particularly sensitive to the quality of prompt writing; so it’s a very different beast, much more difficult to handle than RLHF’d GPT-4, but the pay-offs for the qualified interlocutor might be really high).
Although, of course, I’d love to have impressions from other people, and I’d love to read discussions about this… For that we need more people with access to base-GPT-4 to at least notice this post :-)
Yes, I used it quite a bit. So, yes, all of us can compare to some extent.
But I’ve also read Janus enough (here and on twitter) to know that RLHF mutilates models quite a bit (both via “mode collapse” and via other multiple pathologies; the net result is drastic restrictions of the set of simulations the model can create).
So it potentially might be that base-GPT-4 is drastically more powerful than RLHF’d GPT-4 if one knows how to handle it right...
So, in fact, I particularly wanted Janus’ impressions to be recorded and shared. That’s because I really wanted to know how base-GPT-4 looks through the prism of their general insights, given their writings on the Simulator theory and on LLMs in general (and their ability to deal with potentially high non-triviality of dealing with non-RLHF’d GPT-4; in this sense, note their remark on how base-GPT-4 is particularly sensitive to the quality of prompt writing; so it’s a very different beast, much more difficult to handle than RLHF’d GPT-4, but the pay-offs for the qualified interlocutor might be really high).
Although, of course, I’d love to have impressions from other people, and I’d love to read discussions about this… For that we need more people with access to base-GPT-4 to at least notice this post :-)