Presumably they have a way for it to learn from each time it succeeds or fails (RL + feedback the weights that genned the winning tokens or—the ones that caused errors) or this whole plugin feature will fail to be useful.
Growing pains for sure. Let’s see if OAI will improve it, via RL or whatever other method. Probably we will see it start to work more reliably, but we will not know why (since OAI has not been that ‘open’ recently).
It does, and it actually doesn’t do it very well. I made a post where you can see it fail to use Wolfram Alpha.
Presumably they have a way for it to learn from each time it succeeds or fails (RL + feedback the weights that genned the winning tokens or—the ones that caused errors) or this whole plugin feature will fail to be useful.
Growing pains for sure. Let’s see if OAI will improve it, via RL or whatever other method. Probably we will see it start to work more reliably, but we will not know why (since OAI has not been that ‘open’ recently).