The digits given by the model are wrong (one has e*sqrt(3) ~4.708). Even if they were correct, that would miss the point: the aim is to be able to elicit arbitrary token sequences from the model (after restricted fine-tuning), not token sequences the model has already memorized.
The problem is not “it’s hard to get any >50 digit sequence out of GPT-3.5”, but “it’s hard to make GPT-3.5 precisely ‘stitch together’ sequences it already knows”.
The digits given by the model are wrong (one has e*sqrt(3) ~4.708). Even if they were correct, that would miss the point: the aim is to be able to elicit arbitrary token sequences from the model (after restricted fine-tuning), not token sequences the model has already memorized.
The problem is not “it’s hard to get any >50 digit sequence out of GPT-3.5”, but “it’s hard to make GPT-3.5 precisely ‘stitch together’ sequences it already knows”.