The current trend is for these models to become smaller and smaller to the point where they will soon run locally on a smartphone.
I agree they will become smaller, but I would guess this is not the current trend. Why do you think that?
I agree with OP but will not explain further.
Can read about quantization and the compression of GPT models here:
https://beuke.org/quantization/
Current theme: default
Less Wrong (text)
Less Wrong (link)
I agree they will become smaller, but I would guess this is not the current trend. Why do you think that?
I agree with OP but will not explain further.
Can read about quantization and the compression of GPT models here:
https://beuke.org/quantization/