It is possible to create a good model of a person with current LLMs who will behave 70-90 percent like me. The model could even claim that it is conscious. I experimented with my model, but it is most likely not conscious (or all LLMs are conscious).
It is possible to create a good model of a person with current LLMs who will behave 70-90 percent like me. The model could even claim that it is conscious. I experimented with my model, but it is most likely not conscious (or all LLMs are conscious).