I’m using ChatGPT for hypothesis generation. This conversation suggests that people are actually brushing their tongues. Previously, I was aware that tongue scraping is a thing, but usually that’s not done with a brush.
On Facebook, I saw one person writing about a programming problem that they had. Another person threw that problem into ChatGPT and ChatGPT gave the right answer.
ChatGTP is not yet good enough to solve every problem that you throw at it on it’s own, but it can help you with brainstorming what might be happening with your problem.
ChatGPT can also correctly answer questions like “Write a Wikidata SPARQL query that shows all women who are poets and who live in Germany”
It’s again an easy-to-verify answer but it’s an answer that allows you to research further. The ability to iterate in a fast matter is useful in combination with other research steps.
This is probably key. If GPT can solve something much faster that’s indeed a win. (With the SPARQL example I guess it would take me 10-20 minutes to look up the required syntax and fields, and put them together. GPT cuts that down to a few seconds, this seems quite good.)
My issue is that I haven’t found a situation yet where GPT is reliably helpful for me. Maybe someone who has found such situations, and reliably integrated “ask GPT first” as a step into some of their workflows could give their account? I would genuinely be curious about practical ways people found to use these models.
My experience has been quite bad so far unfortunately. For example I tried to throw a problem at it that I was pretty sure didn’t have an easy solution, but I just wanted to check that I didn’t miss anything obvious. The answer I would expect in this case is “I don’t know of any easy solution”, but instead I got pages of hallucinated BS. This is worse than if I just hadn’t asked GPT at all since now I have to waste my time reading through its long answers just to realize it’s complete BS.
I’m using ChatGPT for hypothesis generation. This conversation suggests that people are actually brushing their tongues. Previously, I was aware that tongue scraping is a thing, but usually that’s not done with a brush.
On Facebook, I saw one person writing about a programming problem that they had. Another person threw that problem into ChatGPT and ChatGPT gave the right answer.
Yeah I guess many programming problems fall into the “easy to verify” category. (Though definitely not all.)
ChatGTP is not yet good enough to solve every problem that you throw at it on it’s own, but it can help you with brainstorming what might be happening with your problem.
ChatGPT can also correctly answer questions like “Write a Wikidata SPARQL query that shows all women who are poets and who live in Germany”
It’s again an easy-to-verify answer but it’s an answer that allows you to research further. The ability to iterate in a fast matter is useful in combination with other research steps.
This is probably key. If GPT can solve something much faster that’s indeed a win. (With the SPARQL example I guess it would take me 10-20 minutes to look up the required syntax and fields, and put them together. GPT cuts that down to a few seconds, this seems quite good.)
My issue is that I haven’t found a situation yet where GPT is reliably helpful for me. Maybe someone who has found such situations, and reliably integrated “ask GPT first” as a step into some of their workflows could give their account? I would genuinely be curious about practical ways people found to use these models.
My experience has been quite bad so far unfortunately. For example I tried to throw a problem at it that I was pretty sure didn’t have an easy solution, but I just wanted to check that I didn’t miss anything obvious. The answer I would expect in this case is “I don’t know of any easy solution”, but instead I got pages of hallucinated BS. This is worse than if I just hadn’t asked GPT at all since now I have to waste my time reading through its long answers just to realize it’s complete BS.