i think the point is that the answer is not reliable. it might be completely correct or borderline wrong, or something in between, and there’s no way to tell without verifying everything it says - and then one could look it up oneself in the first place already.
Im no fan of generative ai, but this argument drives me crazy, there are a lot of things that are easy to verify but hard to come up with, quite famously in fact.
The main problem I see is that Google just shouldn’t include AI results. And they definitely shouldn’t put their unreliable LLM front and center on the results page. When you google something, you want accurate information, which the LLM might have, but only if that data was readily available to begin with. So the stuff it can help with is stuff the search would put first already.
For anything requiring critical thought or research, the LLM will often hallucinate or misrepresent. The danger is that people do not always apply critical thinking. Defaulting to showing an LLM response is extremely dangerous, and it’s basically pointless.
I don’t know. I find it to be a helpful tool. There’s definitely times it’s wrong (very very wrong sometimes) and there’s sometimes it’s right. It’s up to the user to figure that out.
Maybe I’m old and cynical, but I don’t take anything I read on the Internet, especially something automatically generated, at face value. It’s just another tool I could use to help get to the answer I’m looking for.
it’s still a crapshoot and therefore useless.
Could you be a little more constructive and point me at the points that are wrong and useless?
Thank you.
i think the point is that the answer is not reliable. it might be completely correct or borderline wrong, or something in between, and there’s no way to tell without verifying everything it says - and then one could look it up oneself in the first place already.
Im no fan of generative ai, but this argument drives me crazy, there are a lot of things that are easy to verify but hard to come up with, quite famously in fact.
If you’re using AI verbatim without looking up answers and verifying results, then that’s on you.
When you Google something, do you take the first result and just assume it’s fact? You shouldn’t for AI either.
If you’re going to do the research anyway, why bother with AI?
The main problem I see is that Google just shouldn’t include AI results. And they definitely shouldn’t put their unreliable LLM front and center on the results page. When you google something, you want accurate information, which the LLM might have, but only if that data was readily available to begin with. So the stuff it can help with is stuff the search would put first already.
For anything requiring critical thought or research, the LLM will often hallucinate or misrepresent. The danger is that people do not always apply critical thinking. Defaulting to showing an LLM response is extremely dangerous, and it’s basically pointless.
I don’t know. I find it to be a helpful tool. There’s definitely times it’s wrong (very very wrong sometimes) and there’s sometimes it’s right. It’s up to the user to figure that out.
Maybe I’m old and cynical, but I don’t take anything I read on the Internet, especially something automatically generated, at face value. It’s just another tool I could use to help get to the answer I’m looking for.
Same as most human iterations then?
At least I fact check everything I read. Like I did with this post and the church of the anti-AI got angry they got fact checked.
i didn’t get angry, that was the other guys. just trying to explain it rationally
The part where it gives random results of varying quality, sparky.