We all know exactly the kinds of bad outputs we've seen from AI.
I just ran a query for ChatGPT to recommend various databases for a specific use case. Out of the seven databases it recommended, only one was a database actually appropriate; one suggestion was marginally acceptable; three of the recommendations weren't even databases.
I then asked it to provide a list of the most important battles in North Africa prior to the entry of the United States into World War 2.
It gave me five answers. Three of which occurred after the entry of the United States into World War 2.
AIs provides extremely plausible answers. Sometimes it will actually generate correct, useful output; but you cannot yet rely on it for correctness.
I'd like to see a side by side comparison with a random human on the street. Maybe with a sample size of 100 or so. How well do you think the humans would do vs whatever outdated model you were playing with here?
There is clearly significant value to this tech and I'm still dumbfounded how strongly some people try to deny it.
Anyone reckon there's a chance that GPT hallucinates because it was trained on online material (e.g. reddit and other forums)? I'd have to say on topics I know GPT is about as hit or miss as a random internet comment, especially in that they'll both give a confidently stated answer whether the answer is factual or not.
Is it possible GPT just thinks[0] that any answer stated confidently is preferable over not giving an answer?
Promise I'm not just being snarky, legitimate wonder!
[0]: I know it doesn't actually think, you know what I mean
You're judging fish by its ability to climb a tree. Being able to recall facts is a nice side effect for LLMs, not their bread and butter. If you need facts, plug in some RAG to it.
Well, learning and conversing are different things done for different reasons.
I'll agree that GPT4 has completely replaced Google, stackoverflow, etc for me.
The only time I use Google now is for janky more human like situations.
For example, today I had to transfer DC roles from a Windows 2012 R2 server to a new 2022. They only have one DC. And the old DC has a basically unused CA service set up.
ChatGPT would have me "fix" everything first, wheras, I did find a forum post with a situation almost identical to mine that helped me cowboy it rather than being overly meticulous.
There is still value to human experience. For now.
I just ran a query for ChatGPT to recommend various databases for a specific use case. Out of the seven databases it recommended, only one was a database actually appropriate; one suggestion was marginally acceptable; three of the recommendations weren't even databases.
I then asked it to provide a list of the most important battles in North Africa prior to the entry of the United States into World War 2.
It gave me five answers. Three of which occurred after the entry of the United States into World War 2.
AIs provides extremely plausible answers. Sometimes it will actually generate correct, useful output; but you cannot yet rely on it for correctness.