If I program something to always reply “2” when you ask it “how many [thing] in [thing]?” It’s not really good at counting. Could it be good? Sure. But that’s not what it was designed to do.
Similarly, LLMs were not designed to count things. So it’s unsurprising when they get such an answer wrong.
Assume it is not. If you’re asking an LLM for information you don’t understand, you’re going to have a bad time. It’s not a learning tool, and using it as such is a terrible idea.
If you want to use it for search, don’t just take it at face value. Click into its sources, and verify the information.
This is hardly programmer humor… there is probably an infinite amount of wrong responses by LLMs, which is not surprising at all.
I don’t know, programs are kind of supposed to be good at counting. It’s ironic when they’re not.
Funny, even.
Eh
If I program something to always reply “2” when you ask it “how many [thing] in [thing]?” It’s not really good at counting. Could it be good? Sure. But that’s not what it was designed to do.
Similarly, LLMs were not designed to count things. So it’s unsurprising when they get such an answer wrong.
I can evaluate this because it’s easy for me to count. But how can I evaluate something else, how can I know whether the LLM ist good at it or not?
Assume it is not. If you’re asking an LLM for information you don’t understand, you’re going to have a bad time. It’s not a learning tool, and using it as such is a terrible idea.
If you want to use it for search, don’t just take it at face value. Click into its sources, and verify the information.