OP clearly expects LLMs to exhibit mind-like behaviors. Lying absolutely implies agency, but even if you don’t agree, OP is confused that
It did not simply analyze the best type of graphics card for the situation
The whole point of the post is that OP is upset that LLMs are generating falsehoods and parroting input back into its output. No one with a basic understanding of LLMs would be surprised by this. If someone said their phone’s autocorrect was “lying”, you’d be correct in assuming they didn’t understand the basics of what autocorrect is, and would be completely justified in pointing out that that’s nonsense.
ba dum chhhhhhhhhhhhhhh