Indeed!
They do *sort of* have to be this way, in the sense that we don't have a solution to the factuality problem in large language models. What they do, architecturally, is spout stuff that sounds sort of like their training set, but without any notion of true or false.
So as far as this particular specific technology is concerned, they do have to be this way; it's just what they do! The problem isn't as much incentive as it is basic knowledge. People are working hard on altering or adding to the basic technology to make them emit truths where that's important, but it's definitely an unsolved problem.