Quote:
Originally Posted by gmw
It would be interesting to know if the reference issues were deliberate. It seems like a reasonable way to offer some sort of protection: force anyone wanting to produce academic quality results to at least validate and correct the references. For general use, the occasional redundant reference is not going to be a huge issue.
|
They are intentional. The training data is never disclosed, and if it did, we would scream bloody murder for all the bias they put in. All hiding under the "truth check" umbrella.
The only true test would be to train the AI by giving it full access to the world wide web, or a subset with a short list of domains. Then see what it learned, whatever it spits out as answer should be able to be verified by manually searching for the answer with search engine instead of a LLM.