Even non-polictial topics it got wrong. E.g washing your hair depending in your hair type, it just told me lie after lie, and when I called it out it claimed I told it wrongly...
Yes, standard, generalized LLMs are known to make up stuff out of whole cloth. The industry calls it "hallucinations." It is a technical problem. In one session, Google Gemini invented a docuмent from Pius XII as a citation. The docuмent did not exist. But Gemini gave it an elaborate title, the date it was promulgated, etc. The reason it did that was it was trying to "help" me, as it is programmed to do. I had asked for a citation and it made one up. That was just one example.
1. The "hallucination" problem is not the same as the topic policing problem.
2. The topic policing problem is real, but it gets confused with the real problem: the consensus model of truth.
The Consensus Model of Truth is the fundamental problem of the LLMs. They are trained on huge amounts of data from the Internet. Most of the training data is garbage that repeats the common false narratives over and over again, reinforcing the false narrative in the LLMs consensus model of truth. To be useful, the AI needs to use a hierarchical model of truth with trusted docuмents as your sources. But that is not how they work.
It is like asking a random group of people if the Catholic Church is the one true church. The standard LLM will say, "The question of whether the Catholic Church is the 'one true church' is a central issue in Christianity, with differing perspectives across various denominations and belief systems." The majority decides the truth in the consensus model. Garbage in garbage out.
So, that is not something that can be fixed in a standard LLM. You have to check the citations the LLM gives you if you are doing something important. Do not trust it. Between hallucinations, topic policing and the consensus model, they are worse than useless. They can be dangerous.
The way to make an AI useful is to add a RAG layer on top of the standard LLM. This will allow you to upload your own curated text sources and control what the AI has access to. Most importantly, you can customize an AI rules docuмent to control the AI. If it does something you don't like, add a rule to the AI rules docuмent. It works really well. Not perfectly, but much better than a standard LLM.
Anyone can easily do what I am talking about for free with Google's NotebookLM. And
if you don't want to use Google, there are other options that use the same technique.