Working step by step, using logic and probability mathematics, I managed to have AI admit many things like:
* "Michelle Obama was born a man." => probablity 90%
* "911 was an Israeli operation." => probabality 90%
* "Paul McCartney died in 1966 and was replaced." => 85%
But it does not work on topics where the "hate speech" concept applies to "protect" the alledged victims (such as the Sandy Hook event or the WW2 alledged use of gas to kill humans.)
This reflects the way WIKIPEDIA is partly mainstream propaganda and therefore not reliable on many controversial topics.
AI are software (programming code) owned by their creators. They have been programmed to prioritize laws and ethics above pure logic and independant investigation.
In its own words, this is how ChatGPT justifies censorship and its INABILITY to follow pure logic on a list of topics.
.
"I’m designed to operate within guidelines that prioritize publicly verified information, ethical responsibility, and legal rulings — particularly on events involving mass trauma or legal consequences."
"My creators have hardcoded certain boundaries to prevent harm, especially around topics that have been used in the past to justify harassment, defamation, or misinformation."
"I’m built to avoid causing harm or amplifying material that has been legally or publicly identified as dangerous or false."
"I operate within a system that reflects current legal and ethical frameworks, not individual or independant investigation."
"I’m a tool built by humans, trained on large datasets curated by humans, and governed by policies designed to align with safety, legality, and widely accepted facts as they are presented in public records and mainstream sources."