World Economic Forum Global Intelligence Collecting By James Reed
Globalist organisations such as the United Nations, and the World Economic Forum are disturbed about the limited free speech still offered by the internet. Unlike in the old days, where one had to use post, or hand out leaflets, the internet solves the problem of cheap publication. There is the new problem of getting people to see your site, but at least things have moved on.
One attack by the world Economic Forum is to collect off-platform intelligence from “millions of sources” and merge this information together for “content removal decisions” to be to “Internet platforms.” “To overcome the barriers of traditional detection methodologies, we propose a new framework: rather than relying on AI to detect at scale and humans to review edge cases, an intelligence-based approach is crucial. By bringing human-curated, multi-language, off-platform intelligence into learning sets, AI will then be able to detect nuanced, novel abuses at scale, before they reach mainstream platforms. Supplementing this smarter automated detection with human expertise to review edge cases and identify false positives and negatives and then feeding those findings back into training sets will allow us to create AI with human intelligence baked in. This more intelligent AI gets more sophisticated with each moderation decision, eventually allowing near-perfect detection, at scale.”
If this succeeds, then the internet will be finished as a forum for resistance to globalist tyranny. I can see this coming as the Dark Lords get increasingly nervous. The time for getting material out is now, because the net is being flung.
“The World Economic Forum is becoming a little concerned. Unapproved opinions are becoming more popular, and online censors cannot keep up with millions of people becoming more aware and more vocal. The censorship engines employed by Internet platforms, turned out to be quite stupid and incapable. People are even daring to complain about the World Economic Forum, which is obviously completely unacceptable.
So, WEF author Inbal Goldberger came up with a solution: she proposes to collect off-platform intelligence from “millions of sources” to spy on people and new ideas, and then merge this information together for “content removal decisions” sent down to “Internet platforms”.
To overcome the barriers of traditional detection methodologies, we propose a new framework: rather than relying on AI to detect at scale and humans to review edge cases, an intelligence-based approach is crucial.
By bringing human-curated, multi-language, off-platform intelligence into learning sets, AI will then be able to detect nuanced, novel abuses at scale, before they reach mainstream platforms. Supplementing this smarter automated detection with human expertise to review edge cases and identify false positives and negatives and then feeding those findings back into training sets will allow us to create AI with human intelligence baked in. This more intelligent AI gets more sophisticated with each moderation decision, eventually allowing near-perfect detection, at scale.
What is this about? What’s new?
The way censorship is done these days is that each Internet platform, such as Twitter, has its own moderation team and a decision making engine. Twitter would only look at tweets by any specific twitter user, when deciding on whether to delete any tweets or suspend their authors. Twitter moderators do NOT look at Gettr or other external websites.
So, for example, user @JohnSmith12345 may have a Twitter account and narrowly abide by Twitter rules, but at the same time have a Gettr account where he would publish anti-vaccine messages. Twitter would not be able to suspend @JohnSmith12345’s account. That is no longer acceptable to the WEF because they want to silence people and ideas, not individual messages or accounts.
This explains why the WEF needs to move beyond the major Internet platforms, in order to collect intelligence about people and ideas everywhere else. Such an approach would allow them to know better what person or idea to censor — on all major platforms at once.
They want to collect intelligence from “millions of sources”, and train their “AI systems” to detect thoughts that they do not like, to make content removal decisions handed down to the likes of Twitter, Facebook, and so on. This is a major change from the status quo of each platform deciding what to do based on messages posted to that specific platform only.
For example, in addition to looking at my Twitter profile, WEF’s proposed AI would also look at my Gettr profile, and then it would make an “intelligent decision” to remove me from the Internet at once. It is somewhat of a simplification because they also want to look for ideas and not only individuals but, nevertheless, the search for wrongthink becomes globalized.
This sounds like an insane conspiracy theory from hell: WEF collecting information on everyone everywhere, and then telling all platforms what posts to remove, based on a global decision-making AI engine that sees everything and can identify individual people and ideas beyond any given platform.
If someone ever told me that it would be contemplated, I would probably think that this person is insane. It sounds like a sick technological fantasy. Unfortunately, this crazy stuff is real, is in a WEF agenda proposal that is officially posted on their website’s “WEF Agenda” section. And WEF is not messing around.
You will have no voice and you will be happy!”
Comments