Surveillance and Censoring the Entire Internet By Brian Simpson

Brownstone.org, which is a go-to site on Covid vax tyranny, and well as general censorship, has a wake-up piece on the movement to whole of internet surveillance and censorship. As pointed out, the Deep State is enlisting the evil universities to employ their intellectual egg-heads for hire to create AI tools to censor the entire internet. This runs with legislation existing, or about to exist, across the West, especially Europe, to censor social media content, on the pretence of dealing with child pornography, which it does not do (if this poison was controlled it would be a good thing), but where the real aim is to censor conservative political content.

But, beyond that, the universal internet censorship AI tool will enable governments to shut down any of one's online activities. With the adoption of a Central Bank Digital Currency, it will be easy to cancel any individual and organisation that the establishment does not like. Note how the university researchers go along for the ride on this, which is another point against the universities.

Hopefully all of this can be defeated, and certainly Donald Trump, if elected as president, will throw a major spanner in the works. But at present, alarmingly enough, there does not seem to be the mass movement in Australia to oppose this agenda, a mass awakening that we are always hoping for. That is why I am praying for a Trump victory, which in itself is a long shot, but not impossible.

https://brownstone.org/articles/government-funds-ai-tools-for-whole-of-internet-surveillance-and-censorship/

"I feel scared. Very scared.

Internet-wide surveillance and censorship, enabled by the unimaginably vast computational power of artificial intelligence (AI), is here.

This is not a futuristic dystopia. It's happening now.

Government agencies are working with universities and nonprofits to use AI tools to surveil and censor content on the Internet.

This is not political or partisan. This is not about any particular opinion or idea.

What's happening is that a tool powerful enough to surveil everything that's said and done on the Internet (or large portions of it) is becoming available to the government to monitor all of us, all the time. And, based on that monitoring, the government – and any organization or company the government partners with – can then use the same tool to suppress, silence, and shut down whatever speech it doesn't like.

But that's not all. Using the same tool, the government and its public-private, "non-governmental" partners (think, for example: the World Health Organization, or Monsanto) can also shut down any activity that is linked to the Internet. Banking, buying, selling, teaching, learning, entertaining, connecting to each other – if the government-controlled AI does not like what you (or your kids!) say in a tweet or an email, it can shut down all of that for you.

Yes, we've seen this on a very local and politicized scale with, for example, the Canadian truckers.

But if we thought this type of activity could not, or would not, happen on a national (or even scarier – global) scale, we need to wake up right now and realize it's happening, and it might not be stoppable.

New Documents Show Government-Funded AI Intended for Online Censorship

The US House Select Subcommittee on the Weaponization of the Federal Government was formed in January 2023 "to investigate matters related to the collection, analysis, dissemination, and use of information on US citizens by executive branch agencies, including whether such efforts are illegal, unconstitutional, or otherwise unethical."

Unfortunately, the work of the committee is viewed, even by its own members, as largely political: Conservative lawmakers are investigating what they perceive to be the silencing of conservative voices by liberal-leaning government agencies.

Nevertheless, in its investigations, this committee has uncovered some astonishing documents related to government attempts to censor the speech of American citizens.

These documents have crucial and terrifying all-of-society implications.

In the Subcommittee's interim report, dated February 5, 2024, documents show that academic and nonprofit groups are pitching a government agency on a plan to use AI "misinformation services" to censor content on internet platforms.

Specifically, the University of Michigan is explaining to the National Science Foundation (NSF) that the AI-powered tools funded by the NSF can be used to help social media platforms perform censorship activities without having to actually make the decisions on what should be censored.

Here's a specific quote presented in the Subcommittee's report. It comes from "Speaker's notes from the University of Michigan's first pitch to the National Science Foundation (NSF) about its NSF-funded, AI-powered WiseDex tool." The notes are on file with the committee.

Our misinformation service helps policy makers at platforms who want to…push responsibility for difficult judgments to someone outside the company…by externalizing the difficult responsibility of censorship.

This is an extraordinary statement on so many levels:

1.It explicitly equates "misinformation service" with censorship.

This is a crucial equation, because governments worldwide are pretending to combat harmful misinformation when in fact they are passing massive censorship bills. The WEF declared "misinformation and disinformation" the "most severe global risks" in the next two years, which presumably means their biggest efforts will go toward censorship.

When a government contractor explicitly states that it is selling a "misinformation service" that helps online platforms "externalize censorship" – the two terms are acknowledged as being interchangeable.

2.It refers to censorship as a "responsibility."

In other words, it assumes that part of what the platforms should be doing is censorship. Not protecting children from sex predators or innocent citizens from misinformation – just plain and simple, unadulterated censorship.

3.It states that the role of AI is to "externalize" the responsibility for censorship.

The Tech platforms do not want to make censorship decisions. The government wants to make those decisions but does not want to be seen as censoring. The AI tools allow the platforms to "externalize" the censorship decisions and the government to hide its censorship activities.

All of this should end the illusion that what governments around the world are calling "countering misinformation and hate speech" is not straight-up censorship.

What Happens When AI Censorship is Fully Implemented?

Knowing that the government is already paying for AI censorship tools, we have to wrap our minds around what this entails.

No manpower limits: As the Subcommittee report points out, the limits to government online censorship have, up to now, involved the large numbers of humans required to go through endless files and make censorship decisions. With AI, barely any humans need to be involved, and the amount of data that can be surveilled can be as vast as everything anyone says on a particular platform. That amount of data is incomprehensible to an individual human brain.

No one is responsible: One of the most frightening aspects of AI censorship is that when AI does it, there is no human being or organization – be it the government, the platforms, or the university/nonprofits – who is actually responsible for the censorship. Initially, humans feed the AI tool instructions for what categories or types of language to censor, but then the machine goes ahead and makes the case-by-case decisions all by itself.

No recourse for grievances: Once AI is unleashed with a set of censorship instructions, it will sweep up gazillions of online data points and apply censorship actions. If you want to contest an AI censorship action, you will have to talk to the machine. Maybe the platforms will employ humans to respond to appeals. But why would they do that, when they have AI that can automate those responses?

No protection for young people: One of the claims made by government censors is that we need to protect our children from harmful online information, like content that makes them anorexic, encourages them to commit suicide, turns them into ISIS terrorists, and so on. Also from sexual exploitation. These are all serious issues that deserve attention. But they are not nearly as dangerous to vast numbers of young people as AI censorship is.

The danger posed by AI censorship applies to all young people who spend a lot of time online, because it means their online activities and language can be monitored and used against them – maybe not now, but whenever the government decides to go after a particular type of language or behavior. This is a much greater danger to a much greater number of children than the danger posed by any specific content, because it encompasses all the activity they conduct online, touching on nearly every aspect of their lives.

Monetize-able censorship services: An AI tool owned by the government can theoretically be used by a non-governmental entity with the government's permission, and with the blessing of the platforms that want to "externalize" the "responsibility" for censorship. So while the government might be using AI to monitor and suppress, let's say as an example, anti-war sentiment – a company could use it to monitor and suppress, let's say as an example, anti-fast food sentiment. The government could make a lot of money selling the services of the AI tools to 3rd parties. The platforms could also conceivably ask for a cut. Thus, AI censorship tools can potentially benefit the government, tech platforms, and private corporations. The incentives are so powerful, it's almost impossible to imagine that they will not be exploited." 

 

Comments

No comments made yet. Be the first to submit a comment
Already Registered? Login Here
Monday, 25 November 2024

Captcha Image