“THE THOUGHT POLICE”
Today from nationwide military bases, Federal, state and local police departments are unified and operating from high-tech Satellite / Drone division. As an example in Los Angeles, California the LAPD Real-time Assessment and Critical Response Division (RACR) is a nationwide model connecting a unified, fused paradigm, i.e., electromagnetic frequency technological paradigm with massive military personnel human experimentation on civilians. The protest in Los Angeles, one of the United States most infamous cities by the “Stop the Spying Coalition” protest does not reveal the deepest level of what is technologically happening today across the USA and globally through “Fusion Center” state of the art operation centers powerful government agency leadership. Again, this is Global!
FUSION CENTERS ACROSS THE USA
AND, unpublicized mind invasive psychotronic technologies, mind-reading, etc., etc., etc…
Frankly, with advanced technology in hand, those at the helm have proven to be little more than borderline sociopath tendencies and many should not be in the same room with mind invasive, frequency subliminal influencing frequency, beamed assaults, and tracking technologies.
People have on average 6,000 thoughts and many random each day and many are not acted on and merely passing thoughts or a brief emotional response. This is unless Targeted by the “Thought Police” from state-of-the-art operation centers today two to three in each state across the USA who are not above also implanting thoughts, to create the reality they seek, with subliminal influence, using patented beam mind invasive technologies around for DECADES and it done effortlessly and without leaving the building.
THE GLOBAL MASS POPULATION CONTROL CONNECTION
“The group is organizing a wide base of mathematicians in the hopes of cutting off police from using such technologies. The letter’s authors cite “deep concerns over the use of machine learning, AI, and facial recognition technologies to justify and perpetuate oppression.”
ELECTROMAGNETIC FREQUENCIES AND DNA, A FRACTAL ANTENNA
VIDEO: DR. DAVID MARTIN EXPOSES THE FALSE
FOUNDATION OF EUGENICS: “YOU DON’T HAVE DNA”
Artificial intelligence (AI) technology is rapidly proliferating around the world. Startling developments keep emerging, from the onset of deep fake videos that blur the line between truth and falsehood, to advanced algorithms that can beat the best players in the world in multiplayer poker. Businesses harness AI capabilities to improve analytic processing; city officials tap AI to monitor traffic congestion and oversee smart energy metering. Yet a growing number of states are deploying advanced AI surveillance tools to monitor, track, and surveil citizens to accomplish a range of policy objectives—some lawful, others that violate human rights, and many of which fall into a murky middle ground.
In order to appropriately address the effects of this technology, it is important to first understand where these tools are being deployed and how they are being used. Unfortunately, such information is scarce. To provide greater clarity, this paper presents an AI Global Surveillance (AIGS) Index—representing one of the first research efforts of its kind. The index compiles empirical data on AI surveillance use for 176 countries around the world. It does not distinguish between legitimate and unlawful uses of AI surveillance. Rather, the purpose of the research is to show how new surveillance capabilities are transforming the ability of governments to monitor and track individuals or systems. It specifically asks:
- Which countries are adopting AI surveillance technology?
- What specific types of AI surveillance are governments deploying?
- Which countries and companies are supplying this technology?
- AI surveillance technology is spreading at a faster rate to a wider range of countries than experts have commonly understood. At least seventy-five out of 176 countries globally are actively using AI technologies for surveillance purposes. This includes: smart city/safe city platforms (fifty-six countries), facial recognition systems (sixty-four countries), and smart policing (fifty-two countries).
- China is a major driver of AI surveillance worldwide. Technology linked to Chinese companies—particularly Huawei, Hikvision, Dahua, and ZTE—supply AI surveillance technology in sixty-three countries, thirty-six of which have signed onto China’s Belt and Road Initiative (BRI). Huawei alone is responsible for providing AI surveillance technology to at least fifty countries worldwide. No other company comes close. The next largest non-Chinese supplier of AI surveillance tech is Japan’s NEC Corporation (fourteen countries).
- Chinese product pitches are often accompanied by soft loans to encourage governments to purchase their equipment. These tactics are particularly relevant in countries like Kenya, Laos, Mongolia, Uganda, and Uzbekistan—which otherwise might not access this technology. This raises troubling questions about the extent to which the Chinese government is subsidizing the purchase of advanced repressive technology.
- But China is not the only country supplying advanced surveillance tech worldwide. U.S. companies are also active in this space. AI surveillance technology supplied by U.S. firms is present in thirty-two countries. The most significant U.S. companies are IBM (eleven countries), Palantir (nine countries), and Cisco (six countries). Other companies based in liberal democracies—France, Germany, Israel, Japan—are also playing important roles in proliferating this technology. Democracies are not taking adequate steps to monitor and control the spread of sophisticated technologies linked to a range of violations.
- Liberal democracies are major users of AI surveillance. The index shows that 51 percent of advanced democracies deploy AI surveillance systems. In contrast, 37 percent of closed autocratic states, 41 percent of electoral autocratic/competitive autocratic states, and 41 percent of electoral democracies/illiberal democracies deploy AI surveillance technology.1 Governments in full democracies are deploying a range of surveillance technology, from safe city platforms to facial recognition cameras. This does not inevitably mean that democracies are abusing these systems. The most important factor determining whether governments will deploy this technology for repressive purposes is the quality of their governance.
- Governments in autocratic and semi-autocratic countries are more prone to abuse AI surveillance than governments in liberal democracies. Some autocratic governments—for example, China, Russia, Saudi Arabia—are exploiting AI technology for mass surveillance purposes. Other governments with dismal human rights records are exploiting AI surveillance in more limited ways to reinforce repression. Yet all political contexts run the risk of unlawfully exploiting AI surveillance technology to obtain certain political objectives.
- There is a strong relationship between a country’s military expenditures and a government’s use of AI surveillance systems: forty of the world’s top fifty military spending countries (based on cumulative military expenditures) also use AI surveillance technology.2
- The “Freedom on the Net 2018” report identified eighteen countries out of sixty-five that had accessed AI surveillance technology developed by Chinese companies.3 The AIGS Index shows that the number of those countries accessing Chinese AI surveillance technology has risen to forty-seven out of sixty-five countries in 2019.
- The AIGS Index presents a country-by-country snapshot of AI tech surveillance with the majority of sources falling between 2017 and 2019. Given the opacity of government surveillance use, it is nearly impossible to pin down by specific year which AI platforms or systems are currently in use.
- The AIGS Index uses the same list of independent states included in the Varieties of Democracy (V-Dem) project with two exceptions, totaling 176.4 The V-Dem country list includes all independent polities worldwide but excludes microstates with populations below 250,000.
- The AIGS Index does not present a complete list of AI surveillance companies operating in particular countries. The paper uses open source reporting and content analysis to derive its findings. Accordingly, there are certain built-in limitations. Some companies, such as Huawei, may have an incentive to highlight new capabilities in this field. Other companies have opted to downplay their association with surveillance technology and have purposely kept documents out of the public domain.
A full version of the index can be accessed online here:
An interactive map keyed to the index that visually depicts the global spread of AI surveillance technology can be accessed here:
All reference source material used to build the index has been compiled into an open Zotero library. It is available here:
INTRODUCING THE AI GLOBAL SURVEILLANCE (AIGS) INDEX
AI technology was once relegated to the world of science fiction, but today it surrounds us. It powers our smartphones, curates our music preferences, and guides our social media feeds. Perhaps the most notable aspect of AI is its sudden ubiquity.
In general terms, the goal of artificial intelligence is to “make machines intelligent” by automating or replicating behavior that “enables an entity to function appropriately and with foresight in its environment,” according to computer scientist Nils Nilsson.5 AI is not one specific technology. Instead, it is more accurate to think of AI as an integrated system that incorporates information acquisition objectives, logical reasoning principles, and self-correction capacities. An important AI subfield is machine learning, which is a statistical process that analyzes a large amount of information in order to discern a pattern to explain the current data and predict future uses.6 Several breakthroughs are making new achievements in the field possible: the maturation of machine learning and the onset of deep learning; cloud computing and online data gathering; a new generation of advanced microchips and computer hardware; improved performance of complex algorithms; and market-driven incentives for new uses of AI technology.7
Unsurprisingly, AI’s impact extends well beyond individual consumer choices. It is starting to transform basic patterns of governance, not only by providing governments with unprecedented capabilities to monitor their citizens and shape their choices but also by giving them new capacity to disrupt elections, elevate false information, and delegitimize democratic discourse across borders.
The focus of this paper is on AI surveillance and the specific ways governments are harnessing a multitude of tools—from facial recognition systems and big data platforms to predictive policing algorithms—to advance their political goals. Crucially, the index does not distinguish between AI surveillance used for legitimate purposes and unlawful digital surveillance. Rather, the purpose of the research is to shine a light on new surveillance capabilities that are transforming the ability of states—from autocracies to advanced democracies—to keep watch on individuals.
The AIGS Index provides a detailed empirical picture of global AI surveillance trends and describes how governments worldwide are using this technology. It addresses three primary questions:
- Which countries are adopting AI surveillance technology?
- What specific types of AI surveillance are governments deploying?
- Which countries and companies are supplying this technology?
The AIGS Index is contained in Appendix 1. It includes detailed information for seventy-five countries where research indicates governments are deploying AI surveillance technology. The index breaks down AI surveillance tools into the following subcategories: 1) smart city/safe city, 2) facial recognition systems, and 3) smart policing. A full version of the index can be accessed online at
An interactive map keyed to the index that visually depicts the global spread of AI surveillance technology can be accessed at
All reference source material used to build the index has been compiled into an open Zotero library. It is available at
The majority of sources referenced by the index occur between 2017 and 2019. A small number of sources date as far back as 2012. The index uses the same list of countries found in the Varieties of Democracy (V-Dem) project with two minor exceptions.8 The V-Dem country list includes all independent polities worldwide but excludes microstates with populations below 250,000. The research collection effort combed through open-source material, country by country, in English and other languages, including news articles, websites, corporate documents, academic articles, NGO reports, expert submissions, and other public sources. It relied on systematic content analysis for each country incorporating multiple sources to determine the presence of relevant AI surveillance technology and corresponding companies. Sources were categorized into tiered levels of reliability and accuracy. First-tier sources include major print and news magazine outlets (such as the New York Times, Economist, Financial Times, and Wall Street Journal). Second-tier sources include major national media outlets. Third-tier sources include web articles, blog posts, and other less substantiated sourcing; these were only included after multiple corroboration.
Given limited resources and staffing constraints (one full-time researcher plus volunteer research assistance), the index is only able to offer a snapshot of AI surveillance levels in a given country. It does not provide a comprehensive assessment of all relevant technology, government surveillance uses, and applicable companies. Because research relied primarily on content analysis and literature reviews to derive its findings, there are certain built-in limitations. Some companies, such as Huawei, may have an incentive to highlight new capabilities in this field. Other companies may wish to downplay links to surveillance technology and purposely keep documents out of the public domain.
Field-based research involving on-the-ground information collection and verification would be useful to undertake. A number of countries—such as Angola, Azerbaijan, Belarus, Hungary, Peru, Sri Lanka, Tunisia, and Turkmenistan—provided circumstantial or anecdotal evidence of AI surveillance, but not enough verifiable data to warrant inclusion in the index.
A major difficulty was determining which AI technologies should be included in the index. AI technologies that directly support surveillance objectives—smart city/safe city platforms, facial recognition systems, smart policing systems—are included in the index. Enabling technologies that are critical to AI functioning but not directly responsible for surveillance programs are not included in the index.
Another data collection challenge is that governments (and many companies) purposely hide their surveillance capabilities. As such, it is difficult to precisely determine the extent to which states are deploying algorithms to support their surveillance objectives, or whether AI use is more speculative than real.
The index does not differentiate between governments that expansively deploy AI surveillance techniques versus those that use AI surveillance to a much lesser degree (for example, the index does not include a standardized interval scale correlating to levels of AI surveillance). This is by design. Because this is a nascent field and there is scant information about how different countries are using AI surveillance techniques, attempting to score a country’s relative use of AI surveillance would introduce a significant level of researcher bias. Instead, a basic variable was used: is there documented presence of AI surveillance in a given country? If so, what types of AI surveillance technology is the state deploying? Future research may be able to assess and analyze levels of AI surveillance on a cross-comparative basis.
Finally, instances of AI surveillance documented in the index are not specifically tied to harmful outcomes. The index does not differentiate between unlawful and legitimate surveillance. In part, this is because it is exceedingly difficult to determine what specifically governments are doing in the surveillance realm and what the associated impacts are; there is too much that is unknown and hidden.
FINDINGS AND THREE KEY INSIGHTS
The findings indicate that at least seventy-five out of 176 countries globally are actively using AI technologies for surveillance purposes. This includes: smart city/safe city platforms (fifty-six countries), facial recognition systems (sixty-four countries), and smart policing (fifty-two countries). Three key insights emerge from the AIGS Index’s findings.
First, global adoption of AI surveillance is increasing at a rapid pace around the world. Seventy-five countries, representing 43 percent of total countries assessed, are deploying AI-powered surveillance in both lawful and unlawful ways. The pool of countries is heterogeneous—they come from all regions, and their political systems range from closed autocracies to advanced democracies. The “Freedom on the Net 2018” report raised eyebrows when it reported that eighteen out of sixty-five assessed countries were using AI surveillance technology from Chinese companies.9 The report’s assessment period ran from June 1, 2017 to May 31, 2018. One year later, the AIGS Index finds that forty-seven countries out of that same group are now deploying AI surveillance technology from China.
Unsurprisingly, countries with authoritarian systems and low levels of political rights are investing heavily in AI surveillance techniques. Many governments in the Gulf, East Asia, and South/Central Asia are procuring advanced analytic systems, facial recognition cameras, and sophisticated monitoring capabilities. But liberal democracies in Europe are also racing ahead to install automated border controls, predictive policing, safe cities, and facial recognition systems. In fact, it is striking how many safe city surveillance case studies posted on Huawei’s website relate to municipalities in Germany, Italy, the Netherlands, and Spain.
Regionally, there are clear disparities. The East Asia/Pacific and the Middle East/North Africa regions are robust adopters of these tools. South and Central Asia and the Americas also demonstrate sizable take-up of AI surveillance instruments. Sub-Saharan Africa is a laggard—less than one-quarter of its countries are invested in AI surveillance. Most likely this is due to technological underdevelopment (African countries are struggling to extend broadband access to their populations; the region has eighteen of twenty countries with the lowest levels of internet penetration).10 Given the aggressiveness of Chinese companies to penetrate African markets via BRI, these numbers will likely rise in the coming years. Figure 1 shows the percentage breakdown by region of countries adopting AI surveillance.
Second, China is a major supplier of AI surveillance. Technology linked to Chinese companies are found in at least sixty-three countries worldwide. Huawei alone is responsible for providing AI surveillance technology to at least fifty countries. There is also considerable overlap between China’s Belt and Road Initiative and AI surveillance—thirty-six out of eighty-six BRI countries also contain significant AI surveillance technology. However, China is not the only country supplying advanced surveillance technology. France, Germany, Japan, and the United States are also major players in this sector. U.S. companies, for example, have an active presence in thirty-two countries. Figure 2 breaks down the leading companies in the sector.
Third, liberal democracies are major users of AI surveillance. The index shows that 51 percent of advanced democracies deploy AI surveillance systems. In contrast, 37 percent of closed autocratic states, 41 percent of electoral autocratic/competitive autocratic states, and 41 percent of electoral democracies/illiberal democracies deploy AI surveillance technology. Liberal democratic governments are aggressively using AI tools to police borders, apprehend potential criminals, monitor citizens for bad behavior, and pull out suspected terrorists from crowds. This doesn’t necessarily mean that democracies are using this technology unlawfully. The most important factor determining whether governments will exploit this technology for repressive purposes is the quality of their governance—is there an existing pattern of human rights violations? Are there strong rule of law traditions and independent institutions of accountability? That should provide a measure of reassurance for citizens residing in democratic states.
But advanced democracies are struggling to balance security interests with civil liberties protections. In the United States, increasing numbers of cities have adopted advanced surveillance systems. A 2016 investigation by Axios’s Kim Hart revealed, for example, that the Baltimore police had secretly deployed aerial drones to carry out daily surveillance over the city’s residents: “From a plane flying overhead, powerful cameras capture aerial images of the entire city. Photos are snapped every second, and the plane can be circling the city for up to 10 hours a day.”11 Baltimore’s police also deployed facial recognition cameras to monitor and arrest protesters, particularly during 2018 riots in the city.12 The ACLU condemned these techniques as the “technological equivalent of putting an ankle GPS [Global Positioning Service] monitor on every person in Baltimore.”13
On the U.S.-Mexico border, an array of hi-tech companies also purvey advanced surveillance equipment. Israeli defense contractor Elbit Systems has built “dozens of towers in Arizona to spot people as far as 7.5 miles away,” writes the Guardian’s Olivia Solon. Its technology was first perfected in Israel from a contract to build a “smart fence” to separate Jerusalem from the West Bank. Another company, Anduril Industries, “has developed towers that feature a laser-enhanced camera, radar and a communications system” that scans a two-mile radius to detect motion. Captured images “are analysed using artificial intelligence to pick out humans from wildlife and other moving objects.”14 It is unclear to what extent these surveillance deployments are covered in U.S. law, let alone whether these actions meet the necessity and proportionality standard.
The United States is not the only democracy embracing AI surveillance. In France, the port city of Marseille initiated a partnership with ZTE in 2016 to establish the Big Data of Public Tranquility project. The goal of the program is to reduce crime by establishing a vast public surveillance network featuring an intelligence operations center and nearly one thousand intelligent closed-circuit television (CCTV) cameras (the number will double by 2020). Local authorities trumpet that this system will make Marseille “the first ‘safe city’ of France and Europe.”15 Similarly, in 2017, Huawei “gifted” a showcase surveillance system to the northern French town of Valenciennes to demonstrate its safe city model. The package included upgraded high definition CCTV surveillance and an intelligent command center powered by algorithms to detect unusual movements and crowd formations.16
The fact that so many democracies—as well as autocracies—are taking up this technology means that regime type is a poor predictor for determining which countries will adopt AI surveillance.
A better predictor for whether a government will procure this technology is related to its military spending. A breakdown of military expenditures in 2018 shows that forty of the top fifty military spending countries also have AI surveillance technology.17 These countries span from full democracies to dictatorial regimes (and everything in between). They comprise leading economies like France, Germany, Japan, and South Korea, and poorer states like Pakistan and Oman. This finding is not altogether unexpected; countries with substantial investments in their militaries tend to have higher economic and technological capacities as well as specific threats of concern. If a country takes its security seriously and is willing to invest considerable resources in maintaining robust military-security capabilities, then it should come as little surprise that the country will seek the latest AI tools. The motivations for why European democracies acquire AI surveillance (controlling migration, tracking terrorist threats) may differ from Egypt or Kazakhstan’s interests (keeping a lid on internal dissent, cracking down on activist movements before they reach critical mass), but the instruments are remarkably similar. Future research might examine country-level internal security figures and compare them to levels of AI surveillance.
DISTINGUISHING BETWEEN LEGITIMATE AND UNLAWFUL SURVEILLANCE
State surveillance is not inherently unlawful. Governments have legitimate reasons to undertake surveillance that is not rooted in a desire to enforce political repression and limit individual freedoms. For example, tracking tools play a vital role in preventing terrorism. They help security forces deter bad acts and resolve problematic cases. They give authorities the ability to monitor critical threats and react accordingly. But technology has changed the nature of how governments carry out surveillance and what they choose to monitor. The internet has proliferated the amount of transactional data or “metadata” available about individuals, such as information about sent and received emails, location identification, web-tracking, and other online activities. As former UN special rapporteur Frank La Rue noted in a milestone 2013 surveillance report:
Communications data are storable, accessible and searchable, and their disclosure to and use by State authorities are largely unregulated. Analysis of this data can be both highly revelatory and invasive, particularly when data is combined and aggregated. As such, States are increasingly drawing on communications data to support law enforcement or national security investigations. States are also compelling the preservation and retention of communication data to enable them to conduct historical surveillance.18
It goes without saying that such intrusions profoundly affect an individual’s right to privacy—to not be subjected to what the Office of the UN High Commissioner for Human Rights (OHCHR) called “arbitrary or unlawful interference with his or her privacy, family, home or correspondence.”19 Surveillance likewise may infringe upon an individual’s right to freedom of association and expression. Under international human rights law, three principles are critical to assessing the lawfulness of a particular surveillance action.
First, does domestic law allow for surveillance? La Rue’s successor, David Kaye, issued a report in 2019 that affirmed that legal regulations should be “formulated with sufficient precision to enable an individual to regulate his or her conduct accordingly and it must be made accessible to the public.” Legal requirements should not be “vague or overbroad,” which would allow unconstrained discretion to government officials. The legal framework itself should be “publicly accessible, clear, precise, comprehensive and non-discriminatory.”20
Second, does the surveillance action meet the “necessity and proportionality” international legal standard, which restricts surveillance to situations that are “strictly and demonstrably necessary to achieve a legitimate aim”?21
Third, are the interests justifying the surveillance action legitimate? Disagreements abound when it comes to determining what constitutes legitimate surveillance and what is an abuse of power. While governments commonly justify surveillance on national security or public order grounds, the OHCHR warns that such restrictions may “unjustifiably or arbitrarily” restrict citizens’ rights to freedom of opinion and expression. It contends that legitimate surveillance requires states to “demonstrate the risk that specific expression poses to a definite interest in national security or public order,” and that a “robust, independent oversight system” that entrusts judiciaries to authorize relevant surveillance measures and provide remedies in cases of abuse is required.22 Kaye adds that legitimate surveillance should only apply when the interest of a “whole nation is at stake,” and should exclude surveillance carried out “in the sole interest of a Government, regime or power group.”23
The legal standards required to legitimately carry out surveillance are high, and governments struggle to meet them. Even democracies with strong rule of law traditions and robust oversight institutions frequently fail to adequately protect individual rights in their surveillance programs. Countries with weak legal enforcement or authoritarian systems “routinely shirk these obligations.”24 As the OHCHR’s inaugural report on privacy in the digital age concludes, states with “a lack of adequate national legislation and/or enforcement, weak procedural safeguards and ineffective oversight” bring reduced accountability and heightened conditions for unlawful digital surveillance.25
AI surveillance exacerbates these conditions and makes it likelier that democratic and authoritarian governments may carry out surveillance that contravenes international human rights standards. Frank La Rue explains: “Technological advancements mean that the State’s effectiveness in conducting surveillance is no longer limited by scale or duration. Declining costs of technology and data storage have eradicated financial or practical disincentives to conducting surveillance. As such, the State now has a greater capability to conduct simultaneous, invasive, targeted and broad-scale surveillance than ever before.”26
AI surveillance in particular offers governments two major capabilities. One, AI surveillance allows regimes to automate many tracking and monitoring functions formerly delegated to human operators. This brings cost efficiencies, decreases reliance on security forces, and overrides potential principal-agent loyalty problems (where the very forces operating at the behest of the regime decide to seize power for themselves).
Two, AI technology can cast a much wider surveillance net than traditional methods. Unlike human operatives “with limited reserves of time and attention,” AI systems never tire or fatigue.27 As a result, this creates a substantial “chilling effect” even without resorting to physical violence; citizens never know if an automated bot is monitoring their text messages, reading their social media posts, or geotracking their movements around town.28
This paper recognizes that AI surveillance technology is “value neutral.” In and of themselves, these tools do not foment repression, and their presence does not mean that a government is using them for antidemocratic purposes. The index does not specify, country-by-country, whether these instruments are being used by governments in lawful or illegitimate manners. Rather, the purpose of the index is to identify which countries possess sufficiently advanced tools that allow them to pursue a range of surveillance objectives.
HOW MUCH IS CHINA DRIVING THE SPREAD OF AI SURVEILLANCE?
Empirically, the AIGS Index shows that Chinese companies—led by Huawei—are leading suppliers of AI surveillance around the world. Overall, China is making a sustained push for leadership and primacy in AI.29 A growing consensus singles out China as a global driver of “authoritarian tech.” Experts claim that Chinese companies are working directly with Chinese state authorities to export “authoritarian tech” to like-minded governments in order to spread influence and promote an alternative governance model.30 But is this accurate?
There is some truth to this argument—a subset of Chinese exports goes directly to countries like Zimbabwe and Venezuela that are gross human rights violators and which would otherwise be unable to access such technology. But AI surveillance is not solely going from one authoritarian country (China) to other authoritarian states. Rather, transfers are happening in a much more heterogeneous fashion. China is exporting surveillance tech to liberal democracies as much as it is targeting authoritarian markets. Likewise, companies based in liberal democracies (for example, Germany, France, Israel, Japan, South Korea, the UK, the United States) are actively selling sophisticated equipment to unsavory regimes.
Saudi Arabia is a good case in point. Huawei is helping the government build safe cities, but Google is establishing cloud servers, UK arms manufacturer BAE has sold mass surveillance systems, NEC is vending facial recognition cameras, and Amazon and Alibaba both have cloud computing centers in Saudi Arabia and may support a major smart city project.31 The index shows that repressive countries rarely procure such technology from a single source. In Thailand, government officials repeatedly emphasized the importance of “foreign policy balancing” and not affiliating too strongly with any one side: “Always been that way. That’s why we’re still a kingdom. We compromise, we negotiate, and we balance.”32
That being said, there are special reasons why experts are applying greater scrutiny to Chinese companies. Huawei is the leading vendor of advanced surveillance systems worldwide by a huge factor. Its technology is linked to more countries in the index than any other company. It is aggressively seeking new markets in regions like sub-Saharan Africa. Huawei is not only providing advanced equipment but also offering ongoing technological support to set up, operate, and manage these systems.
A recent investigative report by the Wall Street Journal provides an eye-opening example. The reporters found that Huawei technicians in both Uganda and Zambia helped government officials spy on political opponents. This included “intercepting their encrypted communications and social media, and using cell data to track their whereabouts.” Not only did Huawei employees play a “direct role in government efforts to intercept the private communications of opponents,” but they also encouraged Ugandan security officials to travel to Algeria so they could study Huawei’s “intelligent video surveillance system” operating in Algiers.33 Uganda subsequently agreed to purchase a similar facial recognition surveillance system from Huawei costing $126 million.34
The Australian Strategic Policy Institute’s project on Mapping China’s Tech Giants indicates that Huawei is responsible for seventy-five “smart city-public security projects,” and has seen a colossal increase in its business line: “In 2017, Huawei listed 40 countries where its smart-city technologies had been introduced; in 2018, that reach had reportedly more than doubled to 90 countries (including 230 cities).”35 Huawei is directly pitching the safe city model to national security agencies, and China’s Exim Bank appears to be sweetening the deal with subsidized loans. The result is that a country like Mauritius obtains long-term financing from the Chinese government, which mandates contracting with Chinese firms.36The Mauritian government then turns to Huawei as the prime contractor or sub-awardee to set up the safe city and implement advanced surveillance controls.
It is also increasingly clear that firms such as Huawei operate with far less independence from the Chinese government than they claim. Huawei was founded in 1987 by Ren Zhengfei, a former officer in the People’s Liberation Army who served in its “military technology division,” Anna Fifield at the Washington Post has noted.37 There are consistent reports that Huawei receives significant subsidies from the Chinese government.38 There also appear to be strong connections between Huawei’s leadership and China’s security and intelligence apparatus. Sun Yafang, for example, chairwoman of Huawei’s board from 1999 to 2018, once worked in China’s Ministry of State Security.39 Max Chafkin and Joshua Brustein reported in Bloomberg Businessweek that there are allegations that Ren may have been a “high-ranking Chinese spymaster and indeed may still be.”40 Experts maintain that the Chinese Communist Party increasingly is establishing “party ‘cells’ in private companies to enable enhanced access and control.41 Huawei has publicly averred that it would “definitely say no” to any demands by the Chinese government to hand over user data.42 But this contravenes a 2015 Chinese national security law that mandates companies to allow third-party access to their networks and to turn over source code or encryption keys upon request.43 Huawei’s declared ownership structure is remarkably opaque. A recent academic study by Christopher Balding and Donald C. Clarke concluded that 99 percent of Huawei shares are controlled by a “trade union committee,” which in all likelihood is a proxy for Chinese state control of the company.”44
Even if Chinese companies are making a greater push to sell advanced surveillance tech, the issue of intentionality remains perplexing—to what extent are Chinese firms like Huawei and ZTE operating out of their own economic self-interest when peddling surveillance technology versus carrying out the bidding of the Chinese state? At least in Thailand, recent research interviews did not turn up indications that Chinese companies are pushing a concerted agenda to peddle advanced AI surveillance equipment or encourage the government to build sophisticated monitoring systems. An official from Thailand’s Ministry of Interior noted that while AI technology is “out there” and something the government is thinking more about, “China hasn’t offered any AI. It doesn’t give AI—Thais have to ask.”45The smart city/safe city model also garnered skepticism. Somkiat Tangkitvanich, a leading technology expert in Thailand, commented, “the idea of a smart city is a joke.” He relayed a recent conversation he had with Thailand’s information and communications technologies (ICT) minister: “He [the minister] boasted about the smart city in Phuket. . . . He told me that we are thinking about giving wristbands to tourists so that we can track them, we can help them. Something like that. But it’s not really implemented. Smart city in Phuket turns out to be providing free Wi-Fi and internet to tourists!”46 This serves as a useful reminder that more on-the-ground research is needed to separate hyperbole from fact in this area.
TYPES OF AI SURVEILLANCE
The following sections will describe key AI surveillance techniques and how governments worldwide are deploying them to support specific policy objectives.
States use AI technology to accomplish a broad range of surveillance goals. This section details three primary AI surveillance tools incorporated in the AIGS Index: smart city/safe city platforms, facial recognition systems, and smart policing. It also describes enabling technologies—such as cloud computing and Internet of Things (IOT) networks—that are integral for AI surveillance tools to function. Enabling technologies are not incorporated in the index.
Importantly, AI surveillance is not a standalone instrument of repression. It forms part of a suite of digital repression tools—information and communications technologies used to surveil, intimidate, coerce, and harass opponents in order to inflict a penalty on a target and deter specific activities or beliefs that challenge the state.47 (See Appendix 2 for more information.) Table 1 summarizes each technique and its corresponding level of global deployment.
SMART CITIES/SAFE CITIES
The World Bank describes smart cities as “technology-intensive” urban centers featuring an array of sensors that gather information in real-time from “thousands of interconnected devices” in order to facilitate improved service delivery and city management.48 They help municipal authorities manage traffic congestion, direct emergency vehicles to needed locations, foster sustainable energy use, and streamline administrative processes. But there is growing concern that smart cities are also enabling a dramatic increase in public surveillance and intrusive security capabilities. IBM, one of the original coiners of the term, designed a brain-like municipal model where information relevant to city operations could be centrally processed and analyzed.49 A key component of IBM’s smart city is public safety, which incorporates an array of sensors, tracking devices, and surveillance technology to increase police and security force capabilities.
Huawei has been up-front about trumpeting public safety technologies for smart cities. It is marketing “safe cities” to law enforcement communities to “predict, prevent, and reduce crime” and “address new and emerging threats.”50 In a 2016 white paper, Huawei describes a “suite of technology that includes video surveillance, emergent video communication, integrated incident command and control, big data, mobile, and secured public safety cloud” to support local law enforcement and policing as well as the justice and corrections system.51 Huawei explicitly links its safe city technology to confronting regional security challenges, noting that in the Middle East, its platforms can prevent “extremism”; in Latin America, safe cities enable governments to reduce crime; and that in North America, its technology will help the United States advance “counterextremism” programs.52
How do these platforms work in practice to advance surveillance goals? The IT firm Gartner, which partners with Microsoft on smart cities, provides an example:
Saudi Arabia’s Makkah Region Development Authority (MRDA) created a crowd-control system to increase safety and security of Hajj pilgrims. Data is collected via a wristband embedding identity information, special healthcare requirements and a GPS. In addition, surveillance cameras are installed to collect and analyze real-time video along the Al Mashaaer Al Mugaddassah Metro Southern Line (MMMSL), as well as in the holy sites, such as Great Mosque of Mecca, Mount Arafat, Jamarat and Mina.53
Unsurprisingly, such systems lend themselves to improper use. Recently, Huawei’s safe city project in Serbia, which intends to install 1,000 high-definition (HD) cameras with facial recognition and license plate recognition capabilities in 800 locations across Belgrade, sparked national outrage.54 Huawei posted a case study (since removed) about the benefits of safe cities and described how similar surveillance technology had facilitated the apprehension of a Serbian hit-and-run perpetrator who had fled the country to a city in China: “Based on images provided by Serbian police, the . . . [local] Public Security Bureau made an arrest within three days using new technologies.”55 Rather than applaud the efficiency of the system, Serbian commentators observed that in a country racked by endemic corruption and encroaching authoritarianism, such technology offers a powerful tool for Serbian authorities to curb dissent and perpetrate abuses.
Smart city platforms with a direct public security link are found in at least fifty-six of seventy-five countries with AI surveillance technology.
FACIAL RECOGNITION SYSTEMS
Facial recognition is a biometric technology that uses cameras—both video or still images—to match stored or live footage of individuals with images from a database. Not all facial recognition systems focus on individual identification via database matching. Some systems are designed to assess aggregate demographic trends or to conduct broader sentiment analysis via facial recognition crowd scanning.
Unlike ordinary CCTV, which has been a mainstay of police forces for twenty-five years, facial recognition cameras are much more intrusive. They can scan distinctive facial features in order to create detailed biometric maps of individuals without obtaining consent. Often facial recognition surveillance cameras are mobile and concealable. For example, security forces in Malaysia have entered into a partnership with the Chinese tech company Yitu to equip officers with facial recognition body cameras. This will allow security officials to “rapidly compare images caught by live body cameras with images from a central database.”56
Huawei is a major purveyor of facial recognition video surveillance, particularly as part of its safe city platforms. It describes the technology’s benefits in the Kenya Safe City project:
As part of this project, Huawei deployed 1,800 HD cameras and 200 HD traffic surveillance systems across the country’s capital city, Nairobi. A national police command center supporting over 9,000 police officers and 195 police stations was established to achieve monitoring and case-solving. The system worked during Pope Francis’ visit to Kenya in 2015, where more than eight million people welcomed his arrival. With Huawei’s HD video surveillance and a visualized integrated command solution, the efficiency of policing efforts as well as detention rates rose significantly.57
Experts detail several concerns associated with facial recognition.
First, few rules govern access to and the use of image databases (repositories that store captured images from facial recognition cameras). How governments use this information, how long images are stored, and where authorities obtain such images in the first place are opaque issues and vary by jurisdiction. Recent disclosures that U.S. law enforcement agencies (the Federal Bureau of Investigation and Immigration and Customs Enforcement) scanned through millions of photos in state driver’s license databases without prior knowledge or consent come as little surprise. The vacuum of legal checks and balances has led to a “surveillance-first, ask-permission-later system,” Drew Harrell noted in the Washington Post.58
Second, the accuracy of facial recognition technology varies significantly. Certain tests have disclosed unacceptably high false-match rates. A recent independent report of the UK’s Metropolitan Police found that its facial recognition technology had an extraordinary error rate of 81 percent.59 Similarly, Axon, a leading supplier of police body cameras in the United States, announced that it would cease offering facial recognition on its devices. Axon’s independent ethics board stated: “Face recognition technology is not currently reliable enough to ethically justify its use.”60
But other assessments demonstrate much more favorable results. Evaluations conducted between 2014 and 2018 of 127 algorithms from thirty-nine developers by the U.S. National Institute for Standards and Technology showed that “facial recognition software got 20 times better at searching a database to find a matching photograph.” The failure rate in the same period dropped from 4.0 percent to 0.2 percent.61
One reason for the discrepancy is that under ideal conditions, facial recognition can perform very well. But when unexpected variables are thrown in—poor weather or fuzzy database images—then failure rates start to shoot up. Facial recognition technology also has been unable to shake consistent gender and racial biases, which lead to elevated false positives for minorities and women—“the darker the skin, the more errors arise—up to nearly 35 percent for images of darker skinned women” noted Steve Lohr in the New York Times.62
Citizens are starting to fight back against facial recognition systems. Protesters in Hong Kong, for example, have covered up their faces and disabled their smartphone facial recognition logins to prevent law enforcement access. They have also turned the tables on the police by taking pictures of unbadged officers and using facial recognition image searching to expose the officers’ identities online.63
Facial recognition systems are rapidly spreading around the world. The index identifies at least sixty-four countries that are actively incorporating facial recognition systems in their AI surveillance programs.
The idea behind smart policing is to feed immense quantities of data into an algorithm—geographic location, historic arrest levels, types of committed crimes, biometric data, social media feeds—in order to prevent crime, respond to criminal acts, or even to make predictions about future criminal activity. As Privacy International notes: “With the proliferation of surveillance cameras, facial recognition, open source and social media intelligence, biometrics, and data emerging from smart cities, the police now have unprecedented access to massive amounts of data.” Therefore, one major component to smart policing is to create automated platforms that can disaggregate immense amounts of material, facilitate data coming in from multiple sources, and permit fine-tuned collection of individual information.
One area that has received considerable recent attention is predictive policing. The technique accelerated in the United States after the National Institute of Justice started issuing grants for pilot predictive policing projects in 2009. At its core, these programs claim to predict with remarkable accuracy, based on massive data aggregation, where future crimes will be committed and which individuals are likely to commit those crimes. Predictive policing has exploded in popularity. The PredPol predictive analytics program, for example, is deployed “by more than 60 police departments around the country.”64
But there are growing concerns about algorithmic bias and prejudice, as well as the effectiveness of these predictions. Recent reporting by Caroline Haskins for Vice describes how PredPol’s predictive crime forecasting algorithm operates. Predpol’s software generates crime forecasts for police officers “on a scale as small as 500 by 500 square feet,” which can pinpoint specific houses. It assumes that “certain crimes committed at a particular time are more likely to occur in the same place in the future.”65 PredPol reveals that “historical event datasets are used to train the algorithm for each new city (ideally 2 to 5 years of data). PredPol then updates the algorithm each day with new events as they are received from the department.” New predictions are highlighted in special red boxes superimposed on Google Maps representing high-risk areas that warrant special attention from police patrols.66 A key shortcoming in PredPol’s methodology is that it generates future predictions based on data from past criminal activity and arrests. Certain minority neighborhoods that have suffered from “overpolicing” and biased police conduct show up with higher frequency in PredPol’s dashboard. This may not represent fine-tuned algorithmic crime prediction as much as it involves the perpetuation of structurally biased policing.
China has enthusiastically embraced predictive policing as part of its Xinjiang crackdown. Human Rights Watch reports on the creation of an Integrated Joint Operations Platform (IJOP), which collects data from CCTV cameras, facial recognition devices, and “wifi sniffers” (devices that eavesdrop on activities or communications within wireless networks). IJOP procures additional data from license plates and identification cards scanned at checkpoints, as well as health, banking, and legal records.67 Chinese authorities are supplementing IJOP with mandatory DNA samples from all Xinjiang residents aged twelve to sixty-five.68 This information is fed into IJOP computers, and algorithms sift through troves of data looking for threatening patterns. Once IJOP flags an individual, that person is picked up by security forces and detained for questioning.69
Smart policing techniques are used in at least fifty-three of seventy-five countries with AI surveillance.
AI SURVEILLANCE ENABLING TECHNOLOGIES
A second category of technology is not directly responsible for supporting surveillance programs, but provides critical capabilities that are essential for implementing applications. Advanced video surveillance and facial recognition cameras could not function without cloud computing capabilities. As one expert put it, if video surveillance is the “eyes” then cloud services are the “brains” that “connect cameras and hardware to the cloud computing models via 5G networks.”70 However, cloud computing in isolation is not inherently oriented toward surveillance. Therefore, these secondary technologies are placed in an “enabling technologies” category and described below.71 They are not included in the AIGS Index.
AUTOMATED BORDER CONTROL SYSTEMS
These are found primarily in international airports and border crossings. According to the consulting firm Accenture, ABC systems use “multi-model biometric matching”—facial image recognition combined with e-passports or other biometric documents—to process passengers.72 The process initiates when a passenger steps in front of a multi-camera wall. Digital mirrors located adjacent to the cameras attract passengers’ eyes for image capture. A risk assessment is then performed through automated testing of identities against an individual’s passport and certain security watch-lists.73 Those who are not cleared by the automated system must go into secondary screening with human agents.
Governments are piloting new features, such as automated lie detection technology, in ABC systems. For example, the European Union is testing a technology called iBorderCtrl in three countries—Greece, Hungary, and Latvia—to screen migrants at border crossings. Individuals are asked questions about their countries of origin and circumstances of departure. The answers are then evaluated by an AI-based lie-detecting system.74 Travelers found to have honestly answered questions are given a code allowing them to cross. All others are transferred to human border guards for additional questioning. The technology behind iBorderCtrl is based on “affect recognition science,” which purports to read facial expressions and infer emotional states in order to render legal judgments or policy decisions. Psychologists have widely criticized these tools, maintaining that it is difficult to rely on facial expressions alone to accurately determine a person’s state of mind.75 Despite scientific skepticism about these techniques, governments continue to explore their use.
Governments and companies are increasingly storing data in massive off-site locations—known as the cloud—that are accessible through a network, usually the internet.76 Cloud computing is a general use technology that includes everything from turn-by-turn GPS maps, social network and email communications, file storage, and streaming content access. The National Institute of Standards and Technology defines cloud computing as a “model for enabling ubiquitous, convenient, on-demand network access to a shared pool of configurable computing resources (e.g., networks, servers, storage, applications, and services) that can be rapidly provisioned and released with minimal management effort or service provider interaction.”77 In basic terms, cloud computing data centers function as the backbone of the internet, instantly storing, communicating, and transporting produced information. As such, cloud computing is essential to effectively running AI systems. Microsoft, IBM, Amazon, Huawei, and Alibaba have all established these data centers to facilitate AI operations.
A growing number of countries have fully embraced cloud computing and outsourced all of their data storage needs to a single corporate platform. In 2018, for example, Iceland signed a service agreement with Microsoft for the company to be the sole IT supplier for the country’s entire public sector.78 The cloud computing trend is not without problems. For one, cloud servers present enticing targets for cyber hackers. Security firms like NSO Group claim they are able to penetrate cloud servers and access a target’s “location data, archived messages or photos,” leading many to question whether cloud computing companies can keep personal information, corporate secrets, classified government material, or health records safe (however they generally represent a more secure method of storage than legacy on-site data storage facilities).79 A related concern is forced data disclosures—even if cloud servers remain technically secure, governments may coerce companies into disclosing certain data (such as email communications or text messages of regime critics) held in the cloud.
INTERNET OF THINGS
The IOT is based on the reality that more and more devices will be connected to each other via the internet, allowing data to be shared for analytic processing in the cloud.80 A major IOT hurdle is lack of interoperability between devices. At present, iPhones, Alexa speakers, Nest thermostats, and OnStar auto systems function from different platforms and use different information sources. The IOT’s goal is to “help tame this Tower of Babel” and ensure device integration and data aggregation (although companies like Amazon, Apple, and Google are also setting up distinct ecosystems that only have limited interoperability with other platforms).81 While the IOT will bring greater efficiencies, it may also transform traditional non-networked devices, such as smart speakers, into omnipresent surveillance instruments:
The Internet of Things promises a new frontier for networking objects, machines, and environments in ways that we [are] just beginning to understand. When, say, a television has a microphone and a network connection, and is reprogrammable by its vendor, it could be used to listen in to one side of a telephone conversation taking place in its room—no matter how encrypted the telephone service itself might be. These forces are on a trajectory towards a future with more opportunities for surveillance.82
Controversy surrounding IOT technology is growing. In early 2019, Amazon disclosed that thousands of its workers listened to conversations recorded by Echo smart speakers. In some cases, its workers debated whether recordings of possible crimes should be turned over to law enforcement authorities.83 Amazon analyzed these transcripts without the knowledge or consent of its customers. Similarly, it came to light that Google and Facebook contractors have been regularly listening to recordings between their platforms and individual consumers.84
IOT-powered mobile surveillance is another possibility for this class of technology. A new device was recently demonstrated that plugs into a Tesla Model S or Model 3 car and turns its built-in cameras “into a system that spots, tracks, and stores license plates and faces over time,” journalist Andy Greenberg described. When the owner has parked the car, “it can track nearby faces to see which ones repeatedly appear.” The purpose of the device is to warn car owners against thieves and vandals. But as the device’s inventor Truman Kain acknowledges, “it turns your Tesla into an AI-powered surveillance station” and provides “another set of eyes, to help out and tell you it’s seen a license plate following you over multiple days, or even multiple turns of a single trip.”85
The spread of AI surveillance continues unabated. Its use by repressive regimes to engineer crackdowns against targeted populations has already sounded alarm bells. But even in countries with strong rule of law traditions, AI gives rise to troublesome ethical questions. Experts express concerns about facial recognition error rates and heightened false positives for minority populations. The public is increasingly aware about algorithmic bias in AI training datasets and their prejudicial impact on predictive policing algorithms and other analytic tools used by law enforcement. Even benign IOT applications—smart speakers, remote keyless entry locks, automotive intelligent dash displays—may open troubling pathways for surveillance. Pilot technologies that states are testing on their borders—such as iBorderCtrl’s affective recognition system—are expanding despite criticisms that they are based on faulty science and unsubstantiated research. The cumulative impact gives pause. Disquieting questions are surfacing regarding the accuracy, fairness, methodological consistency, and prejudicial impact of advanced surveillance technologies. Governments have an obligation to provide better answers and fuller transparency about how they will use these new intrusive tools.
The purpose of the index and working paper is to highlight emergent trends for a technology that is not well understood yet will increasingly shape modern life. The good news is that there is ample time to initiate a much-needed public debate about the proper balance between AI technology, government surveillance, and the privacy rights of citizens. But as these technologies become more embedded in governance and politics, the window for change will narrow.
Special thanks to Luke Lamey, undergraduate at Georgetown University, for research assistance in compiling references for the AI Global Surveillance Index. Many thanks also go to Jon Bateman (Carnegie Endowment for International Peace), Thomas Carothers (Carnegie Endowment for International Peace), Adrian Shabhaz (Freedom House), Brian Wampler (Boise State University), and Nick Wright (Intelligent Biology, Georgetown University) for generously giving their time to read through prior drafts of this paper and to offer invaluable feedback and advice.
APPENDIX 1: AIGS INDEX
CA: Closed Autocracy
EA: Electoral Autocracy
ED: Electoral Democracy
LD: Liberal Democracy
EAP: East Asia & Pacific
EUR: Europe & Eurasia
MENA: Middle East & North Africa
SCA: South & Central Asia
WHA: Western Hemisphere
- Aggregate Regime Score: Comprised of evenly weighted average of country scores from Freedom in the World 2019,
- https://freedomhouse.org/report/countries-world-freedom-2019; the EIU Democracy Index 2018,
- https://www.eiu.com/topic/democracy-index; and the V-Dem Dataset version 9, Electoral Democracy Index,
- Regime Type: The AIGS Index uses a four-part regime category typology established by V-Dem and Regimes of the World: closed autocracies, electoral autocracies, electoral democracies, and liberal democracies. Countries with an average score less than 2.5 are labeled closed autocracies. Countries with an average score between 4.9 and 2.5 are labeled electoral autocracies. Countries with an average score between 7.9 and 5.0 are labeled electoral democracies. Countries with an average score of 8 or higher are labeled liberal democracies. See Anna Lührmann, Marcus Tannenberg, and Staffan I. Lindberg, “Regimes of the World (RoW): Opening New Avenues for the Comparative Study of Political Regimes,” Politics & Governance 6, no. 1 (2018).
- “Freedom on the Net 2018” data can be accessed at
- Military expenditure data compiled by the Stockholm International Peace Research Institute and can be accessed at
- Data on Belt and Road Initiative participating countries compiled from Dan Kliman and Abigail Grace’s report, “Power Play: Addressing China’s Belt and Road Strategy,” Center for a New American Security, 2018,
- https://s3.amazonaws.com/files.cnas.org/documents/CNASReport-Power-Play-Addressing-Chinas-Belt-and-Road-Strategy.pdf?mtime=20180920093003, as well as from open source reporting.
- Data for China’s overseas direct investment compiled by the Economist Intelligence Unit, “China Going Global Investment Index 2017,”
APPENDIX 2: TAXONOMY OF DIGITAL REPRESSION
1 The index uses a four-part regime category typology established by V-Dem and Regimes of the World: closed autocracies, electoral autocracies, electoral democracies, and liberal democracies. As the authors describe it, “In closed autocracies, the chief executive is either not subjected to elections or there is no meaningful, de-facto competition in elections. Electoral autocracies hold de-facto multiparty elections for the chief executive, but they fall short of democratic standards due to significant irregularities, limitations on party competition or other violations of Dahl’s institutional requisites for democracies. To be counted as electoral democracies, countries not only have to hold defacto free and fair and multiparty elections, but also . . . achieve a sufficient level of institutional guarantees of democracy such as freedom of association, suffrage, clean elections, an elected executive, and freedom of expression. A liberal democracy is, in addition, characterized by its having effective legislative and judicial oversight of the executive as well as protection of individual liberties and the rule of law.” Anna Lührmann, Marcus Tannenberg, and Staffan I. Lindberg, “Regimes of the World (RoW): Opening New Avenues for the Comparative Study of Political Regimes,” Politics & Governance 6, no. 1 (2018).
2 Information from the Stockholm International Peace Research Institute (SIPRI), SIPRI Military Expenditure Database, 2019, https://www.sipri.org/databases/milex.
3 “Freedom on the Net 2018: The Rise of Digital Authoritarianism,” Freedom House, October 30, 2018.
4 In contrast to V-Dem, the AIGS Index does not evaluate the Palestinian Territories (West Bank, Gaza) and Zanzibar due to inconsistent treatment by external sources. See Michael Coppedge, John Gerring, Carl Henrik Knutsen, Staffan I. Lindberg, Jan Teorell, David Altman, Michael Bernhard, et al, “V-Dem Country-Year Dataset 2019,” Varieties of Democracy (V-Dem) Project, 2019,
5 Nils J. Nilsson, The Quest for Artificial Intelligence (United States: Cambridge University Press, 2009), 4.
6 “Preparing for the Future of Artificial Intelligence,” whitehouse.gov, October 2016,
8 As described in the executive summary, in contrast to V-Dem, the AIGS Index does not evaluate the Palestinian Territories (West Bank, Gaza) and Zanzibar due to inconsistent treatment by external sources. “V-Dem Country-Year Dataset 2019,” 2019.
9 “Freedom on the Net 2018: The Rise of Digital Authoritarianism,” October 30, 2018, 9.
10 “Digital 2019: Internet Trends in Q3 2019,” DataReportal—Global Digital Insights, July 19, 2019.
11 Kim Hart, “Baltimore Wrestles With Aerial Surveillance,” Axios, July 31, 2019,
12 Kevin Rector and Alison Knezevich, “Maryland’s Use of Facial Recognition Software Questioned by Researchers, Civil Liberties Advocates,” Baltimore Sun, October 18, 2016,
13 “Persistent Surveillance’s Cynical Attempt to Profit Off Baltimore’s Trauma,” ACLU of Maryland, June 8, 2018,
14 Olivia Solon, “‘Surveillance Society’: Has Technology at the US-Mexico Border Gone Too Far?,” Guardian, June 13, 2018,
15 Alvaro Artigas, “Surveillance, Smart Technologies and the Development of Safe City Solutions: The Case of Chinese ICT Firms and Their International Expansion to Emerging Markets,” IBEI Working Paper, 2017,
16 Theodore Terschlusen, “Valencienne: Demain Les Caméras de Vidéosurveillance Seront Intelligentes Et…Chinoises,” September 2, 2017,
17 SIPRI Military Expenditure Database, 2019.
18 “Report of the Special Rapporteur on the Promotion and Protection of the Right to Freedom of Opinion and Expression, Frank La Rue,” A/HRC/23/40, April 17, 2013,
19 “The Right to Privacy in the Digital Age,” Report of the Office of the United Nations High Commissioner for Human Rights, A/HRC/27/37, June 30, 2014,
https://www.ohchr.org/EN/HRBodies/HRC/RegularSessions/Session27/Documents/A.HRC.27.37_en.pdf. See also article 12 of the Universal Declaration of Human Rights.
20 “Surveillance and Human Rights, Report of the Special Rapporteur on the Promotion and Protection of the Right to Freedom of Opinion and Expression,” A/HRC/41/35, May 28, 2019,
21 “Necessary and Proportionate,” Necessary and Proportionate: International Principles on the Application of Human Rights to Communications Surveillance, March 4, 2016.
22 “The Right to Privacy in the Digital Age,” A/HRC/27/37.
23 “Surveillance and Human Rights,” A/HRC/41/35.
24 Steven Feldstein, “Can a U.N. Report Help Rein in Expansive and Abusive Digital Surveillance?,” World Politics Review, July 9, 2019,
25 “The Right to Privacy in the Digital Age,” A/HRC/27/37.
26 “Report of the Special Rapporteur on the Promotion and Protection of the Right to Freedom of Opinion and Expression, Frank La Rue,” A/HRC/23/40.
27 Steven Feldstein, “The Road to Digital Unfreedom: How Artificial Intelligence Is Reshaping Repression,” Journal of Democracy 30, no. 1 (2019): 42.
28 Nicholas Wright notes: “People will know that the omnipresent monitoring of their physical and digital activities will be used to predict undesired behavior, even actions they are merely contemplating. . . . In order to prevent the system from making negative predictions, many people will begin to mimic the behaviors of a “responsible” member of society. These may be as subtle as how long one’s eyes look at different elements on a phone screen. This will improve social control not only by forcing people to act in certain ways, but also by changing the way they think.” Nicholas Wright, “How Artificial Intelligence Will Reshape the Global Order,” Foreign Affairs, July 10, 2018,
29 See Jeffrey Ding, “Deciphering China’s AI Dream,” Future of Humanity Institute, University of Oxford, 2018,
30 See for example: “China’s Digital Authoritarianism: Surveillance, Influence, and Political Control,” U.S. House of Representatives, Permanent Select Committee on Intelligence, committee hearing, May 16, 2019,
Robert Morgus and Justin Sherman, “Authoritarians Are Exporting Surveillance Tech, And With It Their Vision for the Internet,” Council on Foreign Relations, December 5, 2018,
“Freedom on the Net 2018: The Rise of Digital Authoritarianism,” October 30, 2018; and Paul Mozur, Jonah M. Kessel, and Melissa Chan, “Made in China, Exported to the World: The Surveillance State,” New York Times, April 24, 2019,
31 “Yanbu: A Smart Industrial Oil Kingdom City — Huawei Publications,” Huawei, 2019,
Sebastian Moss, “Google Cloud Continues to Grow, Is Coming to Saudi Arabia,” Data Centre Dynamics, April 24, 2018,
Rob Evans, “BAE ‘Secretly Sold Mass Surveillance Technology to Repressive Regimes,’” Guardian, June 14, 2017,
Triska Hamid, “NEC Profits from Middle East Cyber Fears,” The National, December 10, 2013,
and Alaa Shahine, Erik Schatzker, Vivian Nereim, and Glen Carey, “Saudis Are Talking to Amazon, Alibaba About New City, Prince Says,” Bloomberg, October 26, 2017,
32 Korkit Danchaivichit (deputy secretary general of the National Telecommunications and Broadcasting Commission), interview with the author, May 17, 2019.
33 Joe Parkinson, Nicholas Bariyo and Josh Chin, “Huawei Technicians Helped African Governments Spy on Political Opponents,” Wall Street Journal, August 14, 2019, https://www.wsj.com/articles/huawei-technicians-helped-african-governments-spy-on-political-opponents-11565793017.
34 “Uganda’s Cash-Strapped Cops Spend $126 Million on CCTV from Huawei,” Reuters, August 16, 2019, https://www.reuters.com/article/us-uganda-crime-idUSKCN1V50RF.
35 Danielle Cave Thomas, Samantha Hoffman, Alex Joske, and Fergus Ryan, Elise, “Mapping China’s Tech Giants,” Australian Strategic Policy Institute, Issues Paper, Report No. 15/2019, https://www.aspi.org.au/report/mapping-chinas-tech-giants.
36 “Building a Safe Mauritius, the Inspiration for Heaven,” Huawei, 2019,
37 Anna Fifield, “Bloodthirsty Like a Wolf: Inside the Military Style Discipline at China’s Tech Titan Huawei,” Washington Post, December 13, 2018,
38 The EU’s former top trade official, Karel De Gucht, is on record observing: “They [Huawei] get subsidies. If you have a line of a couple of tens of billions with the bank that you can use at your discretion this is a huge subsidy, no?” Shawn Donnan and Christian Oliver, “EU Commissioner Attacks China’s Telecoms Subsidies,” Financial Times, March 27, 2014,
Likewise, a 2012 U.S. congressional report from the House Intelligence Committee noted that Huawei reaps the “benefit of billions of dollars in Chinese government financing,” and that both Huawei and ZTE “provide a wealth of opportunities for Chinese intelligence agencies to insert malicious hardware or software implants into critical telecommunications components and systems.” “Investigative Report on the U.S. National Security Issues Posed by Chinese Telecommunications Companies Huawei and ZTE,” U.S. House of Representatives, Permanent Select Committee on Intelligence,
October 8, 2012,
39 Max Chafkin and Joshua Brustein, “Why America Is So Scared of China’s Biggest Tech Company,” Bloomberg Businessweek, March 23, 2018,
41 Zhang Lin, “Chinese Communist Party Needs to Curtail Its Presence in Private Businesses,” South China Morning Post, November 25, 2018,
42 Arjun Kharpal, “Huawei CEO: No Matter My Communist Party Ties, I’ll ‘Definitely’ Refuse If Beijing Wants Our Customers’ Data,” CNBC, January 15, 2019,
43 Paul Mozur, “China’s Internet Controls Will Get Stricter, to Dismay of Foreign Business,” New York Times, November 7, 2016,
44 Christopher Balding and Donald C. Clarke, “Who Owns Huawei?” SSRN Scholarly Paper, Rochester, NY: Social Science Research Network, April 17, 2019,
45 Author interview with an official from Thailand’s Ministry of Interior, May 14, 2019.
46 Somkiat Tangkitvanich (president, Thailand Development Research Institute), interview with the author, May 14, 2019.
47 Christian Davenport, “State Repression and Political Order,” Annual Review of Political Science 10 (2007): 2.
48 “Smart Cities,” World Bank, January 8, 2015,
49 “Smart Cities: Utopian Vision, Dystopian Reality,” Privacy International, October 2017,
50 “Huawei Smart City White Paper,” Huawei Enterprise, 2016,
52 “The Road to Collaborative Public Safety,” Huawei, 2017,
53 “Three Rules When Using AI to Add Value to Your IoT Smart Cities,” Gartner, January 29, 2018,
54 Bojan Stojkovski, “Big Brother Comes to Belgrade,” Foreign Policy, June 18, 2019, accessed July 29, 2019,
55 “Huawei Safe City Solution: Safeguards Serbia,” Huawei Enterprise, August 23, 2018,
56 Feldstein, “The Road to Digital Unfreedom: How Artificial Intelligence Is Reshaping Repression,” 40.
57 “Video Surveillance as the Foundation of ‘Safe City’ in Kenya,” Huawei, 2019,
See also Victor Kapiyo and Grace Githaiga, “Kenya, Communications Surveillance in the Digital Age,” Global Information Society Watch,” 2014,
58 Drew Harrell, “FBI, ICE Find State Driver’s License Photos Are a Gold Mine for Facial-Recognition Searches,” Washington Post, July 7, 2019,
59 Rowland Manthorpe and Alexander J Martin, “81% of ‘Suspects’ Flagged by Met’s Police Facial Recognition Technology Innocent, Independent Report Says,” Sky News, July 4, 2019,
60 Charlie Warzel, “A Major Police Body Cam Company Just Banned Facial Recognition,” New York Times, June 27, 2019,
61 “NIST Evaluation Shows Advance in Face Recognition Software’s Capabilities,” U.S. National Institute for Standards and Technology, November 30, 2018,
62 Steve Lohr, “Facial Recognition Is Accurate, If You’re a White Guy,” New York Times, February 9, 2018,
63 Paul Mozur, “In Hong Kong Protests, Faces Become Weapons,” New York Times, July 26, 2019,
64 Randy Rieland, “Artificial Intelligence Is Now Used to Predict Crime. But Is It Biased?” Smithsonian, March 5, 2018.
65 Caroline Haskins, “Revealed: This Is Palantir’s Top-Secret User Manual for Cops,” Vice, July 12, 2019,
66 “The Three Pillars of Predictive Policing,” PredPol, 2018,
67 “China: Big Data Fuels Crackdown in Minority Region,” Human Rights Watch, February 26, 2018,
68 Cate Cadell, “From Laboratory in Far West, China’s Surveillance State Spreads,” Reuters, August 14, 2018,
“China: Minority Region Collects DNA from Millions,” Human Rights Watch, December 13, 2017,
69 Feldstein, “The Road to Digital Unfreedom: How Artificial Intelligence Is Reshaping Repression,” 45.
70 Jeffrey Ding, “ChinAI #59: Is Winter Coming for Hikvision?,” accessed July 28, 2019, https://chinai.substack.com/p/chinai-59-is-winter-coming-for-hikvision.
71 For an in-depth look at the AI machine learning value chain, including an incisive examination of data storage modalities, see: Charlotte Stanton et al., “What the Machine Learning Value Chain Means for Geopolitics,” Carnegie Endowment for International Peace, August 5, 2019,
72 “Accenture Automated Border Clearance Solutions,” Accenture, 2019,
73 “EGate Solutions: Automated Border Control (ABC),” Gemalto, April 12, 2019, https://www.gemalto.com/govt/coesys/eborder-abc.
74 “IBorderCtrl: The Project,” iBorderCtrl, 2016, https://www.iborderctrl.eu/The-project.
75 Lisa Feldman Barrett, Ralph Adolphs, Stacy Marsella, Aleix M. Martinez, and Seth D. Pollak, “Emotional Expressions Reconsidered: Challenges to Inferring Emotion From Human Facial Movements,” Psychological Science in the Public Interest, July 17, 2019,
See also Lucien Begault, “Automated Technologies and the Future of Fortress Europe,” Amnesty International, March 28, 2019,
76 Patricia Moloney Figliola and Eric A Fischer, “Overview and Issues for Implementation of the Federal Cloud Computing Initiative: Implications for Federal Information Technology Reform Management,” Congressional Research Service, January 20, 2015,
77 Peter Mell and Timothy Grance, “The NIST Definition of Cloud Computing,” NIST Special Publication 800-145, September 2011,
78 Esat Dedezade, “Iceland to Become the First ‘Cloud-First-Nation,’” Microsoft News Centre Europe, September 19, 2018,
79 Mehul Srivastava and Tim Bradshaw, “Israeli Group’s Spyware ‘Offers Keys to Big Tech’s Cloud’,” Financial Times, July 19, 2019,
80 “Artificial Intelligence and Life in 2030: One Hundred Year Study on Artificial Intelligence,” Report of the 2015-2016 Study Panel, Stanford University, September 2016,
82 Jonathan L. Zittrain, Matthew G. Olsen, David O’Brien, and Bruce Schneier, “Don’t Panic: Making Progress on the “Going Dark” Debate.” Berkman Center Research Publication 2016-1,
83 Matt Day, Giles Turner, and Natalia Drozdiak, “Amazon Workers Are Listening to What You Tell Alexa,” Bloomberg, April 10, 2019,
84 Greg Bensinger, “Google Employs Humans to Listen to Some Voice-Assistant Recordings,” Washington Post, July 11, 2019,
85 Andy Greenberg, “This Tesla Mod Turns a Model S Into a Mobile ‘Surveillance Station,’” Wired, August 9, 2019,
GLOBAL SURVEILLANCE CONNECTION OF THE FIVE EYES AGREEMENT
GLOBAL – Silent, Invisible, Deadly:
The Weapon that Could Change Warfare
This month, the Force awakens in theaters. Next month, a new force awakens in the New Mexico desert, where the Defense Department is to start testing a weapon worthy of “Star Wars” — a silent, invisible laser that needs just a couple of seconds to burn a hole in targets miles away.
“What it really boils down to is a silent weapon that nobody sees or hears,” said Lt. Gen. Bradley Heithold, commander of Air Force Special Operations Command (AFSOC, pronounced “AFF-sock”).
Heithold is eager to put a laser cannon on four or five of AFSOC’s three dozen or so AC-130 gunships. AC-130s are typically used to cover special operations troops on nighttime missions.
“I think we can do this now,” Heithold said of the laser weapon. “When I say now, I’m talking about by the close of the decade.”
If Heithold is right, that’s a revolutionary turn of events — after decades of costly research.
A laser — the word is an acronym meaning “light amplification by stimulated emission of radiation” — is an intense beam of light that can be produced using a variety of materials. Bursts of energy are “pumped” into a “gain medium,” which can be a gas, liquid or solid, to excite the gain medium’s electrons and release photons.
A 24/7 GLOBAL TESTING, TRAINING AND TARGETING
OPERATION WITH MILITARY AND LAW ENFORCEMENT PERSONNEL
WORKING IN 8 HOUR ROTATING SHIFTS FOR HUMAN EXPERIMENTATION
GLOBAL REPORT – THE COVERT HARASSMENT
CONFERENCE – BERLIN
MIND TECH SWEDEN
GLOBAL HUMAN EXPERIMENTATION
PERFECTING A GLOBAL CONTROL GRID
DISCOVERING THE SECRET PROGRAM
OF USA MIND CONTROL
MIND CONTROLLED POLICE AND MILITARY FORCES
ORIGIN WITH W. ROSS ADEY – USC BRAIN RESEARCH CENTER
Dr. Ross Adey, formerly of the Brain Research Center at the University of Southern California, and the Loma Linda University Medical School, California, worked on the CIA’s infamous Pandora project. His research involved inducing specific behavior modifications by electromagnetic means, as well as inducing calcium efflux events to interfere with brain function resulting in the so-called “confusion weaponry.” (More recently, Dr. Adey had been warning against the biological hazards of EM emissions from mobile phones.
Dr. Adey’s research at the Brain Research Center has shown the existence of a biological reaction to EM radiation. This reaction was found to be dependent on the frequency, amplitude, and dose of the microwave radiation used. In the 80s Dr. Adey performed some crucial experiments using microwave carrier-waves modulated with ELF waves to modify brain tissue responses. He carried out experiments with cat brain tissue, showing that the binding of calcium ions to neuronal sites was affected by weak EM fields. These fields were of a frequency and amplitude similar to fields produced in mammalian bodies and detected on electroencephalographs (EEGs).
Dr. Adey demonstrated how a 147-megahertz (MHz) field, which at tissue level had an intensity of 0.8 milliwatts per square centimeter, caused an efflux or release of calcium ions from the irradiated brain tissue. This response only occurred when the ELF modulation of the microwave carrier-wave had an amplitude modulated at 6-20 hertz (Hz). The maximum stimulation of the neurons took place at 16 Hz, but to either side of this frequency-range parameter, there was no effect.8 [1996 Nexus]
W. ROSS ADEY PROJECT PANDORA “REMOTE BRAIN MANIPULATION”
W. Ross Adey Obituary
More activism from Tim Rifat…
REMOTE VIEWING: THE STORY OF THE REAL X-FILES By Tim Rifat
Secondly, the police have been granted the exclusive use of the 450 MHz microwave frequency range. This is exactly the frequency used by Dr. Ross Adey, the CIA mind control expert, in his experiments on behavioral modification. It seems the police have the exclusive use of this mind control frequency and a vast array of antennae to broadcast this frequency all over the country. Very useful for mass mind control in times of emergency. Adey found that by using 0.75mW/cm2 intensity of pulse modulated microwave at a frequency of 450 MHz, it was discovered that an ELF modulation could be used to control all aspects of human behavior. The Sussex police headquarters is connected to CCTV, closed-circuit television cameras throughout the town. Some of these cameras have microwave telemetry devices that could easily be used to broadcast this frequency. The large antennae that bracket the town could also be used.
MICROWAVE MIND CONTROL by Tim Rifat
By use of my microwave detector, I have found that mobile phones of the newer type, give off a pulse modulated microwave signal of around 0.75mW/cm2 at the earpiece. This may be a coincidence, but it is exactly the intensity required to form behavioral control as found by Dr. Ross Adey, the pioneer of microwave mind control. So in theory, an ELF signal could be added to the microwave network to feed a precise behavioral pattern into every mobile phone user in the UK. If there were widespread riots, the ability to broadcast behavioral stimuli to mollify all the mobile phone users in the country would prove useful since mobile phone users are generally middle class, it means authority has a useful method of controlling the behavior of the key voters. Microwave carrier beams are perfect for transmitting the excitation potential of docility to the phone user to keep them servile in times of trouble. When no ELF signal is broadcast the phone acts in a completely different manner on behavior in humans. In this case, microwave phones cause the neurons to release calcium ions which makes the user tired irritable and when stresses likely to emotional outbursts such as road rage.
MICROWAVE MIND CONTROL OF POLICE AND MILITARY
PERSONNEL THE ENFORCERS
The TETRA system pulses at 17.6 Hz broadcast at 400 MHz which is essentially the Pandora Project funded by the CIA in the late ’60s and early ’70s. Dr. Ross Adey, the chief researcher on the Pandora Project has released a video to leading UK researchers which proves that not only does the TETRA system cause ELF zombification by the massive release of calcium ions in the cerebral cortex and the nervous system, but the activated calcium ions also cause massive hormonal disturbances which lead to frenzied imbalances, emotional and physical states…… Use of the TETRA system by the police will lead to psychotronically controlled officers who may be totally controlled in any situation and are very useful for states of economic or social chaos where extreme and violent behavior is needed without any conscious or moral compunction – so-called police robots. 
The TETRA System: Mass UK Mind Control Technology
and the Zombification Of Britain’s Police is Now A Reality by Tim Rifat
One of the first voices in the West to raise concern was Professor Ross Adey. Although interested in radio and the development of radar from an early age, he qualified in medicine in Australia during the war and then came to Oxford to study brain physiology. He was appointed as a professor at the University of California in 1954 and in 1961 joined their newly formed Brain Research Institute. In 1965 he was appointed as Director of the new Space Biology Laboratory. The projects his teams worked on were related to brain activity of service personnel, including the effects of pulsed ELF and microwave radiation, and were funded by NASA, ARPA, NSF, USAF, and USN. Since then he has remained at the forefront of research into these matters.
“I would be very concerned about a four-watt Tetra hand-held radio against my head day after day”.
So what is the cause of this concern? Back in the mid-1960s Adey and colleagues revolutionized our understanding of the way brains cell operate. During the next 20 years, many papers were published showing that ELF electromagnetic fields affected the way calcium ions move in brain tissue and the way this affects the inner working of the cells. Living cells are enclosed by a strong protective membrane (skin) and communicate with the outside world through electrically charged protein strands that stick out from their surface.
These carry signals into the cell interior where they regulate growth, development, and cell division. By 1976 low-level electromagnetic field effects had been shown to occur across the ELF range, peaking between about 12 and 20 Hz. The effects were seen most clearly with the ELF signals amplitude modulated on to microwaves, such as those that TETRA handsets use. The microwaves penetrate brain tissue and effectively ’carry’ the ELF pulsing deep into the brain. There is also considerable evidence that low levels of pulsed microwaves can cause the blood-brain barrier to allow toxins to reach the brain.
The “Five Eyes Agreement” plays a pivotal role in global targeting utilizing biometric Signal Intelligence (SIGINT) and Human Intelligence (HUMINT) and IMINT (Image Intelligence real-time surveillance) making biometric targeting, tracking, and monitoring effortless through this agreement and any place on Earth.
As the Second World War came to a close the United States and the Soviet Union emerged as rivals and the two primary superpowers vying for global hegemony. Their efforts to obtain superiority over the other resulted in an “arms race” of personnel, aerospace, missile, and nuclear weapons. These efforts spilled into the arena of military and covert intelligence; resulting in both sides competing in the areas of human intelligence (HUMINT), signals intelligence (SIGINT), and image intelligence (IMINT).
SIGNAL INTELLIGENCE (SIGINT)
HUMAN INTELLIGENCE (HUMINT) CONNECTION
TO FIVE EYES AGREEMENT
Alasdair Philips, Director of Powerwatch, examines the potential
hazards of TETRA… The negative aspects of TETRA
Barrie Trower, UK, Confidential Report on Cell Towers and TETRA
REPORTED DIRECTED ENERGY WEAPON ANTENNA
APPEARING ON FAKE CELL TOWERS?
USA STINGRAY SURVEILLANCE AND HIDDEN TOWERS
The Governments, of the U.S., Britain, and Canada, as far as I can tell, have a systematic method for harassing especially ‘marked’ citizens that have either displeased them or have been found to be some type of threat.
Victims globally are living through some of the most heinous human experimentation evil today on this planet. They are being used in the initial stage of biometric, algorithm data collection. The choice to stand and fight for our lives through the toughest of circumstances, honorably, and more importantly, intelligently, is the only positive alternative as the truth snowballs.
The goal and hope are for full disclosure of this beyond belief evil program advancing Artificial Intelligence, nanotechnology, and neuroweapons, operating both overtly and covertly undoubtedly within the United States and globally. The cure, for now, EXPOSE, EXPOSE, EXPOSE! History reveals that the TRUTH is an inevitable force for the greater good of humanity!
UNITED WE STAND ALL OVER THE WORLD
ON TRUTH, AND A UNIFIED, DECISIVE MISSION!
“BIGGER THAN SNOWDEN” WHISTLEBLOWER’S ARCHIVED LINK