Blog Post

Opportunities and Challenges from Artificial Intelligence for Law Enforcement

Tom Campbell, PhD, Founder & President, FutureGrasp, LLC • Jul 20, 2018
Like what you read? CONTACT FutureGrasp for more information, speaking engagements and/or collaboration

Law enforcement (LE) has unique challenges—identification of individuals, prediction of criminal actions, tracking of money flows, tagging and defending against fake news, and enhanced communications across agencies, among others—that demand rapid data acquisition and analysis. As artificial intelligence (AI) offers unprecedented capabilities to acquire and analyze big data, it is no surprise that the LE community is interested in applying AI for its purposes. [1] But given LE’s unique requirements on speed, low error rates, and accurate predictions, is AI up to the task yet of helping LE?

We offer here a review of several facets of AI that could benefit or compromise LE. [2] As AI becomes more advanced, the potential for its application by LE experts (and criminals) will only increase. Let us hope that LE officers in the field and analysts in the office are able to leverage AI to protect people around the world.

DATA, DATA, DATA—THE INTERNET OF THINGS

We are living in a veritable information explosion. Every day we create roughly 2.5 quintillion bytes of new data. [3] The Internet of Things – i.e., the universe of devices connected to the internet – is forecasted to exceed 30 billion devices by 2030. Over a trillion sensors of all forms is projected to be in use by 2020. [4] This digital tsunami tasks our local and federal governments in gathering data, comprehending it, and crafting sound policies. Because of the rapid-response needs of LE, this information must be acquired, curated and analyzed in almost real-time.

AI can assist in managing this surfeit of data. Especially well-positioned here is deep learning (DL), which requires massive datasets in order to yield accurate output. Thus, leveraging DL makes perfect sense for the LE community; each of the technologies described below use some aspect of DL.

FACIAL RECOGNITION (FACE-REC)

There are some people who never forget a face. So-called ‘super-recognizers,’ such individuals can remember people they haven’t seen in decades. For the rest of us, we can use mnemonic tricks to not be embarrassed in conversations. Theoretically, computers should trump everyone on identifying people, as fundamentally face-rec is a challenge of data recall.

In the early days of face-rec competitions with algorithms, we routinely beat our silicon creations…then DL was developed. During an imaging contest using a large database, the Imagenet , one team in 2012 successfully leveraged large computational power with DL to improve face-rec accuracy dramatically. As clarified by MIT Technology Review:

“This was the first time that a deep convolutional neural network had won the competition, and it was a clear victory. In 2010, the winning entry had an error rate of 28.2 percent, in 2011 the error rate had dropped to 25.8 percent. But SuperVision won with an error rate of only 16.4 percent in 2012 (the second best entry had an error rate of 26.2 percent). That clear victory ensured that this approach has been widely copied since then.” [5]

At this writing, the record for face-rec is claimed to be less than one percent error—although exact error values vary as a function of lighting, amount of the face exposed, motion and other factors. Depending on the humans tested, errors by people attempting to recognize others range the gamut from 0% to 50% error. [6]

Recognizing faces is a critical aspect of policing work, yet the techniques for its execution must be trusted. Mug shots and line-ups have been used for decades with human victims attempting to identify criminals. Nevertheless, errors can occur and people can be falsely jailed based upon a mistaken identity. [6]Using algorithms to reduce errors can help, but we must be careful to not be too reliant on them, as even a small error rate can still yield false arrests. [7] Microsoft recently called for the US Government to consider regulating face-rec: “Facial recognition technology raises issues that go to the heart of fundamental human rights protections like privacy and freedom of expression. These issues heighten responsibility for tech companies that create these products. In our view, they also call for thoughtful government regulation and for the development of norms around acceptable uses. In a democratic republic, there is no substitute for decision making by our elected representatives regarding the issues that require the balancing of public safety with the essence of our democratic freedoms. Facial recognition will require the public and private sectors alike to step up – and to act.” [8]

Law enforcement should proceed with caution on the use of face-rec. At the present stage of its development, algorithmically-generated face-rec should be verified independently by humans to avoid mistaken identities.

CAMERAS

Cameras of all forms (including body cameras-bodycams, CCTV-closed circuit television, and drone cameras) can be a powerful tool for reducing violence both at and from police, as well as identifying criminals. The range of intrusiveness of these systems varies from passive (observation only) to active (seeking matches to faces and other crime scene clues in large databases).

On the passive side, bodycams can be used to protect both the citizen and the police officer, but they have their own limitations. “They record audio and video, often with date and time stamps as well as GPS coordinates. They can also be Bluetooth-enabled and set to stream in real time. Some have to be turned on manually, others can be triggered automatically by, for instance, an officer unholstering his weapon.” [9] Bodycams are controversial, and their effectiveness in reducing violence both at and from police officers has been called into question. A 2017 study, claiming to be the largest on bodycams to date, concluded that these systems offer mixed results. “A new study suggests police body cameras have had no measurable effect in Washington, D.C., when it comes to civilian complaints, use of force incidents, policing activity or judicial outcomes involving the city’s Metropolitan Police Department.” [10] Nevertheless, we should note that these are still early days for video recordings of citizen-police interactions, and that more use and data may offer greater clarity on the effectiveness of this LE tool.

On the active side, machine learning and deep learning can be leveraged to identify a wide range of characteristics of people and objects from cameras. Perhaps the most intrusive application of AI for personal identification via cameras is in China:

“It should give Westerners no comfort that China—a one-party state obsessed with social order—is at the forefront of developing bodycams. One Beijing company says it has invented a shoulder-worn, networked model that can recognize faces. Another Chinese firm has equipped police with facial-recognition cameras embedded in glasses that are meant to let officers know in real time if they are looking at someone on a police blacklist. One estimate values China’s surveillance-tech market in 2018 at $120bn. Human-rights campaigners fear that such technology has already been used to monitor activists, enabling arbitrary detention.” [7]

Although privacy limitations may limit such active scanning and analysis in Western countries, the genie is out of the bottle for many nation states. Cameras with AI built-in for face-rec are now commercially available. [11] , [12] It remains to be seen how far LE takes such tools. Applications of cameras have strong implications for other technologies discussed below, however.

UNMANNED AERIAL VEHICLES (UAVs, aka, DRONES)

UAVs have been commercially available for over a decade, but their price-points and capabilities came to levels only just within the past few years so that LE is adopting them. UAVs range in size from an artificial bee [13] to a large aircraft [14]. Their capabilities are just as diverse, with not just cameras, but also weaponization [15] , fire-fighting [16] , and even tree-planting [17]. Only a lack of imagination limits what one can do with these flying machines.

Privacy advocates have been calling for years to place limitations on drone use. After much debate, the US Federal Aviation Administration (FAA, https://www.faa.gov/ ) put into laws the conditions under which one must operate to fly a drone of any kind in public. For example, UAVs are not allowed to fly in US national parks without a permit (for concerns of stressing wildlife, endangering people, etc.). [18] [19] All drones within certain size range can be flown legally only with a permit. [20]

LE may have looser restrictions on UAV applications – for example, with a known threat of terrorism or a manhunt underway for a suspected felon. Nevertheless, even LE must comply in the United States with certain limitations to protect the public’s privacy and safety. “Local and state agencies in the US have passed rules imposing strict regulations on how law enforcement and other government agencies can use drones. Indiana law specifies that police departments can use drones for search-and-rescue efforts, to record crash scenes and to help in emergencies, but otherwise a warrant is required to use a drone. That means police probably won’t be able to fly them near large gatherings unless a terrorist attack or crime is under way.” [21]

The implications of UAVs are extensive for LE. The ability to have a birds-eye view practically anywhere (both outdoors and indoors) offers significant upside for LE in tracking, identification and analysis of criminal individuals and actions. Nevertheless, as noted above, UAV use also brings with it great responsibility to ensure safety and privacy of the public. As UAVs advance technologically, LE must continue to engage with local and national governments to ensure these issues are jointly addressed. INTERPOL is taking a lead in assessing drone threats and challenges to LE via its Global Complex for Innovation. [22]

FAKE NEWS

Fake news has a long history. Ever since printed text became widely available, fake news has been used by people to push propaganda and influence one’s enemies. “Fake news took off at the same time that news began to circulate widely, after Johannes Gutenberg invented the printing press in 1439. ‘Real’ news was hard to verify in that era. There were plenty of news sources—from official publications by political and religious authorities, to eyewitness accounts from sailors and merchants—but no concept of journalistic ethics or objectivity. Readers in search of fact had to pay close attention. In the 16th century, those who wanted real news believed that leaked secret government reports were reliable sources, such as Venetian government correspondence, known as relazioni . But it wasn’t long before leaked original documents were soon followed by fake relazioni leaks. By the 17th century, historians began to play a role in verifying the news by publishing their sources as verifiable footnotes. The trial over Galileo’s findings in 1610 also created a desire for scientifically verifiable news and helped create influential scholarly news sources.” [23]

Since those early days, fake news has expanded to include every new recording or communications technology--see image at bottom of text. When photography was first introduced to the public, faking a photograph taken on real film required staging shots or chemical manipulation. Digital photos make it much easier to simply jigger a few pixels and produce any photo one wishes. Now it is a challenge to have a photo be admitted as evidence in a court of law.

As the internet came online in the latter part of the 20th century, fake news really took off. Now with major online platforms such as Facebook and Twitter, as well as the 24/7 news cycle, we struggle to identify what is real or not while living in our chosen filter bubbles. Nation states such as Russia and China have embraced social media as a means for pushing their own messages or propaganda.

We are now progressing into even more sophisticated fake news. Audio and video can be manipulated at will. One research group claims that all they need is 3.7 seconds of a human’s voice soundwaves to mimic them with an algorithm. [24] Another research group takes such work to another level by superimposing a researcher’s face onto that of another individual, and consequently enabling the researcher to make the victim say anything with any expression the researcher desires. [25]

Further into the realm of difficulty and complexity, augmented and virtual realities (AR and VR) will become the next-gen platforms for fake news. As AR and VR are used more by companies and governments, we may fully expect them to be target platforms for fake news. Essentially, any mode of communications is now fair game for fake news generation.

Law enforcement has an important role to play in fake news, albeit one that appears to be still gelling in terms of its scope. Police officers and nation states must both identify and defend against fake news. During the recent Mexico elections (July 2018), an effort was made to identify fake news in text, audio and video by AI. Human curators then reviewed the tagged media and decided if it was truly false and what to do with it if so. [26] At this writing, it is unclear how effective this work was.

As we expand how we exchange news, nefarious individuals will continue to manipulate our communications and distribute their own narratives or influence efforts. It is upon the shoulders of LE and nation states to ensure that fake news is identified and does not result in skewed elections or criminal actions. Sadly, the veracity of news is a competition between criminals and LE with no foreseeable end.

PREDICTIVE POLICING

In the 2002 movie Minority Report , the character played by Tom Cruise uses a bio-based computing system to predict crimes before they happen and thus arrest perpetrators before they commit their nefarious deeds. We are a ways away from being able to make such fully accurate predictions—and accurate they must be; imagine being falsely accused for murder because of algorithmic error or bias?—but we may still consider that the summation of the technologies described above and others are leading us down this path. Whether and how quickly predictive policing moves out of fiction into every police station remains to be seen. Nevertheless, significant efforts and research dollars are being expended to make predictive policing become the norm in identifying crime locations and individuals in advance. We list below just a few examples of the research now underway, and already deployed tools, to enable future Tom Cruises to wield their predictive hands against crime.

In the United States, the company Palantir has been testing predictive policing in large cities for several years. Unbeknownst to several city council members, Palantir algorithms were used in New Orleans for over six years; the Palantir contract has since been cancelled. [27] The Los Angeles Police Department (LAPD) still uses Palantir’s models, but there are claims that the results have inherent racial bias. [28] Because of concerns of opaqueness (lack of disclosure of the databases) and bias, there are now major lawsuits pending against Palantir in Chicago, Los Angeles, New Orleans and New York City. [29]

Japan recently announced a major effort to enable predictive policing. “The AI-based system would employ a ‘deep learning’ algorithm that allows the computer to teach itself by analyzing big data. It would encompass the fields of criminology, mathematics and statistics while gathering data on times, places, weather and geographical conditions as well as other aspects of crimes and accidents.” [30] The Japanese government hopes to prove the system in at least one prefecture first, and then deploy it further at the 2020 Tokyo Olympics. [31]

Perhaps China has taken the level of predictive policing to its most intrusive level, violating international human rights norms in the process. In the prefecture of Xinjiang, local authorities and police use big data analytics to monitor a wide variety of personally identifiable information (PII) on minority citizens, often without their knowledge. Individuals can then be arrested and imprisoned without trial. [32] Chinese citizens, although said by Chinese officials to be trusting of the technology wielded by their local law enforcement, are known to cover their faces when cameras are pointed at them. [33]

Thus, other than a few nation states such as China, widespread use of predictive policing is not yet the norm. An informal survey made during the recent INTERPOL—UNICRI meeting in Singapore on AI, robotics and law enforcement showed that none of the 20 attending countries had significant predictive policing programs planned or in place at this writing (China was not present at the meeting). [1]This lack of broad application may change, though, as algorithm, data and computational capabilities increase. Ultimately, caution should be exercised by all parties in deployment of predictive policing to ensure that it does not compromise the privacy of innocent civilians and/or result in ill will and lawsuits against LE that may take the technology too far.

CONCLUSIONS & IMPLICATIONS

One can argue that AI is certainly capable of helping to address some of the data acquisition and analysis needs in LE, but with the caveat that its use must be monitored and proofed before arrests or litigation decisions are made. Technologies such as face-rec with algorithm-enabled digital cameras, as well as nascent experiments in predictive policing, offer LE officers new tools by which to perform their tasks. The challenges of fake news are increasing in complexity and may require AI to assist LE in its identification and digital removal. Nevertheless, inherent limitations of AI such as inherent bias, opaque algorithms, and lack of complete datasets should give us pause before unchecked deployment of AI in LE. In the United States and elsewhere there are major lawsuits pending against LE agencies that may have deployed AI algorithms too soon without rigorous verification and validation (V&V) mechanisms in place. Other nations such as China have gone all-in on AI-enabled LE capabilities, but at the cost of human rights for their citizens who in minor cases lose privacy and in worst cases get falsely arrested and jailed without warrants.

Despite technical and ethical challenges, the use of AI by LE is entering an inexorable state. In many criminal cases, there is simply too much data for the traditional officer or detective alone to capture and to assess all relevant evidence. Whether and how individual police departments, small towns, cities and nation states apply the powers of AI algorithms remains to be seen. Moreover, its adoption will most probably be driven by culture and nation state interests. Let us hope that the applications of AI in LE are made after careful consideration of both the benefits to LE agencies and the citizens they are charged to protect.

NOTES (all websites accessed July 2018)


[1] Witness the recent workshop, “1st INTERPOL-UNICRI Global Meeting on the Opportunities and Risks of Artificial Intelligence and Robotics for Law Enforcement 2018,” held in Singapore on 11-12 July 2018. I presented the closing talk, “Looking Into The Future…”

[2] Note this review is not meant to be comprehensive, but merely an introduction to the field. Substantial literature has been written about each of the overview sections here.

[3] “10 Key Marketing Trends for 2017,” IBM Marketing Cloud, https://www-01.ibm.com/common/ssi/cgi-bin/ssialias?htmlfid=WRL12345USEN

[4] “Over 1 trillion sensors could be deployed by 2020,” January 26, 2017, https://www.electronicspecifier.com/around-the-industry/over-1-trillion-sensors-could-be-deployed-by-2020

[5] “The Revolutionary Technique That Quietly Changed Machine Vision Forever,” September 9, 2014, https://www.technologyreview.com/s/530561/the-revolutionary-technique-that-quietly-changed-machine-vision-forever/

[6] “Combining the facial recognition decisions of humans and computers can prevent costly mistakes,” June 4, 2018, https://theconversation.com/combining-the-facial-recognition-decisions-of-humans-and-computers-can-prevent-costly-mistakes-97365

[7] “CEO of a Facial Recognition Company: The Tech is Too Volatile to Give to Law Enforcement,” June 28, 2018, https://www.nextgov.com/emerging-tech/2018/06/ceo-facial-recognition-company-tech-too-volatile-give-law-enforcement/149337/

[9] “Data Detectives,” Technology Quarterly, Justice, The Economist, June 2, 2018.

[10] “New Study Casts Doubt On Effectiveness Of Police Body Cameras. But Is That Fair?,” October 24, 2017, https://www.huffingtonpost.com/entry/dc-police-body-camera-study_us_59ee1bace4b003385ac11440

[11] “Artificial intelligence software suite enables advanced people detection capabilities,” June 25, 2018, https://www.vision-systems.com/articles/2018/06/artificial-intelligence-software-suite-enables-advanced-people-detection-capabilities.html

[12] “AWS DeepLens - Deep learning enabled video camera for developers,” https://www.amazon.com/dp/B075Y3CK37/x_tw_01_01_01_04

[13] “Rise Of The Robot Bees: Tiny Drones Turned Into Artificial Pollinators,” March 23, 2017, https://www.npr.org/sections/thesalt/2017/03/03/517785082/rise-of-the-robot-bees-tiny-drones-turned-into-artificial-pollinators

[14] “Predator RQ-1 / MQ-1 / MQ-9 Reaper UAV,” https://www.airforce-technology.com/projects/predator-uav/

[15] “Homeland Security: Warns of Weaponized Drones as Terror Threat,” November 10, 2017, https://www.usnews.com/news/national-news/articles/2017-11-10/homeland-security-warns-of-weaponized-drones-as-terror-threat

[16] “How firefighters are using drones to save lives,” August 27, 2017, https://www.cnbc.com/2017/08/26/skyfire-consulting-trains-firefighters-to-use-drones-to-save-lives.html

[17] “New tree-planting drones can plant 100,000 trees in a single day,” January 31, 2018, https://www.techspot.com/news/73032-new-tree-planting-drones-can-plant-100000-trees.html

[18] Years ago, I was in Yosemite National Park and saw a small UAV being used to film a valley, despite a large crowd present in the same area. I asked the operator if he had a permit—the individual feigned surprise, then immediately landed the drone, grabbed it, and took-off in his pickup truck.

[22] “Drones: The New Threat and Challenge to Law Enforcement: Threat-Tool-Evidence,” INTERPOL, https://www.INTERPOL.int/

[23] “The Long and Brutal History of Fake News,” December 18, 2016, https://www.politico.com/magazine/story/2016/12/fake-news-history-long-violent-214535

[24] “‘Deep Voice’ Software Can Clone Anyone's Voice With Just 3.7 Seconds of Audio,” March 7, 2018, https://motherboard.vice.com/en_us/article/3k7mgn/baidu-deep-voice-software-can-clone-anyones-voice-with-just-37-seconds-of-audio

[26] “AI to help tackle fake news in Mexican election,” June 30, 2018, https://www.bbc.com/news/technology-44655770

[27] “New Orleans ends its Palantir predictive policing program,” March 15, 2018, https://www.theverge.com/2018/3/15/17126174/new-orleans-palantir-predictive-policing-program-end

[28] “The LAPD Has A New Surveillance Formula, Powered By Palantir,” May 8, 2018, https://theappeal.org/the-lapd-has-a-new-surveillance-formula-powered-by-palantir-1e277a95762a/

[29] “'Predictive policing': Big-city departments face lawsuits,” July 5, 2018, http://www.chicagotribune.com/news/sns-bc-us--predictive-policing-challenges-20180705-story.html

[31] “Japan trials AI-assisted predictive policing before 2020 Tokyo Olympics,” January 29, 2018, https://www.scmp.com/news/asia/east-asia/article/2130980/japan-trials-ai-assisted-predictive-policing-2020-tokyo-olympics

[32] “China: Big Data Fuels Crackdown in Minority Region,” February 26, 2018, https://www.hrw.org/news/2018/02/26/china-big-data-fuels-crackdown-minority-region

[33] “Looking Through the Eyes of China’s Surveillance State,” July 16, 2018, https://www.nytimes.com/2018/07/16/technology/china-surveillance-state.html?partner=IFTTT


FutureGrasp, LLC
Like what you read? CONTACT FutureGrasp for more information, speaking engagements and/or collaboration
Diplomacy turmoil in age of coronavirus
By Thomas A. Campbell, Andrew Hyde, Geoffrey M. Odlum, Ariel Ahram 11 Jun, 2020
The coronavirus pandemic has utterly transformed international diplomacy. Statecraft has long been driven by face-to-face interactions that expand through informal mechanisms, such as body language, spontaneous conversations, and an understanding of the views of other actors beyond rote statements of positions. Such meetings are now difficult if not impossible in the era of social distancing and work from home mandates. Meetings postponed or canceled altogether have slowed a range of negotiations while many important international conferences have also been delayed or gone virtual. Cybersecurity risks have amplified as unsecure networks are leveraged for conversations that would have normally occurred face to face. Diplomats must find new ways to work in this pandemic era, as well as what the current turmoil in operations may mean post-COVID19. The practice of diplomacy has always been intertwined with state-of-the-art methods of information and communications technology (ICT). For example, telegraphy revolutionized the practices of diplomacy by accelerating the speed of communication between embassies and the metropole. The internet has had similar revolutionary impact on the practices of statecraft and diplomacy. Web-enabled ICT multiplied the number of voices and interests involved in international policymaking; complicated international decision-making; reduced exclusive control of States in the process; accelerated and freed the dissemination of information, accurate or not, about any issue or event; and enabled traditional diplomatic services to be delivered faster and more cost-effectively, both to ones’ own citizens and government, and to those of other countries. Following the development of social media, multiple government agencies are now avid posters on Twitter and other platforms in their attempts to shape public discourse at home and abroad. _________ TURN TO TURMOIL The novel coronavirus has imposed suddenly a full-court-press for diplomats to take to internet ICT, but accompanying this rapid change are new communications stresses and nuances. The functionalities and efficiencies of diplomatic ICT - including phone, email, social media messaging, and virtual meetings – are now routinely in question. The limitations of video-conferencing and other ICT-driven modes of communication have long been documented . U.S. Foreign Service Officers (FSOs) are specifically trained in face-to-face interactions and talking to groups. Exclusive reliance upon teleconference platforms such as Zoom and WebEx affect multiple core capabilities of the FSO. It is difficult to detect and project desired body language, thus introducing challenges in convincing foreign interlocutors to share private thoughts and intentions over a monitor. Negotiations are both enabled and constrained online, thus pushing the dialogue in sometimes unexpected directions. Reduced or no travel, fewer in-person meetings, and overall, limited personal interactions compromise statecraft. Video engagement changes how information and intelligence is collected, thus impacting institutional memory. Diplomatic activity and initiatives have slowed considerably as many major meetings and conferences have gone online, with others outright canceled. US Government agencies have suspended many pre-pandemic priorities and pivoted almost entirely to focusing now on coping with the virus. Security in the White House has increased with individuals working near POTUS being given daily coronavirus checks , thus self-limiting the frequency of high-level meetings. Intelligence officers at the CIA are working shifts , such as three days on followed by three days off, to improve social distancing by reducing the given number of people in buildings at any given time. In New York City, itself a pandemic epicenter, all aspects of life have been severely compromised for over a month now, detrimentally affecting United Nations Headquarters operations and activities. For example, COVID-19 has prompted changes in the working methods of the UN General Assembly : “Physical distancing and stay-at-home restrictions mean representatives can no longer meet in person, including to vote on resolutions, which are now circulated under a process known as ‘silence procedure.’ Ambassadors are given a 72-hour window to consult their capitals. If they all agree on a resolution, it is passed. If not, the resolution is not adopted as the ‘silence’ has been broken.” Webinars and online meetings are the new norm, but they impact the large-scale and often slow-moving collaborative work of drafting and editing of diplomatic agreements and statements. These processes now often involve dozens of people all working remotely using a "track changes" mode of interaction. But direct contact and interaction often undergird these collaborations, as interlocutors know each other and can converse face to face before they move into the web environment. Going straight to the web without initial constructive in-person discussions in the best case merely compromises the speed of production; in the worst case it stymies it entirely. Cybersecurity risks increase when executing all diplomacy online. A singular reliance on ICT raises new questions about the reliability and vulnerability of networks to disruption or espionage; unpreparedness opens new cyber-attack vectors for hackers, both amateurs and State-sponsored professionals. Governments may not have the requisite cyberinfrastructure and trained employees to seamlessly continue diplomacy operations from home offices. Sharing of sensitive documents either cannot be done at all, or only in a limited manner. Diplomatic online sessions can be recorded and played back with deep analyses of verbatim words and body language – normally hurried notes taking and brief summarizations are what constitute the record for face-to-face conversations. Since different organizations and different countries can adopt disparate and non-interoperable technological approaches, ICT compatibility confusion throws additional barriers at smooth statecraft. Some diplomats may avoid negotiations altogether because of security concerns, thus compromising statecraft progress in general. _____________ ACCESSING THE ADVANTAGE State engagements must continue apace regardless of the pandemic. Negotiations of peace, non-virus global health issues, human rights, climate change, nuclear nonproliferation, and trade all must occur, in whatever limited way. Embassies need to find ways to provide emergency consular services to visiting or expatriate citizens in need without putting diplomats at risk of infection. Diplomats face challenges in assessing emerging political, economic, and social trends when prevented from face-to-face meetings or travel around a country. National elections scheduled during this period of social distancing may take place without the confidence-building presence of foreign election observers, increasing potential for corrupt or anti-democratic practices to alter election outcomes. Diplomatic professionals will have to consider how the new online reality impacts their ability to advance their national interests, to understand rapidly shifting geopolitical events, and to reach publics overseas. Key is to understand how diplomacy morphs into a new normal in the coming months and years, and to identify means that diplomats and senior policymakers can prepare. Tactical challenges are to identify best ways of doing business and to implement them in real time during the pandemic. Strategic challenges are to consider how diplomacy following the pandemic might inalterably change, what new opportunities are present, and how to position diplomats accordingly. Advantages and opportunities exist in this new online reality, nevertheless. Frictionless meeting logistics with accompanying financial savings for internet-only engagements; increased opportunities to hear directly from key players; younger diplomats shining with their greater digital skills – all these and more are coming to light in the pandemic era. As stated by the United Nations’ Under-Secretary General for Political and Peacebuilding Affairs, Rosemary DiCarlo , “Although we recognize that the limitations of the processes in which face-to-face meetings are restricted, the increased use of technology has the potential to create new opportunities [and] enhance the inclusivity of peace processes - for example, including the participation of women and young people.” Concomitant with this new reality is the increased potential to leverage big data, predictive analytics, and other digital technologies. Investments should be amplified in artificial intelligence (AI) to analyze all data that can be extracted on any given country, region, or topic. Predictive analytics using machine learning (and its subset, deep learning) can then be applied to offer novel insights for meeting preparedness and briefings to senior policymakers; such approaches offer new opportunities in optimized human-machine hybrid decision-making. AI can also be used to automatically assess treaty compliance, as well as to monitor and predict events before they occur . In-situ machine language translations and real-time content assessment for both verbal and written communiques are feasible now. Increased security of voting processes via blockchain holds great potential. Virtual reality avatars can be applied to replicate the feeling of sitting across a table from a counterpart. Ultimately, there is a wide range of technological applications and solutions that await adoption and deployment by forward-leaning diplomatic corps. However long the pandemic lasts, diplomacy has been forever changed. How diplomats respond to the recent step-change in personal interactions and resultant altered statecraft will dictate whether national objectives are met, whether treaties are negotiated, and whether normal operations return - or diplomacy turmoil continues. The opinions and characterizations in this piece are those of the authors and do not necessarily represent those of the U.S. Government. Thomas A. Campbell, Ph.D. (tom.campbell@futuregrasp.com) is Founder & CEO of FutureGrasp, an advisory group that works with organizations globally to identify policy and business opportunities from emerging and disruptive technologies, especially artificial intelligence. Previously, he was the first National Intelligence Officer for Technology with the National Intelligence Council, Office of the Director of National Intelligence (ODNI). Andrew Hyde (andrew.hyde@futuregrasp.com) is a Nonresident Fellow at the Stimson Center, a Washington, D.C. based think tank, with the Transforming Conflict and Governance Program, and a Senior Advisor with FutureGrasp. Previously, he was a Foreign Service Officer at the U.S. Department of State and a Congressional staffer. Geoffrey Odlum (geoffrey.odlum@futuregrasp.com) served as a Foreign Service Officer at the U.S. Department of State from 1989 to 2017. He is currently the President of Odlum Global Strategies, a national security and technology policy consulting firm, and is a Senior Advisor with FutureGrasp. Ariel I. Ahram, Ph.D. (ahram@vt.edu) is Associate Professor and Chair of Government and International Affairs (GIA) at the School of Public and International Affairs, Virginia Tech.
By Thomas A. Campbell, Ph.D. 04 Apr, 2019
It is the position of FutureGrasp, LLC that every nation state should have an AI national plan. To remain economically competitive internally, as well as to facilitate collaboration across borders, it is imperative that governments recognize the need to channel whatever resources they possess to produce plans that enable strategic directions in the rapidly growing technology sector of AI.
By Tom Campbell, PhD; Founder & President; FutureGrasp, LLC 07 Sep, 2018
As we move into future years with more advanced compute and exponential data availability, there will be an inexorable push to simulate more of our world. Populating our digital worlds with digital twins and digital doubles will ultimately serve us greatly in our increasingly complex real worlds.
By Tom Campbell, PhD; Founder & President; FutureGrasp, LLC 10 Aug, 2018
It is often said that history repeats itself—we are seeing these same trends today with autonomous weapons.
By Tom Campbell, Founder & President, FutureGrasp, LLC 19 Jun, 2018
Voice is all the buzz (no pun intended) in Silicon Valley. Many technology leaders (Apple, Amazon, Microsoft, etc.) are betting that verbal commands will soon be the major means of communication for our numerous devices. But is verbal chatter the terminal point for our computer communications?
By Thomas A. Campbell, Ph.D., FutureGrasp, LLC 04 May, 2018
Quantum supremacy, the point at which a quantum computer exceeds the capabilities of any classical computer , is now near at hand. I explain why and what that means for computer science.
By Thomas A. Campbell, Ph.D., FutureGrasp, LLC & Robert Meagley, Ph.D., nR, LLC; ONE Nanotechnologies LLC 08 Feb, 2018
We review the state of the art in advanced semiconductors as relevant to artificial intelligence (AI).
By Thomas A. Campbell, PhD, FutureGrasp, LLC & Robert Meagley, PhD, nR, LLC 02 Feb, 2018
We review the state of the art in advanced compute architectures as relevant to artificial intelligence (AI).
By Thomas A. Campbell, Ph.D., FutureGrasp, LLC & Matt Wolfe, Ph.D., Virginia Tech Applied Research Corporation (VT-ARC) 09 Jan, 2018
We review the challenges to qualitative forecasting of science and technology, and detail the advantages to taking a more quantitative approach using big data analytics and AI.
By Tom Campbell 04 Dec, 2017
The only way we can survive and thrive in this constantly changing and increasingly complex world is to get assistance from our machine creations. Artificial intelligence (AI) must be a core aspect of that help, as it offers unprecedented capabilities to monitor, control and assess a wide range of situations. AI will not (at least in the near term) remove the human from the loop; instead, it will augment our capacities for better data collection, analysis and decision making.
Show More
Share by: