Connect with us

Intelligence

Artificial Intelligence: Potential Intensifier of Strategic Dynamics in South Asia

Published

on

With growing dependency on artificial rationalization, human reasoning and decision-making is under continuous suppression. Where machine learning and deep learning tends to empower machines to carry out functions and break assigned tasks into easier ones, it nevertheless fastens the route towards a world order that is likely to be in absolute control of Artificial Intelligence(AI). Does it indicate cutting humans entirely out of the loop?

This deliberate submission of power to machines has some assured repercussions in the realm of strategic stability which rational actors must take into consideration. The simulation of human cognition – the capacity of human mind to learn, interpret and reason- in machines is what artificial intelligence refers to. It eventually stands as a defining feature of modern societies. By the enhanced use of algorithms, AI optimizes the ability for collecting avast range of data whether numeric or categorical in the form of big data to measure the information and derive results accordingly. Thus, Artificial Intelligence is itself emerging as a vast technological industry for creating intelligent machines. Such machines would be capable of independent decision-making based on the level of subjectivity conceded to AI. This subjectivity defines the rationale of decisions made by machines. Along with enhanced precision and prompt responses, it suggests that over-reliance on AI could probably take the shape of absolute control.

Artificial Intelligence (AI) in the International Arena (IA) acts as a modifier of global affairs and challenges whether bilateral or multilateral. Additionally, it is transforming military strategies with its significant precision and speed via contracting the action-reaction loop. As such AI is being developed for assessing and responding to problems with minimum human supervision. Which, the other way, predicts an autonomous crisis escalation with minimal or no chances of containment. One such example is the development of lethal autonomous weapon systems (LAWS). Analyzing the broad view of global affairs under the predominant existence of nuclear weapons, robotic and computational technology is so far effectively assisting states in maintaining the safety and security mechanisms of nuclear and fissile material/data. It is evident from the events of the cold war era that other than human error, technological error within the realm of nuclear strategy could easily escalate towards nuclear war fighting or its accidental use with a catastrophic domino effect. Despite the precision, speed and human-like reasoning, machines are likely to lack a considerable situational variation with respect to risk assessment of actions and their reactions. The reliance on artificial rationalization means increased unpredictability and competition that resultantly means greater strategic instability around the globe.

Strategic stability demands a credence among nuclear weapon states that their adversaries would not likely be able to undermine their nuclear deterrence by any means. This surety is crucial in the case of South Asia. Comprising of three nuclear weapon states with inter-state rivalries, South Asia demands a stable strategic environment which requires a considerable level of risk assessment and management. Machine learning and big data analysis are some already adopted strategies in South Asia as in other parts of the world to predict and track an adversary’s aggressive posturing. Although, it is technically challenging for a state to be able to locate and target all of its adversary’s dispersed nuclear weapons and delivery systems during crisis-time, AI maximizes this detection and tracking ability. Hence, it could provide a win-win strategic advantage to one party over the other. This likelihood convinces states to pursue greater reliance on advanced AI-supported defence technology while greatly increasing the chances of a possible malfunction or misinterpretation of command.

Strategic stability of South Asia is already fragile. The prediction dynamics of this strategic stability after AI inception has long been a bone of contention. It can be traced that China’s New Generation Artificial Intelligence Development Plan and its AI advancements within strategic realm could lead to more aggression stemming from India’s hegemonic designs. Resultantly, Pakistan’s nuclear deterrence would be reasonably undermined. This can lead to a mutual fog of war in terms of strategic vulnerabilities and disparities. Moreover, the cyber-vulnerabilities and cyber-breach events in South Asia already foretell the emerging uncertainty currently undermining strategic stability in the region.

Furthermore, the prevalence of AI within nuclear realm elevates the risks of an accidental or unauthorized use of nuclear weapons which as an outcome could trigger escalation. Incorporating AI within command and control mechanisms of nuclear weapons states would possibly increase the risk of a misinformed and irrevocable weapons launch. China in pursuit of advanced AI, a bellicose India and balancing Pakistan (vis-a-vis India) would all vulnerable to such misadventures inflicted by an over and uncontrolled reliance on AI. In this regard, keeping the strategic stability of South Asia intact is a much more challenging matter than anywhere else on the globe.

Being an alluring domain, Artificial Intelligence has become a necessary evil which based on the above discussed risks still poses an existential threat to humanity. It presses states around the world and particularly in South Asia as a technologically nascent yet rapidly advancing region to compete in such a way that it may eventually turn into their absolute submission to AI. Another alarming aspect is that ultimately human intelligence adheres to the necessity of the human security perspective whereas AI, if not programmed correctly, may not recognize or emphasize the human safety or security enough. Instead of relinquishing total control and submitting to machines intentionally which could be real risk attracting phenomenon, Artificial intelligence must be employed to assist and empower human cognition to better respond to the collective and individual strategic challenges. 

Research Associate at the Strategic Vision Institute (SVI), a non-partisan think tank based out of Islamabad, Pakistan.

Continue Reading
Comments

Intelligence

Conspiracy Theories, Fake News and Disinformation: Why There’s So Much of It and What We Can Do About it

Published

on

In March 2019, under the aegis of the United States Department of State, a group of researchers released a report called “Weapons of Mass Distraction: Foreign State-Sponsored Disinformation in the Digital Age.” The report mostly focused on foreign states’ propaganda, disinformation and fake news. Taking into account the upcoming US elections, the report can provide practical recommendations for policymakers and stakeholders.

The report begins with a horrific story broadcasted on the Russian state-owned “Channel One” in 2014. The story covered how Ukrainian soldiers crucified a child in front of its mother’s eyes. Later, this story was proved to be fake, and there was neither a killed child, nor shocked mother. Still, the story went viral. It had reached a much broader audience on social mediathan it did on television.

The authors refer to that story as “an example of Kremlin-backed disinformation campaign.” The authors of the report continued to state that “in subsequent years, similar tactics would again be unleashed by the Kremlin on other foreign adversaries, including the United States during the lead-up to the 2016 presidential election.”

Undoubtedly, the fake story did a lot of damage to the reputation of Channel One and other state-funded media. It is clear why authors begin with that story — it was poorly done, obviously faked and quickly exposed. However, it showed how effective and powerful social media could be (despite all of the reputation risks). There is also an important point highlighted in the report, particularly that “the use of modern-day disinformation does not start and end with Russia. A growing number of states, in the pursuit of geopolitical ends, are leveraging digital tools and social media networks to spread narratives, distortions, and falsehoods to shape public perceptions and undermine trust in the truth.” We are used to research, dedicated to propaganda and fake news issues, that establishes only Russia is responsible for disinformation and fake news. This report, on the other hand, addresses propaganda and disinformation as a comprehensive problem.

In the introduction, the authors claim that disinformation is a problem that consists of two major factors: technology giants and their impact and the psychological element of how people consume information on the Internet. Technology giants have disrupted disinformation and propaganda, and the proliferation of social media platforms made the information ecosystem vulnerable to foreign, state-sponsored actors. “The intent [of bad foreign actors] is to manipulate popular opinion to sway policy or inhibit action by creating division and blurring the truth among the target population.”

Another important aspect of disinformation highlighted in the report is the abuse of fundamental human biases and behaviour. The report states that “people are not rational consumers of information. They seek swift, reassuring answers and messages that give them a sense of identity and belonging.” The statement is proved by the research showing that, on average, a false story reaches 1 500 people six times more quickly than a factual account. And indeed, conspiracy stories have become something usual these days. We see it has become even more widespread during the current pandemic — 5G towers, Bill Gates and “evil Chinese scientists” who supposedly invented the coronavirus became scapegoats. And there are a lot more paranoid conspiracy stories spreading on the Internet.

What is the solution? Authors do not blame any country, tech giants or the behavior of people. Rather the opposite, they suggest that the solution should be complex: “the problem of disinformation is therefore not one that can be solved through any single solution, whether psychological or technological. An effective response to this challenge requires understanding the converging factors of technology, media, and human behaviours.”

Define the Problem First

What is the difference between fake news and disinformation? How does disinformation differ from misinformation? It is a rather rare occasion that reports give a whole chapter dedicated to terminology. And the report “The Weapons of Mass Distraction” definitely provides readers with a vast theoretical background. Authors admit that there are a lot of definitions, and it is difficult to ascribe the exact parameters to disinformation. However, it states that “misinformation is generally understood as the inadvertent sharing of false information that is not intended to cause harm, just as disinformation is widely defined as the purposeful dissemination of false information.”

Psychological Factors

As it was mentioned in the beginning, authors do not attach labels and do not focus on one side of the problem. A considerable part of the report is dedicated to psychological factors of disinformation. The section helps readers understand behavioural patterns of how humans consume information, why it is easy to fall for a conspiracy theory, and how to use this information to prevent the spread of disinformation.

The findings are surprising. There are several cognitive biases that make disinformation easy to flourish. And the bad news is that there is little we can do about it.

First of all, confirmation bias and selective exposure lead people to prefer information that confirms their preexisting beliefs make information consistent with one’s preexisting beliefs more persuasive. Moreover, confirmation bias and selective exposure work together with other naïve realism that “leads individuals to believe that their perception of reality is the only accurate view and that those who disagree are simply uninformed or irrational.”

In reality, these cognitive biases are widely used by tech giants. That doesn’t mean that there is a conspiracy theory behind it. That means that it is easy for big tech companies to sell their products using so-called “filter bubbles.” Such a bubble is an algorithm that selectively guesses what information a user would like to see based on information about the user, such as location, past click-behaviour and search history. Filter bubbles work well on such websites like YouTube. A Wall Street Journal investigation found that YouTube’s recommendations often lead users to channels that feature conspiracy theories, partisan viewpoints and misleading videos, even when those users haven’t shown interest in such content.

These days, the most popular way to counter misinformation is fact-checking and debunking the false information. In the report, the researchers presented some evidence that the methods we are used to employing, may not be that effective. “Their analysis determined that users are more active in sharing unverified rumours than they are in later sharing that these rumours were either debunked or verified. The veracity of information, therefore, appears to matter little. A related study found that even after individuals were informed that a story had been misrepresented, more than a third still shared the story.”

The other research finding is that “participants who perceived the media and the word “news” negatively were less likely than others to identify a fake headline and less able to distinguish news from opinion or advertising.” Obviously, there is a reason for that. It’s a lack of trust. The public has low trust towards journalists as a source of information about the coronavirus, says the latest research. Additionally, according to the American Press Institute, only 43 per cent of people said they could easily distinguish factual news from opinion in online-only news or social media. Thus, the majority of people can hardly distinguish news from opinions in a time when trust towards journalism is at its historical minimum. It is therefore no surprise that people perceive news that negatively.

This can have implications for news validation. The report states it can differ from country to country. “Tagging social media posts as “verified” may work well in environments where trust in news media is relatively high (such as Spain or Germany), but this approach may be counterproductive in countries where trust in news media is much lower (like Greece).

A vast research basis also reveals the following essential findings. First, increasing online communities’ exposure to different viewpoints is rather counterproductive. The research presented in the report found that conservative people become more conservative and liberals become more liberal.

Second, the phenomenon called belief perseverance, which is the inability of people to change their minds even after being shown new information, means that facts can matter little in the face of strong social and emotional dynamics.

Third, developing critical thinking skills and increasing media literacy may also be counterproductive or have minimal use. Research shows us that “many consumers of disinformation already perceive themselves as critical thinkers who are challenging the status quo.” Moreover, even debunking false messages cannot be that effective. Showing corrective information did not always reduce the participant’s belief in misinformation. Besides, “consumers of fake news were presented with a fact-check, they almost never read it.”

What can be done here? Authors provide the reader with a roadmap for countering misleading information. Although the roadmap, which is also based on researches, can have very limited use, according to the report.

The main idea is to be proactive. While debunking false messages, developing critical thinking, and other tools have minimal potential, some psychological interventions can help in building resilience against disinformation. Authors compare disinformation and misinformation as a disease, and they propose we need a vaccine that builds resilience to a virus. This strategy means that people should be warned “that they may be exposed to information that challenges their beliefs, before presenting a weakened example of the (mis)information and refuting it.”

Another aspect of the roadmap is showing different perspectives, “which allows people to understand and overcome the cognitive biases that may render them adversarial toward opposing ideas.” According to the authors, this approach should focus less on the content of one’s thoughts and more on their structure. The fact that certain factors can make humans susceptible to disinformation can also be used as part of the solution.

What About the Tech Giants?

The authors admit that social media platforms should be playing a central role to neutralize online disinformation. Despite the fact that tech giants demonstrated their willingness to address disinformation, their incentives are not always prioritized to limit disinformation. Moreover, their incentives are aligned with spreading more of it because of its business model. “Users are more likely to click on or share sensational and inaccurate content; increasing clicks and shares translates into greater advertising revenue. The short-term incentives, therefore, are for the platforms to increase, rather than decrease, the amount of disinformation their users see.”

The technological section of the report is split into three parts dedicated to three tech companies — Facebook, Twitter and Google. While the report focuses on what companies have already done to counter disinformation, we will highlight only the recommendations and challenges that still remain.

Despite all the incentives that have been implemented by Facebook in recent years, the social media platform still remains vulnerable for disinformation. The main vulnerability is behind its messaging apps. WhatsApp has been a great source of disinformation during the Rohingya crisis in 2018 and during the Brazilian presidential elections in the same year. The second vulnerability lies in third-party fact-checking services staffed by human operators. Human operators are struggling to handle the volume of the content: “fake news can easily go viral in the time between its creation and when fact-checkers are able to manually dispute the content and adjust its news feed ranking.”

Despite all the vulnerabilities, including a colossal bot network, Twitter became more influential in countering the threat using such technologies like AI. The question of how proactive the company will be countering the threat still remains. Yet, Twitter now uses best practices, according to the report.

With its video-sharing platform YouTube and ad platform, YouTube might be the most vulnerable platform. The website, with its personalized recommendation algorithm (filter bubbles), has faced strong criticism for reinforcing the viewers’ belief that the conspiracy is, in fact, real. However, YouTube announced in 2019 that it would adjust its algorithms to reduce recommendations of misleading content.

However, it is not just the tech giants who should take responsibility for disinformation. According to the report, it’s countries who should bear the ultimate responsibility for “defending their nations against this kind of disinformation.” Yet, since the situation is still in private hands, what can the government do here?

For example, they could play a more significant role in engaging in regulating social media companies. According to the report, it doesn’t mean total control of social media companies. However, authors admit that this solution may have some implications for possible restriction of freedom of speech and outright censorship, and there is no easy and straightforward way to solve this complex problem.

What can we do about it? According to the report, technology will change, but the problem will not be solved within the next decade. And the fact is, we should learn how to live with the disinformation. At the same time, public policies should focus on mitigating disastrous consequences while maintaining civil liberties, freedom of expression and privacy.

The report provides readers with quite a balanced approach to the problem. While other research projects attach labels on countries or technologies, the authors of the report “Weapons of Mass Distraction” admit the solution will not be easy. It is a complex problem that will require a complex solution.

From our partner RIAC

Continue Reading

Intelligence

Engaging with Local Stakeholders to Improve Maritime Security and Governance

Michael Van Ginkel

Published

on

Illicit activity in the maritime domain takes place within a complex cultural, physical, and political environment. When dialogue is initiated with a diverse range of stakeholders, policy recommendations can take into account region-specific limitations and opportunities. As noted in the Stable Seas: Sulu and Celebes Seas maritime security report, sectors like fisheries, coastal welfare, and maritime security are intrinsically linked, making engagement with a diverse range of local stakeholders a necessity. This collaborative approach is essential to devising efficient and sustainable solutions to maritime challenges. Engagement with local stakeholders helps policymakers discover where in these self-reinforcing cycles additional legislation or enforcement would have the greatest positive impact. Political restrictions against pursuing foreign fishing trawlers in Bangladesh, for example, have allowed the trawlers to target recovering populations of hilsa while local artisanal fishers suffer. In the context of the Philippines, the Stable Seas program and the Asia Pacific Pathways to Progress Foundation recently conducted a workshop that highlighted the importance of consistent stakeholder engagement, resulting in a policy brief entitled A Pathway to Policy Change: Improving Philippine Fisheries, Blue Economy, and Maritime Law Enforcement in the Sulu and Celebes Seas.

Physical Environment

Consistent communication with local stakeholders on regional anomalies allows policymakers to modify initiatives to adjust for the physical, cultural, and political context of a maritime issue. The physical environment affects how, where, and why illicit actors operate in the maritime domain. Knowledge held by local stakeholders about uninhabited coastlines, local currents, and the locations of important coastal communities helps policymakers find recognizable patterns in the locations and frequency of maritime incidents. The 36,289 km of coastline in the Philippine archipelago means that almost 60 percent of the country’s municipalities and cities border the sea. The extensive coastline and high levels of maritime traffic make monitoring coastal waters and achieving maritime domain awareness difficult for maritime law enforcement agencies. A Pathway to Policy Change outlines several recommendations by regional experts on ways to improve maritime domain awareness despite limitations imposed by a complex physical environment. The experts deemed collaboration with local government and land-based authorities an important part of addressing the problem. By engaging with stakeholders working in close proximity to maritime areas, policymakers can take into account their detailed knowledge of local environmental factors when determining the method and motive behind illicit activity.

Cultural Environment

Culture shapes how governments respond to non-traditional maritime threats. Competition and rivalry between maritime law enforcement agencies can occur within government structures. A clearer understanding of cultural pressures exerted on community members can help policymakers develop the correct response. Strong ties have been identified between ethnic groups and insurgency recruiting grounds in Mindanao. The Tausug, for instance, tend to fight for the MNLF while the MILF mostly recruits from the Maguindanaons and the Maranao. Without guidance from local stakeholders familiar with cultural norms, correlations could be left unnoticed or the motivations for joining insurgency movements could be misconstrued as being based solely on extremist or separatist ideology. Local stakeholders can offer alternative explanations for behavioral patterns that policymakers need to make accommodations for.

Political Environment

Local stakeholder engagement allows policymakers to work on initiatives that can accommodate limitations imposed by the political environment. Collaboration with local stakeholders can provide information on what government resources, in terms of manpower, capital, and equipment, are available for use. Stakeholders also provide important insights into complex political frameworks that can make straightforward policy implementation difficult. Understanding where resource competition and overlapping jurisdiction exist enables policymakers to formulate more effective initiatives. Despite strong legislation regulating IUU fishing in the Philippines, local stakeholders have pointed out that overlapping jurisdictions have created exploitable gaps in law enforcement. In A Pathway to Policy Change, local experts suggested that the government should lay down an executive order to unify mandates in the fisheries sector to address the issue. Similarly, the Bangsamoro Autonomous Region of Muslim Mindanao (BARMM) is highlighted as a region that heavily influences maritime security in the Sulu and Celebes seas. Working with government officials to understand how policy initiatives need to adjust for the region’s semi-autonomous status ensures maritime issues are properly addressed. BARMM, for instance, issues fishing permits for its own waters in addition to government permits, which can cause inconsistencies. Working alongside local stakeholders allows policymakers to create initiatives that take into account special circumstances within the political system.

Private Sector Engagement

Extending engagement with local stakeholders to the private sector is particularly important during both the policy research and implementation processes. Encouraging private stakeholders to actively help counter illicit activity can help policymakers create a more sustainable and efficient solution to security threats. As A Pathway to Policy Change highlights, private companies already have a strong incentive from a business perspective to involve themselves in environmental and social issues. Governments can encourage further involvement of private stakeholders like blue economy businesses and fishers by offering tax breaks and financial compensation for using sustainable business practices and for helping law enforcement agencies gather information on illicit activity. Offering financial rewards to members of the Bantay Dagat program in the Philippines, for example, would encourage more fishers to participate. Governments can also double down on educational programs to raise awareness of important issues threatening local economic stability. By communicating consistently with local stakeholders, policymakers can both more accurately identify maritime security needs and more comprehensively address them.

Conclusion

The unique physical, cultural, and political context in which maritime issues take place makes the knowledge of local stakeholders an invaluable asset. While many important types of information can be collected without working closely with stakeholders, there are also innumerable important aspects of any given context which cannot be quantified and analyzed from afar. Engagement with stakeholders provides a nuanced understanding of more localized and ephemerial factors that affect regional maritime security. Engaging with local stakeholders allows policymakers to capitalize on opportunities and circumvent limitations created by the political, cultural, and physical environment surrounding maritime issues in order to create sustainable, long-term solutions.

Continue Reading

Intelligence

Turkey Faced With Revolt Among Its Syrian Proxies Over Libyan Incursion

Published

on

Relations between Turkey and Syrian armed groups that used to be considered cordial due to massive support provided by the Turkish authorities to the Syrian opposition are rapidly deteriorating over Turkey’s incursion into the Libyan conflict, according to sources among the Syrian militants fighting in Libya.

Last month, over 2,000 fighters defected from Sultan Murad Division, one of the key armed factions serving the Turkish interests in Syria. The group’s members chose to quit after they were ordered to go to Libya to fight on the side of the Turkey-backed Government of National Accord (GNA). This marks a drastic shift in the attitude of the Syrian fighters towards participation in the Libyan conflict: just a few months ago there was no shortage of mercenaries willing to fly to Libya via Turkey for a lucrative compensation of $2,000 – 5,000 and a promise of Turkish citizenship offered by Ankara.

Both promises turned out to be an exaggeration, if not a complete lie. The militants who traveled to Libya got neither the money nor the citizenship and other perks that were promised to them, revealed a fighter of Ahrar al-Sharqiya faction Zein Ahmad. Moreover, he pointed out that after the fighters arrived in Libya they were immediately dispatched to Tripoli, an arena of regular clashes between GNA forces and units of the Libyan National Army despite Turkish promises of tasking them with maintaining security at oil facilities.

Data gathered by the Syrian Observatory for Human Rights shows that around 9,000 members of Turkey-backed Syrian armed factions are currently fighting in Libya, while another 3,500 men are undergoing training in Syria and Turkey preparing for departure. Among them are former members of terror groups such as Al-Qaeda affiliate in Syria Hayat Tahrir al-Sham, as confirmed by reports of capture of a 23-years-old HTS fighter Ibrahim Muhammad Darwish by the LNA forces. Another example is an ISIS terrorist also captured by the LNA who confessed that he was flown in from Syria via Turkey.

By sending the Syrian fighters to Libya Ankara intended to recycle and repurpose these groups for establishing its influence without the risks and consequences of a large-scale military operation involving major expenses and casualties among Turkish military personnel. However, the recent developments on the ground show that this goal was not fully achieved.

The Syrian fighters sustain heavy casualties due to the lack of training and weaponry. Total count of losses among the Turkey-backed groups reached hundreds and continue to grow as GNA and LNA clash with intermittent success. Until Turkey’s President Recep Erdogan curbs his ambition, destructive nature of involvement of the Syrian armed groups in Libya may result in the downfall of Turkey’s influence over the Syrian opposition.

Continue Reading

Publications

Latest

Trending