An array of experts has contributed deep insight into how artificial intelligence is impacting on the way information is being produced, disseminated and consumed, thus reshaping the communications landscape.
They joined a panel on 5 March, in the framework of Mobile Learning Week at UNESCO HQ in Paris, for a dedicated workshop on the subject.
Guy Berger, Director for Freedom of Expression and Media Development at UNESCO, opened the discussion by describing the problem of disinformation by drawing on definitions of the Council of Europe and the European Union.
This perspective sees disinformation as content that is deliberately and intentionally fabricated, not true nor verifiable, and which is produced with the intention of making a profit, and/or pushing a certain ideological or political agenda.
Through social media algorithms, micro-targeting and persuasion, the dissemination of ‘deep fakes’, AI-generated content and automated trolling, artificial intelligence evidently plays a crucial role in the rapid spread of disinformation.
However, said Berger, AI can also be part of the solution, as was illustrated by the multistakeholder panel that explored current problems and possible ways of facing them. Storyzy is a tech start-up created precisely to address the emerging challenges, by using AI to and classify online sources according to the probability of them spreading disinformation, as explained by Pierre-Albert Ruquier, its co-founder and Marketing Director.
Yet tackling the spread of disinformation by virtue of artificial intelligence poses important issues in the view of the panelists. Wafa Ben-Hassine, a Policy Counsel at the Access Now NGO, noted that while hate speech constituting incitement to violence should be banned according to international standards, disinformation is not always illegal. Yet, AI-based responses to this content often led to situations in which speech protected by international law ends up being taken down by a machine, she warned.
Divina Frau-Meigs, UNESCO Chair for “Savoir-Devenir in sustainable digital development: mastering information cultures”, highlighted how the spread of disinformation pose risks to democracy, particularly in relation to the integrity of elections.
Marc-Antoine Dilhac, Philosophy professor at the University of Montreal and Canada Research Chair in Public Ethics, echoed this concern, calling attention to how, by giving away their data, individuals are feeding algorithms that serve business interests and targeted political advertising.
Elodie Vialle, Head of the Journalism and Technology Desk at Reporters Without Borders, explained how “Journalists, and mostly female journalists, are being increasingly harassed online, through disinformation campaigns amplified by bots”. While recognizing that AI can represent opportunities for journalists (e.g. by facilitating fact-checking), false information spreads six times faster than true information. In this situation, journalism needs to be reinforced and supported, she stressed.,
For his part, Cordel Green, Executive Director at the Broadcasting Commission of Jamaica, identified the current disruption in the media ecosystem as the source of many challenges, including for regulatory bodies. A very small group of tech companies control social networks and have become content creators and aggregators. Meanwhile, audiences are shifting online, traditional media are becoming unprofitable and see their capacities are diminished. “We are losing fact-checking gatekeepers”, he cautioned.
The audience raised points of their own: How to factor in content, in order for AI to be able to discern metaphors, irony, and jokes? How can journalists survive in the new digital age? How can we tackle the effects of AI on human rights, and in terms of amplifying existing inequalities and biases?
Recognising that the spread of disinformation is a multifaceted problem, the panel offered several pragmatic solutions.
Ms Ben-Hassine insisted on holistically tackling the challenges posed by the social media platforms’ business model and the way in which they carry out data collection. She advocated the need for appropriate legal frameworks regulating competition and data protection, as well as transparency in electoral advertising.
Both Professor Dilhac and Mr Ruquier explained that the automated analysis of what constituted disinformation would need to be worked out in tandem with human moderators. AI would be best when it is used to find trolls and fraudulent accounts, then flagging them for human moderators to make the final decision, they said. Mr Green was of a similar view when stating that journalists should look at AI not as the enemy, but as a tool.
Panelists supported multi-stakeholder responses like the Journalism Trust Initiative referred to by Ms Vialle, which brings together media managers, editors, publishers and regulators. The best solution to protect journalists and freedom of expression against the threats of false information is self-regulation, she noted. Similarly, social media platforms should face up to their responsibilities while avoiding the privatization of censorship, she argued.
Speakers all agreed on the importance of further research, as well as of empowering users through Media and Information Literacy, with Ms Frau-Meigs highlighting relevant ongoing efforts, such as those promoted by the Global Alliance for Partnerships on MIL, the Council of Europe and the High Level Expert Group set up by the European Commission to counter online disinformation, among others.
Ten Ways the C-Suite Can Protect their Company against Cyberattack
Cyberattacks are one of the top 10 global risks of highest concern in the next decade, with an estimated price tag of $90 trillion if cybersecurity efforts do not keep pace with technological change. While there is abundant guidance in the cybersecurity community, the application of prescribed action continues to fall short of what is required to ensure effective defence against cyberattacks. The challenges created by accelerating technological innovation have reached new levels of complexity and scale – today responsibility for cybersecurity in organizations is no longer one Chief Security Officer’s job, it involves everyone.
The Cybersecurity Guide for Leaders in Today’s Digital World was developed by the World Economic Forum Centre for Cybersecurity and several of its partners to assist the growing number of C-suite executives responsible for setting and implementing the strategy and governance of cybersecurity and resilience. The guide bridges the gap between leaders with and without technical backgrounds. Following almost one year of research, it outlines 10 tenets that describe how cyber resilience in the digital age can be formed through effective leadership and design.
“With effective cyber-risk management, business executives can achieve smarter, faster and more connected futures, driving business growth,” said Georges De Moura, Head of Industry Solutions, Centre for Cybersecurity, World Economic Forum. “From the steps necessary to think more like a business leader and develop better standards of cyber hygiene, through to the essential elements of crisis management, the report offers an excellent cybersecurity playbook for leaders in public and private sectors.”
“Practicing good cybersecurity is everyone’s responsibility, even if you don’t have the word “security” in your job title,” said Paige H. Adams, Global Chief Information Security Officer, Zurich Insurance Group. “This report provides a practical guide with ten basic tenets for business leaders to incorporate into their company’s day-to-day operations. Diligent application of these tenets and making them a part of your corporate culture will go a long way toward reducing risk and increasing cyber resilience.”
“The recommendation to foster internal and external partnerships is one of the most important, in my view,” said Sir Rob Wainwright, Senior Cyber Partner, Deloitte. “The dynamic nature of the threat, not least in terms of how it reflects the recent growth of an integrated criminal economy, calls on us to build a better global architecture of cyber cooperation. Such cooperation should include more effective platforms for information sharing within and across industries, releasing the benefits of data integration and analytics to build better levels of threat awareness and response capability for all.”
The Ten Tenets
1. Think Like a Business Leader – Cybersecurity leaders are business leaders first and foremost. They have to position themselves, teams and operations as business enablers. Transforming cybersecurity from a support function into a business-enabling function requires a broader view and a stronger communication skill set than was required previously.
2. Foster Internal and External Partnerships – Cybersecurity is a team sport. Today, information security teams need to partner with many internal groups and develop a shared vision, objectives and KPIs to ensure that timelines are met while delivering a highly secure and usable product to customers.
3. Build and Practice Strong Cyber Hygiene – Five core security principles are crucial: a clear understanding of the data supply chain, a strong patching strategy, organization-wide authentication, a secure active directory of contacts, and encrypted critical business processes.
4. Protect Access to Mission-Critical Assets – Not all user access is created equal. It is essential to have strong processes and automated systems in place to ensure appropriate access rights and approval mechanisms.
5. Protect Your Email Domain Against Phishing – Email is the most common point of entry for cyber attackers, with the median company receiving over 90% of their detected malware via this channel. The guide highlights six ways to protect employees’ emails.
6. Apply a Zero-Trust Approach to Securing Your Supply Chain – The high velocity of new applications developed alongside the adoption of open source and cloud platforms is unprecedented. Security-by-design practices must be embedded in the full lifecycle of the project.
7. Prevent, Monitor and Respond to Cyber Threats – The question is not if, but when a significant breach will occur. How well a company manages this inevitability is ultimately critical. Threat intelligence teams should perform proactive hunts throughout the organization’s infrastructure and keep the detection teams up to date on the latest trends.
8. Develop and Practice a Comprehensive Crisis Management Plan – Many organizations focus primarily on how to prevent and defend while not focusing enough on institutionalizing the playbook of crisis management. The guide outlines 12 vital components any company’s crisis plan should incorporate.
9. Build a Robust Disaster Recovery Plan for Cyberattacks – A disaster recovery and continuity plan must be tailored to security incident scenarios to protect an organization from cyberattacks and to instruct on how to react in case of a data breach. Furthermore, it can reduce the amount of time it takes to identify breaches and restore critical services for the business.
10. Create a Culture of Cybersecurity – Keeping an organization secure is every employee’s job. Tailoring trainings, incentivizing employees, building elementary security knowledge and enforcing sanctions on repeat offenders could aid thedevelopment of a culture of cybersecurity.
In the Fourth Industrial Revolution, all businesses are undergoing transformative digitalization of their industries that will open new markets. Cybersecurity leaders need to take a stronger and more strategic leadership role. Inherent to this new role is the imperative to move beyond the role of compliance monitors and enforcers.
Moving First on AI Has Competitive Advantages and Risks
Financial institutions that implement AI early have the most to gain from its use, but also face the largest risks. The often-opaque nature of AI decisions and related concerns of algorithmic bias, fiduciary duty, uncertainty, and more have left implementation of the most cutting-edge AI uses at a standstill. However, a newly released report from the World Economic Forum, Navigating Uncharted Waters, shows how financial services firms and regulators can overcome these risks.
Using AI responsibly is about more than mitigating risks; its use in financial services presents an opportunity to raise the ethical bar for the financial system as a whole. It also offers financial services a competitive edge against their peers and new market entrants.
“AI offers financial services providers the opportunity to build on the trust their customers place in them to enhance access, improve customer outcomes and bolster market efficiency,” says Matthew Blake, Head of Financial Services, World Economic Forum. “This can offer competitive advantages to individual financial firms while also improving the broader financial system if implemented appropriately.”
Across several dimensions, AI introduces new complexities to age-old challenges in the financial services industry, and the governance frameworks of the past will not adequately address these new concerns.
Explaining AI decisions
Some forms of AI are not interpretable even by their creators, posing concerns for financial institutions and regulators who are unsure how to trust solutions they cannot understand or explain. This uncertainty has left the implementation of cutting-edge AI tools at a standstill. The Forum offers a solution: evolve past “one-size-fits-all” governance ideas to specific transparency requirements that consider the AI use case in question.
For example, it is important to clearly and simply explain why a customer was rejected for a loan, which can significantly impact their life. It is less important to explain a back-office function whose only objective is to convert scans of various documents to text. For the latter, accuracy is more important than transparency, as the ability of this AI application to create harm is limited.
Beyond “explainability”, the report explores new challenges surrounding bias and fairness, systemic risk, fiduciary duty, and collusion as they relate to the use of AI.
Bias and fairness
Algorithmic bias is another top concern for financial institutions, regulators and customers surrounding the use of AI in financial services. AI’s unique ability to rapidly process new and different types of data raise the concern that AI systems may develop unintended biases over time; combined with their opaque nature such biases could remain undetected. Despite these risks, AI also presents an opportunity to decrease unfair discrimination or exclusion, for example by analyzing alternative data that can be used to assess ‘thin file’ customers that traditional systems cannot understand due to a lack of information.
The widespread adoption of AI also has the potential to alter the dynamics of the interactions between human actors and machines in the financial system, creating new sources of systemic risk. As the volume and velocity of interactions grow through automated agents, emerging risks may become increasingly difficult to detect, spread across various financial institutions, Fintechs, large technology companies, and other market participants. These new dynamics will require supervisory authorities to reinvent themselves as hubs of system-wide intelligence, using AI themselves to supervise AI systems.
As AI systems take on an expanded set of tasks, they will increasingly interact with customers. As a result, fiduciary requirements to always act in the best interests of the customer may soon arise, raising the question if AI systems can be held “responsible” for their actions – and if not, who should be held accountable.
Given that AI systems can act autonomously, they may plausibly learn to engage in collusion without any instruction from their human creators, and perhaps even without any explicit, trackable communication. This challenges the traditional regulatory constructs for detecting and prosecuting collusion and may require a revisiting of the existing legal frameworks.
“Using AI in financial services will require an openness to new ways of safeguarding the ecosystem, different from the tools of the past,” says Rob Galaski, Global Leader, Banking & Capital Markets, Deloitte Consulting. “To accelerate the pace of AI adoption in the industry, institutions need to take the lead in developing and proposing new frameworks that address new challenges, working with regulators along the way.”
For each of the above described concerns, the report outlines the key underlying root causes of the issue and highlights the most pressing challenges, identifies how those challenges might be addressed through new tools and governance frameworks, and what opportunities might be unlocked by doing so.
The report was prepared in collaboration with Deloitte and follows five previous reports on financial innovation. The World Economic Forum will continue its work in Financial Services, with a particular focus on AI’s connections to other emerging technologies in its next phase of research through mid-2020.
US Blacklist of Chinese Surveillance Companies Creates Supply Chain Confusion
The United States Department of Commerce’s decision to blacklist 28 Chinese public safety organizations and commercial entities hit at some of China’s most dominant vendors within the security industry. Of the eight commercial entities added to the blacklist, six of them are some of China’s most successful digital forensics, facial recognition, and AI companies. However, the two surveillance manufacturers who made this blacklist could have a significant impact on the global market at large—Dahua and Hikvision.
Putting geopolitics aside, Dahua’s and Hikvision’s positions within the overall global digital surveillance market makes their blacklisting somewhat of a shock, with the immediate effects touching off significant questions among U.S. partners, end users, and supply chain partners.
Frost & Sullivan’s research finds that, currently, Hikvision and Dahua rank second and third in total global sales among the $20.48 billion global surveillance market but are fast-tracking to become the top two vendors among IP surveillance camera manufacturers. Their insurgent rise among IP surveillance camera providers came about due to both companies’ aggressive growth pipelines, significant product libraries of high-quality surveillance cameras and new imaging technologies, and low-cost pricing models that provide customers with higher levels of affordability.
This is also not the first time that these two vendors have found themselves in the crosshairs of the U.S. government. In 2018, the U.S. initiated a ban on the sale and use of Hikvision and Dahua camera equipment within government-owned facilities, including the Department of Defense, military bases, and government-owned buildings. However, the vague language of the ban made it difficult for end users to determine whether they were just banned from new purchases of Dahua or Hikvision cameras or if they needed to completely rip-and-replace existing equipment with another brand. Systems integrators, distributors, and even technology partners themselves remained unsure of how they should handle the ban’s implications, only serving to sow confusion among U.S. customers.
In addition to confusion over how end users in the government space were to proceed regarding their Hikvision and Dahua equipment came the realization that both companies held significant customer share among commercial companies throughout the U.S. market—so where was the ban’s line being drawn for these entities? Were they to comply or not? If so, how? Again, these questions have remained unanswered since 2018.
Hikvision and Dahua each have built a strong presence within the U.S. market, despite the 2018 ban. Both companies are seen as regular participants in industry tradeshows and events, and remain active among industry partners throughout the surveillance ecosystem. Both companies have also attempted to work with the U.S. government to alleviate security concerns and draw clearer guidelines for their sales and distribution partners throughout the country. They even established regional operations centers and headquarters in the country.
While blacklisting does send a clearer message to end users, integrators, and distributors—for sales and usage of these companies’ technologies—remedies for future actions still remain unclear. When it comes to legacy Hikvision and Dahua cameras, the onus appears to be on end users and integrators to decide whether rip-and-replace strategies are the best way to comply with government rulings or to just leave the solutions in place and hope for the best.
As far as broader global impacts of this action, these will remain to be seen. While the 2018 ban did bring about talks of similar bans in other regions, none of these bans ever materialized. Dahua and Hikvision maintained their strong market positioning, even achieving higher-than-average growth rates in the past year. Blacklisting does send a stronger message to global regulators though, so market participants outside the U.S. will just have to adopt a wait-and-see posture to see how, if at all, they may need to prepare their own surveillance equipment supply chains for changes to come.
A lesson in Naomi Wolf’s promiscuities and an open space where poetry matters
Shut the door. Shut out the quiet light. Tell yourself to swim away from the tigers with arms pillars of...
Overcoming today’s challenges for tomorrow’s security
In a world where technology such as artificial intelligence and robotics is evolving rapidly, defence organisations that are steeped in...
Afghanistan: EU reinforces humanitarian support with €40 million as crisis worsens
The European Commission has allocated an additional €40 million in emergency assistance for those affected by the worsening humanitarian situation...
Regional Conference on Air Quality Management in the Western Balkans
Government representatives from North Macedonia, Kosovo, Bosnia and Herzegovina (BiH) and Serbia met today in Skopje for a regional conference...
A New Currency Offers New Hope for Zimbabwe
For many Zimbabweans queuing up outside banks last week, it must have felt like the beginning of a new era....
It’s Hard to Find a Black Cat in a Dark Room, Especially If It Isn’t There: RAND on the Search for Cyber Coercion
What is cyber coercion and how have states used cyber operations to coerce others? These are the questions addressed in...
Post-Brexit UK will continue to offer significant opportunities
PwC’s new report, Brexit and beyond: Assessing the impact on Europe’s asset and wealth managers, outlines the chief findings from...
Africa3 days ago
The Geopolitics of natural resources of Western Sahara
Europe2 days ago
U.S. President Trump to meet Bulgaria’s Prime Minister at the White House: What to expect?
Defense3 days ago
Is this the end of NATO-era?
South Asia2 days ago
The era emerged from “RuwanWeliSaya”: Aftermath of Presidential Election in Sri Lanka
South Asia2 days ago
Sri Lanka’s election results and their implications
Americas2 days ago
Poll Shows Trump’s Israel Policy Is Opposed Even by Republicans
Urban Development2 days ago
Banking on nature: a Mexican city adapts to climate change
Newsdesk3 days ago
ADB Program to Help Improve Education and Health in Armenia