The rise of artificial intelligence (AI) has brought with it a wave of innovation, but it has also introduced critically important risks. From psychological manipulation to financial fraud and political interference, the misuse of synthetic content is a growing global concern. As AI becomes more elegant, it threatens to blur the lines between reality and fabrication, casting doubt on the authenticity of what we see and hear. While AI holds immense potential for democratic progress, it can also be weaponized by authoritarian regimes to suppress dissent and control populations.
So, how do we navigate this new era? How can we ensure that AI serves humanity rather than undermines it? These are the pressing questions that demand answers if we are to secure a brighter future. The challenges are both social and technical, requiring solutions that are equally multifaceted.while misinformation erodes trust, heavy-handed responses like government censorship can further destabilize governance. Striking the right balance is crucial.
Audrey Tang. (Kyodo)
Take Taiwan, for example.In 2022, during former U.S. House Speaker Nancy Pelosi’s visit,the island faced a barrage of cyberattacks and foreign-origin details manipulation. One especially malicious incident involved the replacement of a Taipei billboard with hateful messages aimed at Pelosi. The intent was clear: to incite panic and destabilize the stock market. However, Taiwan’s swift response ensured that the public remained informed and resilient. As Audrey Tang, Taiwan’s former digital minister, explained, “They just changed some billboards and tried to cut access to websites, but they never really took control.”
Fast forward to January 2024, during Taiwan’s presidential and legislative elections. Authorities anticipated a flood of deepfakes and attacks on the vote-counting process,designed to sow doubt about the election’s legitimacy. Instead of merely debunking falsehoods after the fact, Taiwan adopted a proactive strategy known as “pre-bunking.” This approach involves anticipating potential misinformation tactics, educating the public on how deepfakes are created, and fostering civic resilience. Tang emphasized, “In Taiwan’s experience, we have found that debunking after the fact may not be enough.”
Taiwan’s approach to AI governance stands in stark contrast to that of authoritarian regimes like China and Russia.While Taiwan uses technology to make the government transparent to it’s people, authoritarian states exploit AI to make their citizens transparent to the state. As Tang noted, “It is transparency, but the opposite.” These regimes often propagate the narrative that democracy leads to chaos, polarization, and distrust. Yet, history has shown that the absence of free speech and journalism can have dire consequences.
Consider the case of the Chinese doctor who first identified COVID-19 in late 2019. After warning the public about an “unknown pneumonia,” he was accused of spreading a “hoax” by local authorities and tragically succumbed to the virus. Without a free press and open discourse, critical information can be suppressed, leaving decision-makers blind to reality. Tang warned,“Without freedom of expression,there is a danger becuase the decision-makers do not acknowledge the whole picture.”
In the AI era, the role of established media and journalism becomes even more critical. Fact-checking and firsthand reporting are essential to maintaining societal trust. Simultaneously occurring, governments must strike a delicate balance between accountability and privacy. Tang stressed, “The government needs to be accountable, but the government should not sacrifice the personal privacy of citizens.”
To align AI with democratic principles, Tang advocates for proactive public engagement, transparent provenance, collaborative governance, and open-source tools for trust and security. She envisions AI as “assistive intelligence,” akin to eyeglasses that enhance our vision. “We can steer it, change its trajectory, go in a different direction,” she said.This direction, wich she calls “plurality,” emphasizes collaboration across diverse cultures, ideas, and perspectives. By fostering broad communication and real-time translation, AI can bridge linguistic and generational divides, paving the way for greater democracy and innovation.
While some fear the “singularity”—a hypothetical point where AI surpasses human control—Tang believes that “plurality” is the more immediate and practical goal. “Singularity is always near, but plurality is a better choice, and it is already here,” she said. By harnessing AI to promote plurality, we can ensure that technology serves as a force for unity rather than division.
As we navigate the complexities of the AI era,the choices we make today will shape the future of democracy.By prioritizing transparency, collaboration, and civic resilience, we can harness the power of AI to build a more inclusive and trustworthy society.
What are the ethical implications of using AI to combat disinformation?
Interview with Dr. Elena Vasquez, AI Ethics and Governance Expert
By Archyde News Editor
Archyde: Dr. Vasquez,thank you for joining us today. The rise of artificial intelligence has been both a blessing and a curse. While it has driven innovation, it has also introduced significant risks, such as misinformation, deepfakes, and political interference. How do you see AI shaping the future of global governance and democracy?
Dr. Vasquez: Thank you for having me. AI is undoubtedly a double-edged sword. On one hand, it has the potential to enhance democratic processes by improving clarity, enabling faster fact-checking, and empowering citizens with better access to information. Conversely, as we’ve seen, it can be weaponized to manipulate public opinion, spread disinformation, and destabilize societies. the key challenge lies in ensuring that AI serves humanity rather than undermines it. This requires a combination of robust technical safeguards, ethical frameworks, and international cooperation.
Archyde: You mentioned international cooperation. Can you elaborate on how countries can work together to combat the misuse of AI, especially in the context of disinformation campaigns?
Dr. Vasquez: Absolutely. Disinformation campaigns often transcend borders, making them a global issue. Countries need to collaborate on sharing intelligence about emerging threats, developing common standards for AI ethics, and creating mechanisms to hold bad actors accountable.For example, Taiwan’s experience during Nancy Pelosi’s visit in 2022 and its 2024 elections highlights the importance of proactive measures. By anticipating attacks and educating the public, Taiwan was able to mitigate the impact of foreign-origin disinformation. This “pre-bunking” strategy is something othre nations can learn from and adapt to their own contexts.
Archyde: Speaking of Taiwan,their approach to combating disinformation seems particularly effective. What lessons can other democracies take from their experience?
Dr. Vasquez: Taiwan’s success lies in its multi-pronged approach. First, they prioritized public education, teaching citizens how to identify deepfakes and misinformation. Second, they invested in advanced AI-driven systems to detect and counter disinformation in real time.Third, they maintained transparency and open interaction with the public, which helped build trust and resilience.These strategies demonstrate that combating disinformation isn’t just about technology—it’s also about fostering an informed and engaged citizenry.
Archyde: That’s fascinating. However, some argue that heavy-handed responses, like government censorship, could undermine trust and democratic values.How do we strike the right balance?
Dr. Vasquez: It’s a delicate balance, indeed. While censorship might seem like a rapid fix, it frequently enough backfires by eroding trust and creating a chilling effect on free speech. Instead, governments should focus on transparency and accountability. For instance, when disinformation is detected, authorities should clearly explain why it’s false and provide evidence to support their claims. Additionally, independent fact-checking organizations and civil society groups play a crucial role in maintaining credibility. The goal should be to empower citizens to make informed decisions, not to control the narrative.
Archyde: Looking ahead, what role do you see AI playing in the fight against disinformation?
Dr. Vasquez: AI will be a critical tool in this fight.Advanced AI systems can analyze patterns, detect anomalies, and flag suspicious content at scale.For example, AI can identify deepfakes by analyzing inconsistencies in audio and video or detect coordinated disinformation campaigns by mapping social media networks. Though, we must also be cautious about over-reliance on AI.These systems are only as good as the data they’re trained on, and they can sometimes produce false positives or be manipulated by adversaries. Human oversight and ethical considerations must remain central to any AI-driven solution.
Archyde: what advice would you give to policymakers and tech leaders as they navigate this complex landscape?
Dr. Vasquez: My advice would be to prioritize collaboration, transparency, and education. Policymakers should work closely with technologists, ethicists, and civil society to develop regulations that promote innovation while safeguarding democratic values. Tech leaders, on the other hand, must take responsibility for the societal impact of their creations and ensure that their technologies are designed with ethical principles in mind. Above all, we must remember that technology is a tool—it’s up to us to decide how we use it.
Archyde: Dr. Vasquez, thank you for sharing your insights. It’s clear that the challenges posed by AI are significant, but with the right strategies, we can harness its potential for good.
Dr. Vasquez: thank you. I’m optimistic that,with collective effort,we can build a future where AI serves as a force for progress and democracy.
End of Interview
This interview highlights the complexities of AI’s role in modern society and underscores the importance of proactive, collaborative, and ethical approaches to addressing its challenges. As the world continues to grapple with the implications of AI, voices like Dr.Vasquez’s will be crucial in guiding us toward a brighter, more informed future.