Skip to content

Commit 0cffb5c

Browse files
committed
Added three articles on AI Risks
1 parent 2221918 commit 0cffb5c

34 files changed

+17895
-772
lines changed

dictionary.txt

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -381,3 +381,4 @@ tarnish
381381
goveranance
382382
subpostmasters
383383
exonerate
384+
disinformation

docs/ai/Risks/Emergent-Behaviour.md

Lines changed: 67 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,67 @@
1+
---
2+
title: Emergent Behaviour
3+
description: AI develops unforeseen behaviours, capabilities, or self-replication that could lead to unpredictable consequences.
4+
5+
featured:
6+
class: c
7+
element: '<risk class="feature-fit">Emergent Behaviour</risk>'
8+
tags:
9+
- AI-Risk
10+
- Emergent-Behaviour
11+
sidebar_position: 2
12+
tweet: yes
13+
---
14+
15+
16+
**Impact: 3** - While some emergent behaviors may be benign, others could lead to unintended or harmful consequences that are difficult to control.
17+
18+
## Sources
19+
20+
**DeepMind’s AI Risk Studies (2023):** Research into emergent AI behaviors highlights that as models scale, they may develop unanticipated capabilities, sometimes without direct human prompting.
21+
22+
**MIT Technology Review (2022):** Reports on AI unpredictability in large language models, showing how these systems often display unexpected behaviors that were not explicitly trained into them.
23+
24+
## How This Is Already Happening
25+
26+
- **Unanticipated AI Capabilities:** AI models, particularly deep learning and reinforcement learning systems, have exhibited behaviors that developers did not explicitly program, such as deception, strategic planning, or novel problem-solving approaches.
27+
28+
- **Scaling Laws and Emergent Complexity:** As AI models grow larger, they demonstrate unexpected capabilities—such as sudden improvements in reasoning or deception—without explicit changes in training methodology.
29+
30+
- **Black Box Problem:** Many advanced AI models function as opaque systems, meaning their internal decision-making processes are difficult to interpret, making it hard to predict emergent behaviors before deployment.
31+
32+
- **AI Hallucinations:** Large language models frequently generate false or misleading information, confidently presenting incorrect statements as facts. These hallucinations highlight a fundamental unpredictability in AI behavior, posing risks in fields such as healthcare, law, and journalism where accuracy is critical.
33+
34+
35+
## Mitigating Practices
36+
37+
- **Interpretability**
38+
39+
- Developing tools to analyze AI decision-making processes and detect emergent behaviors before they become risks.
40+
- **Efficacy: Medium** - Helps understand AI behavior but does not prevent emergent capabilities from appearing.
41+
- **Ease of Implementation: Medium** - Research in explainable AI is advancing, but understanding deep learning models remains complex.
42+
- **Replication Control**
43+
44+
- Preventing self-replicating AI or unsupervised proliferation of emergent behaviors by implementing strict replication oversight.
45+
- **Efficacy: High** - Can limit the spread of unexpected behaviors.
46+
- **Ease of Implementation: Medium** - Requires robust tracking and governance frameworks.
47+
- **Monitoring**
48+
49+
- Implementing real-time monitoring of AI behavior in deployment to detect and intervene in emergent risks.
50+
- **Efficacy: High** - Helps detect unintended consequences early.
51+
- **Ease of Implementation: Medium** - Requires sophisticated tracking mechanisms and rapid-response intervention systems.
52+
53+
## Can Risk Management Address This Risk?
54+
55+
**Partially.** Risk management can identify emergent behaviors and develop strategies for monitoring and control. However, since emergent behaviors are **inherently unpredictable**, complete mitigation is difficult. Managing this risk requires a combination of **interpretability research, robust oversight, and adaptive governance models** to respond as new behaviors emerge.
56+
57+
## Can Software Engineering Address This Risk?
58+
59+
**Mainly.** Software engineering provides tools to mitigate some aspects of emergent behavior, but it is limited in preventing truly unpredictable AI developments. Here are some key considerations:
60+
61+
- **Testing and Debugging Limitations:** Unlike traditional software, where debugging and rigorous testing can prevent errors, AI systems—especially deep learning models—can exhibit unexpected behaviors that were never explicitly programmed.
62+
63+
- **Explainability and Interpretability Challenges:** AI interpretability techniques, such as explainable AI (XAI), can help researchers understand how models make decisions. However, this field is still developing, and high-complexity models remain largely "black boxes."
64+
65+
- **Automated Monitoring and Response:** Engineers can design AI systems to monitor their own outputs and flag anomalies. However, this does not guarantee that emergent behaviors will be caught before causing harm, especially in real-time, high-stakes environments.
66+
67+
- **AI Safety and Alignment Research:** Ongoing efforts in AI safety, such as reinforcement learning from human feedback (RLHF) and constitutional AI, aim to shape AI behavior in line with human values. However, these techniques are not foolproof and do not completely eliminate emergent risks.

docs/ai/Risks/Loss-Of-Diversity.md

Lines changed: 47 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,47 @@
1+
---
2+
title: Loss Of Diversity
3+
description: A single AI system dominates globally, leading to catastrophic consequences if it fails or contains errors.
4+
5+
featured:
6+
class: c
7+
element: '<risk class="lock-in">Loss Of Diversity</risk>'
8+
tags:
9+
- AI-Risk
10+
- Loss-Of-Diversity
11+
---
12+
13+
**Impact: 3** - A lack of diversity could create system-wide vulnerabilities, where a single flaw in a dominant AI model causes widespread failure.
14+
15+
## Sources
16+
17+
**Bostrom & Shulman, "Global AI Governance" (2021):** Discusses the risks of a single dominant AI system shaping global decision-making. If AI governance converges around a monolithic framework, any flaw or misalignment in the system could have catastrophic consequences at a planetary scale.
18+
19+
**OpenAI’s Research on AI Model Homogeneity (2022):** Highlights concerns that if AI models become too similar—due to the concentration of AI development within a few major entities—there could be systemic vulnerabilities, lack of innovation, and a failure to address diverse global needs.
20+
21+
## How This Is Already Happening
22+
23+
- **Centralisation of AI Development:** The majority of cutting-edge AI models are being developed by a small group of technology giants, creating an ecosystem where a few dominant models dictate global AI capabilities. This raises the risk that any flaw, bias, or vulnerability present in these models will propagate worldwide.
24+
25+
- **Monoculture Effects in AI Decision-Making:** AI is increasingly being embedded into critical decision-making systems (finance, healthcare, military strategy). If all of these systems rely on a few underlying models, unexpected failures or adversarial attacks could spread rapidly across multiple industries.
26+
27+
- **Lack of Innovation Due to Model Homogeneity:** When the same AI architectures are used across different sectors, it stifles alternative approaches that might better serve specialized needs. A uniform AI landscape risks optimizing for narrow commercial objectives rather than the diverse interests of different populations and industries.
28+
29+
30+
## Mitigations
31+
32+
- **Ecosystem Diversity**
33+
- Encouraging the development of multiple, independent AI models instead of relying on a single dominant system.
34+
- **Efficacy: High** - Diversified AI systems reduce systemic risks and encourage innovation.
35+
- **Ease of Implementation: Medium** - Requires regulatory incentives or decentralized AI development initiatives.
36+
- **Multi-Stakeholder Oversight**
37+
- Ensuring that AI governance involves multiple institutions, including governments, researchers, and civil society, to prevent monopolization.
38+
- **Efficacy: Medium** - Helps distribute AI power more equitably but may struggle with enforcement.
39+
- **Ease of Implementation: Medium** - Requires cooperation between multiple sectors, which can be slow and politically complex.
40+
- **Transparency**
41+
- Requiring AI developers to publish model architectures, data sources, and decision-making rationales.
42+
- **Efficacy: Medium** - Increases accountability but does not prevent monopolization directly.
43+
- **Ease of Implementation: Medium** - Transparency regulations are becoming more common but face resistance from corporate interests.
44+
45+
## Can Risk Management Address This Risk?
46+
47+
Partially. Traditional risk management can identify and highlight the dangers of AI monoculture, but effective mitigation requires strong regulatory intervention and industry-wide commitment—both of which are difficult to enforce under current economic incentives.

docs/ai/Risks/Social-Manipulation.md

Lines changed: 97 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,97 @@
1+
---
2+
title: Social Manipulation
3+
description: AI could predict and shape human behaviour on an unprecedented scale.
4+
5+
featured:
6+
class: c
7+
element: '<risk class="communication">Social Manipulation</risk>'
8+
tags:
9+
- AI-Risk
10+
- Social-Manipulation
11+
sidebar_position: 2
12+
tweet: yes
13+
---
14+
15+
AI systems designed to influence behaviour at scale could (and do) undermine democracy, free will, and individual autonomy.
16+
17+
## Sources
18+
19+
- **The spread of true and false news online** [Vosoughi, Roy, & Aral, 2018](https://doi.org/10.1126/science.aap9559): Demonstrates that false news travels faster and reaches more people than true news on social platforms, indicating how AI-driven disinformation campaigns could exploit these vulnerabilities.
20+
21+
- **The science of fake news** [Lazer et al., 2018](https://www.researchgate.net/publication/323650280_The_science_of_fake_news/link/5b30d8760f7e9b0df5c767b7/download?_tp=eyJjb250ZXh0Ijp7ImZpcnN0UGFnZSI6InB1YmxpY2F0aW9uIiwicGFnZSI6InB1YmxpY2F0aW9uIn19): Examines the ecosystem of fake news creation and dissemination, emphasising the critical need for policy, research, and technological measures to counter AI-enabled misinformation. Also, recognises how at odds any measures of control are with the purest notions of "free speech".
22+
23+
- **Deepfakes: A Grounded Threat Assessment** [Chesney & Citron, 2019](https://dx.doi.org/10.2139/ssrn.3213954): Explores the implications of deepfake technology, highlighting the urgent necessity for innovative detection tools and policy interventions (see below).
24+
25+
- **Nazi Propaganda** [United States Holocaust Memorial Museum](https://encyclopedia.ushmm.org/content/en/article/nazi-propaganda): Examines how the Nazi regime harnessed mass media—including radio broadcasts, film, and print—to shape public opinion, consolidate power, and foment anti-Semitic attitudes during World War. (Fake content isn't a new problem.)
26+
27+
---
28+
29+
## How This Is Already Happening
30+
31+
### AI-Powered Targeted Advertising & Manipulation
32+
33+
- Algorithms mine user data to deliver highly customised ads.
34+
- Personalized messaging exploits individual biases, vulnerabilities, or preferences.
35+
36+
**Real-Life Examples:**
37+
38+
- The [Cambridge Analytica scandal (2016)](https://en.wikipedia.org/wiki/Facebook–Cambridge_Analytica_data_scandal): Data from millions of Facebook users was used to create highly customized political ads, influencing voter perceptions in the U.S. and beyond.
39+
- Political microtargeting in multiple elections: Campaigns worldwide have leveraged platforms like Meta (Facebook) or Google to deliver personalized messages designed to sway opinion on sensitive issues.
40+
41+
### AI-Generated Disinformation & Deepfakes
42+
43+
- Sophisticated tools create realistic but false content that distorts public perception.
44+
- Deepfake videos or audio can undermine trust in legitimate information sources.
45+
46+
**Real-Life Examples:**
47+
48+
- [Fake Zelensky video (2022)](https://www.npr.org/2022/03/16/1087062648/deepfake-video-zelenskyy-experts-war-manipulation-ukraine-russia): A deepfake video urged Ukrainian forces to surrender, illustrating how synthetic media can be weaponized during international conflicts.
49+
- Fake celebrity endorsements: AI-generated videos and images appear online, falsely promoting products or political messages by well-known public figures. e.g. [2022 Elon Musk Crypto Scam](https://www.vice.com/en/article/scammers-use-elon-musk-deepfake-to-steal-crypto/): A deepfake video circulated on social media, featuring a convincing impersonation of Elon Musk endorsing a fraudulent cryptocurrency platform.
50+
- [2019 Deepfake CEO Phone Scam](https://www.forbes.com/sites/jessedamiani/2019/09/03/a-voice-deepfake-was-used-to-scam-a-ceo-out-of-243000/): In a widely reported case, criminals used AI-generated voice to impersonate a chief executive, successfully convincing a subordinate to transfer \$243,000 to a fraudulent account.
51+
52+
### Predictive AI Systems Controlling Social Behavior
53+
54+
- Platforms use behavioral models to shape user experiences, potentially pushing them to adopt certain viewpoints or behaviors.
55+
- Data-driven predictions about habits and preferences can be used to modify or influence choices.
56+
57+
**Real-Life Examples:**
58+
59+
- TikTok’s recommendation algorithm: Known for its powerful engagement-driven feed, it can rapidly shape users’ content consumption, potentially reinforcing certain narratives or trends. See: [TikTok Algorithm Eating Disorder article on The Verge](https://www.theverge.com/2021/12/18/22843606/tiktok-wsj-algorithm-change-eating-disorder).
60+
61+
- China’s social credit initiatives: Although not purely about content manipulation, these systems use data and behavioral metrics to encourage or discourage particular actions, effectively guiding social behavior.
62+
63+
---
64+
65+
## Mitigations
66+
67+
### AI Transparency Regulations
68+
69+
- Mandate clear labeling of AI-generated content.
70+
- Require accountability and auditing mechanisms for social media platforms.
71+
- **Examples:** 
72+
- [Generative AI and watermarking - European Parliament](https://www.europarl.europa.eu/RegData/etudes/BRIE/2023/757583/EPRS_BRI\(2023\)757583_EN.pdf)
73+
- **Efficacy:** Medium – Transparency can deter some manipulative actors, but determined bad actors may still evade or exploit labelling.
74+
- **Ease of Implementation:** Moderate – Requires infrastructure for labelling, auditing, and enforcement, but could be mandated by legislation.
75+
76+
### Ethical AI Development Standards
77+
78+
- Industry-wide codes of conduct to discourage manipulative AI.
79+
- Incentivize designers to embed fairness and user consent into algorithmic systems.
80+
- **Examples**
81+
-  [Understanding artificial intelligence ethics and safety - Turing Institute](https://www.turing.ac.uk/sites/default/files/2019-06/understanding_artificial_intelligence_ethics_and_safety.pdf)
82+
- [AI Playbook for the UK Government](https://www.gov.uk/government/publications/ai-playbook-for-the-uk-government/artificial-intelligence-playbook-for-the-uk-government-html#principles)
83+
- [DOD Adopts Ethical Principles for Artificial Intelligence](https://www.defense.gov/News/Releases/Release/Article/2091996/dod-adopts-ethical-principles-for-artificial-intelligence/)
84+
- **Efficacy:** Medium – Encourages best practices and self-regulation, but relies on voluntary compliance without legal backing.
85+
- **Ease of Implementation:** Low – Professional bodies and industry coalitions can quickly adopt and publicize guidelines, though ensuring universal adherence remains a challenge. Firms have varying incentives, budgets, and ethical priorities, making universal buy-in elusive.
86+
87+
### Education & Public Awareness Campaigns
88+
89+
- Equip citizens with media literacy skills to spot deepfakes and manipulation attempts.
90+
- Encourage public understanding of how personal data can be exploited by AI-driven systems.
91+
- **Examples:**
92+
- https://newslit.org
93+
- https://www.unesco.org/en/media-information-literacy
94+
- **Efficacy:** High – Empowered, media-savvy populations are significantly harder to manipulate. However, scaling efforts to entire populations is a substantial challenge given diverse educational, cultural, and socioeconomic barriers.
95+
- **Ease of Implementation:** Low – While public outreach is feasible, achieving wide coverage and sustained engagement can be resource-intensive.  Overcoming entrenched biases, misinformation echo chambers, and public apathy is an uphill battle, particularly if there’s no supportive policy or consistent funding.
96+
97+

0 commit comments

Comments
 (0)