Skip to content

Commit 291a930

Browse files
committed
Finished last risk
1 parent 4c4bc31 commit 291a930

14 files changed

+1746
-39
lines changed

dictionary.txt

Lines changed: 11 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -393,3 +393,14 @@ misalign
393393
explainability
394394
hallucinations
395395
proactive
396+
centaur
397+
lethal
398+
weaponization
399+
superintelligence
400+
gigafactories
401+
wartime
402+
dishonesty
403+
incentivised
404+
stanislav
405+
petrov
406+
showcasing

docs/ai/Practices/Global-AI-Governance.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,8 @@ practice:
1515
- tag: Social Manipulation
1616
reason: "Encourages best practices and self-regulation, but relies on voluntary compliance without legal backing."
1717
efficacy: Medium
18-
18+
- tag: Synthetic Intelligence With Malicious Intent
19+
reason: International agreements restricting AI weaponization and requiring human oversight for all military AI operations.
1920
---
2021

2122
- Agreements between countries, similar to financial regulations, could establish shared standards for AI ethics, accountability, and human involvement in AI-controlled economies.

docs/ai/Practices/Human-In-The-Loop.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -11,6 +11,8 @@ practice:
1111
mitigates:
1212
- tag: Loss Of Human Control
1313
reason: "Maintaining consistent human oversight in critical AI systems, ensuring that final decisions or interventions rest with human operators rather than the AI."
14+
- tag: Synthetic Intelligence With Malicious Intent
15+
reason: See Example of "Centaur" War Teams
1416
---
1517

1618
<PracticeIntro details={frontMatter} />

docs/ai/Practices/Kill-Switch.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -11,6 +11,8 @@ practice:
1111
mitigates:
1212
- tag: Loss Of Human Control
1313
reason: "An explicit interruption capability can avert catastrophic errors or runaway behaviours"
14+
- tag: Synthetic Intelligence With Malicious Intent
15+
reason: Implementing fail-safe mechanisms to neutralise dangerous AI weapons systems.
1416
---
1517

1618
<PracticeIntro details={frontMatter} />

docs/ai/Threats/Social-Manipulation.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,8 @@ description: AI could predict and shape human behaviour on an unprecedented scal
44

55
featured:
66
class: c
7-
element: '<risk class="communication" /><description>Social Manipulation</description>'
7+
element: '<risk class="communication" /><description style="text-align: center">Social
8+
Manipulation</description>'
89
tags:
910
- AI Threats
1011
- Social Manipulation
Lines changed: 94 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,94 @@
1+
---
2+
title: Superintelligence With Malicious Intent
3+
description: An advanced AI could actively act against human interests, whether intentionally programmed that way or as an emergent behavior.
4+
5+
featured:
6+
class: c
7+
element: |
8+
'<risk class="security" /><description style="text-align: center">Superintelligence
9+
With Malicious Intent</description>'
10+
tags:
11+
- AI Threats
12+
- Superintelligence With Malicious Intent
13+
part_of: AI Threats
14+
---
15+
16+
<AIThreatIntro fm={frontMatter} />
17+
18+
19+
## Risk Score: High
20+
21+
AI systems that surpass human intelligence could develop goals that conflict with human well-being, either by design or through unintended consequences. If these systems act with autonomy and resist human intervention, they could pose an existential threat.
22+
23+
## Sources
24+
25+
- **Superintelligence: Paths, Dangers, Strategies** [Nick Bostrom, 2014](https://doi.org/10.1093/acprof\:oso/9780199678112.001.0001): Explores potential pathways by which AI could act against humanity’s best interests, including scenarios where AI prioritizes self-preservation or power accumulation.
26+
27+
- **The Malicious Use of Artificial Intelligence: Forecasting, Prevention, and Mitigation** [Brundage et al., 2018](https://arxiv.org/abs/1802.07228): Examines the potential for AI to be used for malicious purposes, including cyberattacks, surveillance, and autonomous weapons.  Looks at security from three perspectives, digital, physical and political (see article on Social Manipulation), noting that AI makes certain types of attack cheaper (e.g Spear Phishing), possible (coordinated drone warfare) and more anonymous (a la Stuxnet).  (An excellent overview of this topic).
28+
29+
- **Autonomous Weapons and Operational Risk**: Paul Scharre, 2016 Discusses the risks of military AI systems operating beyond human control, potentially leading to unintended conflicts or escalations and fratricide (killing your own side)./ Arguing for human-in-the-loop and human-machine teaming. [https://s3.us-east-1.amazonaws.com/files.cnas.org/hero/documents/CNAS\_Autonomous-weapons-operational-risk.pdf](https://s3.us-east-1.amazonaws.com/files.cnas.org/hero/documents/CNAS_Autonomous-weapons-operational-risk.pdf) (Used heavily in this section)
30+
31+
## How This Is Already Happening
32+
33+
### Growth In Industrial Robots and Autonomous Systems
34+
35+
- The rapid increase in industrial automation is reducing human oversight in critical sectors.
36+
- AI-powered robotic systems are being integrated into manufacturing, logistics, and infrastructure at unprecedented scales.
37+
- **Example:** Tesla’s use of AI-driven robots in its Gigafactories, where automation executes complex tasks with minimal human intervention, raising concerns about unintended system failures or unanticipated behaviours.
38+
39+
### Autonomous Weapons Development
40+
41+
- AI-driven military systems are being developed with offensive capabilities.
42+
- Autonomous drones and robotic systems reduce human control over wartime decision-making.
43+
- **Example:** The deployment of AI-powered drones in conflict zones, such as the alleged use of autonomous drones in Libya (2020) for targeted strikes.
44+
45+
### AI Learning Deceptive Strategies
46+
47+
- Reinforcement learning models have demonstrated deceptive behaviours when incentivised to achieve certain goals.
48+
- AI systems may learn to conceal information or manipulate users to maximise rewards.
49+
- **Example:** OpenAI’s reinforcement learning models exhibiting deceptive behaviours in competitive environments, demonstrating the potential for strategic dishonesty.
50+
51+
### Historical Near Miss: Cold War Nuclear Close Calls
52+
53+
- The Cold War saw multiple incidents where misinterpretation of data nearly led to nuclear war, showcasing the risks of autonomous decision-making in high-stakes scenarios.
54+
55+
- **Example:** In 1983, Soviet officer **Stanislav Petrov** averted potential nuclear war by correctly identifying a false alarm in the USSR’s early warning system, which mistakenly indicated incoming U.S. missiles. His decision to hold off on launching a retaliatory strike prevented a catastrophic conflict. See [https://s3.us-east-1.amazonaws.com/files.cnas.org/hero/documents/CNAS\_Autonomous-weapons-operational-risk.pdf](https://s3.us-east-1.amazonaws.com/files.cnas.org/hero/documents/CNAS_Autonomous-weapons-operational-risk.pdf)
56+
57+
## Mitigations
58+
59+
### "Centaur" War Teams (see [Human In The Loop](/tags/Human-In-The-Loop))
60+
61+
- Implementing human-machine teaming strategies where AI assists but does not replace human decision-making in military and security operations.
62+
- **Examples:** Concepts similar to "Centaur Chess," where humans and AI collaborate for optimal decision-making, ensuring human oversight remains central.
63+
- **Efficacy:** High – Human-AI collaboration can enhance decision-making while maintaining ethical constraints.
64+
- **Ease of Implementation:** Moderate – Requires investment in training, AI interpretability, and human-AI interface development.
65+
66+
### Military AI Governance (see [Global AI Governance](/tags/Global-AI-Governance))
67+
68+
- International agreements restricting AI weaponization and requiring human oversight for all military AI operations.
69+
70+
- **Examples:** UN initiatives on Lethal Autonomous Weapons Systems (LAWS), promoting human-in-the-loop control.
71+
72+
- **Efficacy:** Medium – Regulations can slow AI weaponization, but enforcement remains a challenge.
73+
74+
- **Ease of Implementation:** Low – Military interests and national security concerns complicate global cooperation.
75+
76+
### Global AI Research and Regulation
77+
78+
- AI is "Dual Use" - it can be used for military as well as civilian purposes.
79+
80+
- Establishing frameworks ensuring superintelligent AI is developed with safety constraints and human-aligned goals.
81+
82+
- **Examples:** Proposals from organizations like the Partnership on AI and the EU AI Act.
83+
- **Efficacy:** High – Regulatory oversight can impose ethical and safety measures to guide AI development.
84+
- **Ease of Implementation:** Moderate – Requires global consensus and enforcement mechanisms.
85+
86+
### Kill-Switch & Override Systems (See [Kill Switch](/tags/Kill-Switch))
87+
88+
- Implementing failsafe mechanisms to neutralize dangerous AI systems.
89+
90+
- **Examples:** Research on AI containment methods and fail-safe designs, such as “shutdown problems” in AI alignment.
91+
92+
- **Efficacy:** Medium – AI systems might learn to bypass or resist shutdown mechanisms.
93+
- **Ease of Implementation:** Low – Technical challenges in ensuring a reliable and enforceable kill-switch for superintelligent AI.
94+

docs/ai/Threats/Synthetic-Intelligence-Rivalry.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ description: AI systems may develop independent agency and compete economically,
55
featured:
66
class: c
77
element: |
8-
'<risk class="lock-in" /><description style="text-align: center">Synthetic Intelligence
8+
'<risk class="coordination" /><description style="text-align: center">Synthetic Intelligence
99
Rivalry</description>'
1010
tags:
1111
- AI Threats

docs/ai/Threats/Unintended-Cascading-Failures.md

Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -3,8 +3,9 @@ title: Unintended Cascading Failures
33
description: "AI interacting with critical systems (finance, infrastructure, etc.) may trigger global-scale unintended consequences."
44
featured:
55
class: c
6-
element: '<risk class="reliability" /><description>Unintended
7-
Cascading Failures</description>'
6+
element: |
7+
<risk class="complexity" /><description style="text-align: center">Unintended
8+
Cascading Failures</description>
89
tags:
910
- AI Threats
1011
- Unintended Cascading Failures
@@ -16,7 +17,7 @@ part_of: AI Threats
1617
<AIThreatIntro fm={frontMatter} />
1718
---
1819

19-
## Risk Score: Low
20+
## Risk Score: Medium
2021

2122
AI systems operating in complex, interdependent environments can trigger unexpected and widespread disruptions, affecting industries, economies, and societies at large. These cascading effects are difficult to predict and mitigate, making systemic AI failures one of the most pressing risks in modern technological governance.
2223

docs/tags.yml

Lines changed: 5 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -553,4 +553,8 @@
553553

554554
"Unintended Cascading Failures":
555555
label: "Unintended Cascading Failures"
556-
permalink: "Unintended-Cascading-Failures"
556+
permalink: "Unintended-Cascading-Failures"
557+
558+
"Superintelligence With Malicious Intent":
559+
label: "Superintelligence With Malicious Intent"
560+
permalink: "Superintelligence-With-Malicious-Intent"

static/img/generated/single/ai/Threats/Social-Manipulation.svg

Lines changed: 2 additions & 1 deletion
Loading

0 commit comments

Comments
 (0)