Skip to content

Commit b113663

Browse files
authored
Merge pull request #146 from risk-first/practice-pages
AI Risks
2 parents e80df7f + 2b5dc13 commit b113663

File tree

91 files changed

+53262
-29
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

91 files changed

+53262
-29
lines changed

dictionary.txt

Lines changed: 23 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -381,3 +381,26 @@ tarnish
381381
goveranance
382382
subpostmasters
383383
exonerate
384+
disinformation
385+
systemic
386+
underscoring
387+
deepfake
388+
deepfakes
389+
grampian
390+
interpretability
391+
unsupervised
392+
misalign
393+
explainability
394+
hallucinations
395+
proactive
396+
centaur
397+
lethal
398+
weaponization
399+
superintelligence
400+
gigafactories
401+
wartime
402+
dishonesty
403+
incentivised
404+
stanislav
405+
petrov
406+
showcasing
Lines changed: 21 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,21 @@
1+
---
2+
title: Ecosystem Diversity
3+
description: Encouraging the development of multiple, independent AI models instead of relying on a single dominant system.
4+
featured:
5+
class: c
6+
element: '<action>Ecosystem Diversity</action>'
7+
tags:
8+
- Ecosystem Diversity
9+
- AI Practice
10+
practice:
11+
mitigates:
12+
- tag: Loss Of Diversity
13+
reason: "Diversified AI systems reduce systemic risks and encourage innovation."
14+
efficacy: High
15+
---
16+
17+
<PracticeIntro details={frontMatter} />
18+
19+
- Encouraging the development of multiple, independent AI models instead of relying on a single dominant system.
20+
21+
- Requires regulatory incentives or decentralised AI development initiatives.
Lines changed: 40 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,40 @@
1+
---
2+
title: Global AI Governance
3+
description: International cooperation is necessary to prevent AI firms from evading national regulations by relocating to jurisdictions with lax oversight.
4+
featured:
5+
class: c
6+
element: '<action>National AI Regulation</action>'
7+
tags:
8+
- National AI Regulation
9+
- AI Practice
10+
practice:
11+
mitigates:
12+
- tag: Synthetic Intelligence Rivalry
13+
reason: "Can provide international oversight, but effectiveness depends on cooperation among nations."
14+
efficacy: Medium
15+
- tag: Social Manipulation
16+
reason: "Encourages best practices and self-regulation, but relies on voluntary compliance without legal backing."
17+
efficacy: Medium
18+
- tag: Synthetic Intelligence With Malicious Intent
19+
reason: International agreements restricting AI weaponization and requiring human oversight for all military AI operations.
20+
---
21+
22+
<PracticeIntro details={frontMatter} />
23+
24+
- Agreements between countries, similar to financial regulations, could establish shared standards for AI ethics, accountability, and human involvement in AI-controlled economies.
25+
26+
- Challenging to implement due to differing national interests, enforcement issues, and political resistance.
27+
28+
- Industry-wide codes of conduct to discourage manipulative AI.
29+
30+
- Incentivize designers to embed fairness and user consent into algorithmic systems.
31+
32+
- Professional bodies and industry coalitions can quickly adopt and publicise guidelines, though ensuring universal adherence remains a challenge. Firms have varying incentives, budgets, and ethical priorities, making universal buy-in elusive.
33+
34+
## Examples
35+
36+
-  [Understanding artificial intelligence ethics and safety - Turing Institute](https://www.turing.ac.uk/sites/default/files/2019-06/understanding_artificial_intelligence_ethics_and_safety.pdf)
37+
38+
- [AI Playbook for the UK Government](https://www.gov.uk/government/publications/ai-playbook-for-the-uk-government/artificial-intelligence-playbook-for-the-uk-government-html#principles)
39+
40+
- [DOD Adopts Ethical Principles for Artificial Intelligence](https://www.defense.gov/News/Releases/Release/Article/2091996/dod-adopts-ethical-principles-for-artificial-intelligence/)
Lines changed: 30 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,30 @@
1+
---
2+
title: Human In The Loop
3+
description: Consistent human oversight in critical AI systems.
4+
featured:
5+
class: c
6+
element: '<action>Human In The Loop</action>'
7+
tags:
8+
- Human In The Loop
9+
- AI Practice
10+
practice:
11+
mitigates:
12+
- tag: Loss Of Human Control
13+
reason: "Maintaining consistent human oversight in critical AI systems, ensuring that final decisions or interventions rest with human operators rather than the AI."
14+
- tag: Synthetic Intelligence With Malicious Intent
15+
reason: See Example of "Centaur" War Teams
16+
---
17+
18+
<PracticeIntro details={frontMatter} />
19+
20+
- Maintaining consistent human oversight in critical AI systems, ensuring that final decisions or interventions rest with human operators rather than the AI.
21+
- AI may suggest diagnoses or treatments, but a certified professional reviews and confirms before enacting them. In the above NHS Grampian example, the AI is augmenting human decision making with a third opinion, rather than replacing human judgement altogether (yet).
22+
- Some proposals mandate that human operators confirm critical actions (e.g., missile launches), preventing AI from unilaterally making life-or-death decisions. This might work in scenarios where response time isn't a factor.
23+
24+
## Types Of Human In The Loop
25+
26+
> - **Semi-autonomous operation**: machine performs a task and then stops and waits for approval from the human operator before continuing. This control type is often referred to as "human in the loop."
27+
> - **Supervised autonomous operation**, where the machine, once activated, performs a task under the supervision of a human and will continue performing the task unless the human operator intervenes to halt its operation. This control type is often referred to as “human on the loop.”
28+
> - **Fully autonomous operation**, where the machine, once activated, performs a task and the human operator does not have the ability to supervise its operation and intervene in the event of system failure. This control type is often referred to as “human out of the loop.”
29+
30+
From: https://s3.us-east-1.amazonaws.com/files.cnas.org/hero/documents/CNAS_Autonomous-weapons-operational-risk.pdf
Lines changed: 20 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,20 @@
1+
---
2+
title: Interpretability
3+
description: Developing tools to analyse AI decision-making processes and detect emergent behaviors before they become risks.
4+
featured:
5+
class: c
6+
element: '<action>Interpretability</action>'
7+
tags:
8+
- Interpretability
9+
- AI Practice
10+
practice:
11+
mitigates:
12+
- tag: Emergent Behaviour
13+
reason: "An explicit interruption capability can avert catastrophic errors or runaway behaviours"
14+
---
15+
16+
<PracticeIntro details={frontMatter} />
17+
18+
- Helps understand AI behavior but does not prevent emergent capabilities from appearing.
19+
20+
- Research in explainable AI is advancing, but understanding deep learning models remains complex.

docs/ai/Practices/Kill-Switch.md

Lines changed: 24 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,24 @@
1+
---
2+
title: Kill Switch
3+
description: Fail-safe systems capable of shutting down or isolating AI processes if they exhibit dangerous behaviours.
4+
featured:
5+
class: c
6+
element: '<action>Kill Switch Mechanism</action>'
7+
tags:
8+
- Kill Switch
9+
- AI Practice
10+
practice:
11+
mitigates:
12+
- tag: Loss Of Human Control
13+
reason: "An explicit interruption capability can avert catastrophic errors or runaway behaviours"
14+
- tag: Synthetic Intelligence With Malicious Intent
15+
reason: Implementing fail-safe mechanisms to neutralise dangerous AI weapons systems.
16+
---
17+
18+
<PracticeIntro details={frontMatter} />
19+
20+
## Examples
21+
22+
- **Google DeepMind’s ‘Big Red Button’ concept** (2016), proposed as a method to interrupt a reinforcement learning AI without it learning to resist interruption.
23+
24+
- **Hardware Interrupts in Robotics:** Physical or software-based emergency stops that immediately terminate AI operation.
Lines changed: 23 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,23 @@
1+
---
2+
title: Multi-Stakeholder Oversight
3+
description: Governments can implement policies that ensure AI-driven firms remain accountable to human oversight.
4+
featured:
5+
class: c
6+
element: '<action>Multi-Stakeholder Oversight</action>'
7+
tags:
8+
- Multi-Stakeholder Oversight
9+
- AI Practice
10+
practice:
11+
mitigates:
12+
- tag: Loss Of Diversity
13+
reason: "Ensuring that AI governance involves multiple institutions, including governments, researchers, and civil society, to prevent monopolisation."
14+
efficacy: Medium
15+
---
16+
17+
<PracticeIntro details={frontMatter} />
18+
19+
- Ensuring that AI governance involves multiple institutions, including governments, researchers, and civil society, to prevent monopolisation.
20+
21+
- Helps distribute AI power more equitably but may struggle with enforcement.
22+
23+
- Requires cooperation between multiple sectors, which can be slow and politically complex.
Lines changed: 25 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,25 @@
1+
---
2+
title: National AI Regulation
3+
description: Governments can implement policies that ensure AI-driven firms remain accountable to human oversight.
4+
featured:
5+
class: c
6+
element: '<action>National AI Regulation</action>'
7+
tags:
8+
- National AI Regulation
9+
- AI Practice
10+
practice:
11+
mitigates:
12+
- tag: Synthetic Intelligence Rivalry
13+
reason: "Government policies can strongly influence AI firms' behavior if enforced effectively."
14+
efficacy: High
15+
- tag: Loss Of Diversity
16+
reason: "Antitrust Regulations – Breaking up AI monopolies."
17+
---
18+
19+
<PracticeIntro details={frontMatter} />
20+
21+
- Governments can implement policies that ensure AI-driven firms remain accountable to human oversight. This could include requiring AI systems to maintain transparency, adhere to ethical standards, and uphold employment obligations by ensuring that a minimum level of human involvement remains in corporate decision-making.
22+
23+
- Requires strong legal frameworks, enforcement mechanisms, and political will.
24+
25+
- Antitrust Regulations – Breaking up AI monopolies.
Lines changed: 29 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,29 @@
1+
---
2+
title: Public Awareness
3+
description: Equip citizens with media literacy skills to spot deepfakes and manipulation attempts.
4+
featured:
5+
class: c
6+
element: '<action>Public Awareness</action>'
7+
tags:
8+
- Public Awareness
9+
- AI Practice
10+
practice:
11+
mitigates:
12+
- tag: Social Manipulation
13+
reason: "Empowered, media-savvy populations are significantly harder to manipulate. However, scaling efforts to entire populations is a substantial challenge given diverse educational, cultural, and socioeconomic barriers."
14+
efficacy: Medium
15+
---
16+
17+
<PracticeIntro details={frontMatter} />
18+
19+
- Equip citizens with media literacy skills to spot deepfakes and manipulation attempts.
20+
21+
- Encourage public understanding of how personal data can be exploited by AI-driven systems.
22+
23+
- While public outreach is feasible, achieving wide coverage and sustained engagement can be resource-intensive.  Overcoming entrenched biases, misinformation echo chambers, and public apathy is an uphill battle, particularly if there’s no supportive policy or consistent funding.
24+
25+
## Examples
26+
27+
- https://newslit.org
28+
29+
- https://www.unesco.org/en/media-information-literacy
Lines changed: 29 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,29 @@
1+
---
2+
title: Replication Control
3+
description: Replication control becomes relevant when an AI system can duplicate itself—or be duplicated—beyond the reach of any central authority.
4+
featured:
5+
class: c
6+
element: '<action>Replication Control</action>'
7+
tags:
8+
- Replication Control
9+
- AI Practice
10+
practice:
11+
mitigates:
12+
- tag: Loss Of Human Control
13+
reason: "An explicit interruption capability can avert catastrophic errors or runaway behaviours"
14+
- tag: Emergent Behaviour
15+
reason: "Preventing self-replicating AI or unsupervised proliferation of emergent behaviours by implementing strict replication oversight."
16+
efficacy: High
17+
---
18+
19+
<PracticeIntro details={frontMatter} />
20+
21+
- Replication control becomes relevant when an AI system can duplicate itself—or be duplicated—beyond the reach of any central authority (analogous to a computer virus—though with potentially far greater autonomy and adaptability).
22+
23+
- An organization/person builds a very capable AI with some misaligned objectives. If they distribute its model or code openly, it effectively becomes “in the wild.”
24+
25+
- Could controls be put in place to prevent this from happening? TODO: figure this out.
26+
27+
- In open-source communities or decentralised systems, controlling replication requires broad consensus and technical enforcement measures.
28+
29+

0 commit comments

Comments
 (0)