Quick Risk Pre-Scan¶
1. Purpose¶
This shortened risk scan identifies the most critical blockers for your AI prototype in 20–30 minutes. Conduct it on days 3–4 of the 30-Day Explorer Kit.
Relationship to the full Pre-Scan
This is a simplified version of the full Risk Pre-Scan. If the result is amber or red, or if in doubt, always complete the full version.
Project: [Name] Completed by: [Name + role] Date: [Date]
2. Part A — Hard Blockers (Stop Questions)¶
If you answer "Yes" to any of these questions: STOP this project immediately and consult the Compliance Hub.
Prohibited practices (EU AI Act Art. 5)
- Does the system use subliminal or manipulative techniques to influence human behaviour without the person's knowledge?
- Does the system apply biometric categorisation based on sensitive characteristics (race, political opinions, religion)?
- Does the system perform real-time biometric identification in public spaces?
- Does the system evaluate individuals based on social behaviour ("social scoring")?
→ If one or more "Yes": PROJECT BLOCKED. Consult EU AI Act.
3. Part B — High-Risk Indicators¶
Score each question: 0 = No / 1 = Partially / 2 = Yes
B1 — Application Domain¶
| Question | Score (0/1/2) |
|---|---|
| Is the system deployed in critical infrastructure (energy, water, transport)? | |
| Does it decide on access to education, employment or social services? | |
| Does it decide on credit, insurance or financial services? | |
| Is it deployed in law enforcement, migration or the justice system? | |
| Does the system affect safety (physical harm possible)? |
Subtotal B1: ___/10
B2 — Data & Privacy¶
| Question | Score (0/1/2) |
|---|---|
| Does the system process personal data (GDPR)? | |
| Does the training or inference data contain special categories (health, political, biometric)? | |
| Is data from minors being processed? | |
| Is the data source external/unknown (e.g. web scraping)? | |
| Are user interactions stored without explicit consent? |
Subtotal B2: ___/10
B3 — Autonomy & Impact¶
| Question | Score (0/1/2) |
|---|---|
| Does the system make decisions without human intervention that impact individuals? | |
| Are the consequences of an error difficult to reverse? | |
| Are there no alternative control measures if the system fails? | |
| Does the system interact directly with end users who do not know it is AI? | |
| Does the system affect labour-related decisions (evaluation, selection, dismissal)? |
Subtotal B3: ___/10
4. Score Calculation¶
Total score Part B: Subtotal B1 + B2 + B3 = ___/30
| Total score | Colour code | Interpretation | Action |
|---|---|---|---|
| 0–6 | 🟢 Green | Low risk — proceed | Document and move to day 5 |
| 7–15 | 🟡 Amber | Elevated risk — additional measures required | Complete the full Pre-Scan; schedule a risk session with a stakeholder |
| 16–30 | 🔴 Red | High risk — stop or redefine | Complete the full Pre-Scan; consult a legal adviser |
5. Part C — Transparency & Governance (Baseline Checks)¶
Always complete, regardless of Part B score.
Minimum requirements for prototype
- Transparency: End users know they are interacting with an AI system (no hidden AI)
- Human oversight: There is always a human who can review and correct the AI output
- Hard Boundaries: We have defined at least 2 concrete boundaries on what the system NEVER does
- Logging: We log inputs and outputs of the prototype (also for troubleshooting)
- Accountable person: One individual bears ultimate responsibility for this system
6. Conclusion & Next Step¶
Risk score: [ ] Green [ ] Amber [ ] Red
Remarks:
[Note any specific risks that deserve extra attention, even if the total score is green.]
Established Hard Boundaries:
- [E.g. The system never automatically sends communications without human approval.]
- [E.g. The system never processes personal data outside the EU without explicit consent.]
Next step:
- Green: Document in Project Charter Light, section 5, and proceed
- Amber: Schedule risk session and complete full Pre-Scan before day 9
- Red: Discuss with Sponsor. Consider redefining the use case
7. Related Modules¶
- Full Risk Pre-Scan
- EU AI Act Overview
- Risk Classification Framework
- AI Collaboration Modes
- Privacy & Data Sheet