Activities per year
Project Details
Description
My research addresses the issue of explainability of Artificial Intelligence (AI) systems, specifically in the agricultural sector. AI, while beneficial in various fields, can be complex and opaque, potentially infringing on fundamental rights. These ‘black-box’ models have led to the development of Explainable AI (hereinafter: XAI), which aims to make AI decision-making processes understandable to humans.
In agriculture, AI applications are increasingly important. However, these models may lack transparency, which is crucial for farmers to trust and adopt AI systems.
The EU AI Act aims to regulate such AI systems based on their risk level. It includes rules on explainability, which could be interpreted as XAI requirements, but their interpretation and measurement remain unclear.
My research aims to explore the theoretical meaning and practical application of explainability as a legal requirement for (high-risk) AI systems in the EU. With doctrinal and empirical research, legal theory and practical implications are combined. I explore how explainability can be operationalised throughout the AI lifecycle of AI systems for field crop farming, to determine what the practical reality needs.
In agriculture, AI applications are increasingly important. However, these models may lack transparency, which is crucial for farmers to trust and adopt AI systems.
The EU AI Act aims to regulate such AI systems based on their risk level. It includes rules on explainability, which could be interpreted as XAI requirements, but their interpretation and measurement remain unclear.
My research aims to explore the theoretical meaning and practical application of explainability as a legal requirement for (high-risk) AI systems in the EU. With doctrinal and empirical research, legal theory and practical implications are combined. I explore how explainability can be operationalised throughout the AI lifecycle of AI systems for field crop farming, to determine what the practical reality needs.
| Status | Active |
|---|---|
| Effective start/end date | 1/09/24 → 31/08/28 |
Collaborative partners
- Hanze University of Applied Sciences (lead)
- University of Groningen, Faculty of Law
Keywords
- Artificial intelligence
- AI Act
- Agriculture
- Explainability
- AI systems
Research Focus Areas Hanze University of Applied Sciences * (mandatory by Hanze)
- Entrepreneurship
- Healthy Ageing
Research Focus Areas Research Centre or Centre of Expertise * (mandatory by Hanze)
- Digital Transformation
Publinova themes
- Economics and Management
- ICT and Media
- Law
- People and Society
- Health
- Technology
Fingerprint
Explore the research topics touched on by this project. These labels are generated based on the underlying awards/grants. Together they form a unique fingerprint.
Activities
-
Derde consortiumdag DigiAgro
Coppens, C. (Participant), van Beem, E. (Speaker), Groenboom, R. (Participant), Bredek, J. (Speaker) & Bialek, L. (Speaker)
26 Jun 2025Activity: Participating in or organising an event › Participation in workshop, seminar, course
-
Fourth European Workshop on Algorithmic Fairness (EWAF '25)
van Beem, E. (Speaker), van Beem, E. (Chair) & van Beem, E. (Participant)
30 Jun 2025 → 2 Jul 2025Activity: Participating in or organising an event › Participation in conference
-
Explainability of AI in the EU AI Act: functions lost and others found?
van Beem, E. (Speaker)
3 Apr 2025Activity: Talk or presentation › Oral presentation