Artificial intelligence is becoming increasingly present in everyday life – from algorithms that decide what we watch on Netflix, to systems that help doctors with diagnoses, or self-driving cars on our roads. But as AI systems become more involved in decision-making, a crucial question arises: who is responsible if AI makes a mistake?
First – what does it mean when AI “makes a mistake”?
AI doesn’t “make mistakes” the way humans do. It has no consciousness, intention, or sense of guilt. A mistake in the context of AI means the system made a decision that was inaccurate, unfair, or unsafe. Examples include:
- Facial recognition that doesn’t identify certain ethnic groups with equal accuracy
- An algorithm that automatically denies loan applications based on biased data
- A self-driving car that causes a traffic accident
Who bears responsibility?
The answer isn’t simple, but here are a few perspectives:
- Developers and engineers
They create the systems, train the models, and choose the data. If an AI system makes decisions based on biased or incomplete data, the responsibility may lie with those who designed it.
- Companies using AI
If a company uses AI to automate processes (e.g. hiring, medical diagnoses) and that system discriminates or fails, ethical (and legal) responsibility may fall on the employer, even if they didn’t directly develop the system.
- Users
In some cases, end users may be responsible if they used the tool in an unintended way, ignored recommendations, or failed to supervise the AI’s decisions.
- Lawmakers and regulatory bodies
If there is no clear legal framework defining how AI should be used and what may be automated, it’s hard to assign responsibility. The question is: why hasn’t the law kept up with technology?
Ethics vs. Law: Not the Same
Interestingly, what is legally permitted is not always ethically right. AI can, for example, rank job candidates using criteria that aren’t illegal, but are still deeply biased or unfair.
This brings us to a key question: should AI systems be “explainable”? If we don’t know how the system made its decision, it’s difficult to determine where the mistake happened – and even harder to assign blame.
What does the future hold?
Introducing AI into all areas of life calls for a new kind of responsibility – a collective one. We may soon have “AI ethicists,” mandatory risk assessments before systems go live, or even laws regulating algorithmic transparency.
In any case, the question of responsibility won’t disappear. On the contrary – it will become increasingly important as AI systems begin to make more serious decisions.
AI doesn’t exist in a vacuum. If we want fair and responsible technologies, we need to start thinking responsibly today – before a major mistake happens.

Croatia
Bosnia and Herzegovina
Serbia
Crna Gora
North Macedonia
Ukraine
Albania
Kosovo
Austria
Deutschland
Switzerland

