first of all ... should we even bother?
Yes — it is possible to argue this, and some researchers do argue it.
But it is not the majority view, and the argument has limits.
“Explainability is not a worthy exercise because we are not meant to know the internal logic.”
This argument usually comes in three forms:
1. “Opaque complexity is natural and unavoidable.”
Just like:
the human brain
evolution
weather
even some quantum systems
…AI systems may have internal structures too large or high-dimensional for humans to intuitively understand.
So the argument goes:
If the system’s internal logic is inherently incomprehensible, then trying to “explain” it in simple human terms is fundamentally misguided.
2. “AI is not built for human-legible reasoning.”
Neural networks don’t learn logic rules. They learn:
distributed embeddings
nonlinear transformations
emergent features
As a result:
The model doesn’t have human-like reasoning inside it, so there is nothing to "explain" in the way people expect.
This argument says explainability forces humans’ expectations onto an alien computational structure.
3. “Performance matters more than interpretability.”
Some argue that:
aircraft designers don’t fully track airflow around every molecule
we trust antibiotics without understanding all biological processes
we use GPS without knowing general relativity
By analogy:
If the model is reliable, safe, and tested, we don’t need to know why it works.
Explainability is seen as a luxury, not a requirement.
Counterarguments (why explainability is considered important)
Even if the internal logic is inherently non-human, explainability is still critical for:
1. Safety
Without interpretability, we cannot detect:
hidden dangerous behaviors
deceptive alignment
bias
catastrophic failure modes
2. Trust
Regulators, auditors, and users need transparency for high-stakes applications:
medical diagnosis
credit scoring
criminal justice
autonomous vehicles
3. Debugging
Engineers need interpretability to fix:
hallucinations
brittle features
failures under distribution shift
4. Moral/Legal Responsibility
Governments increasingly require explainability:
EU AI Act
US FTC guidelines
Right to Explanation laws
So even if internal logic is alien, external explanations (counterfactuals, feature relevance, causal tests) are still practically necessary.
Philosophical Question: “Are we not meant to know the internal logic?”
This is a deeper, almost metaphysical point.
Some argue:
neural networks evolve internal structures through optimization
these structures are not designed to be interpretable
therefore, pursuing “full understanding” may be like trying to understand every neuron in a human brain:
possible in principle, but not what the system is for
This is similar to arguments in:
evolutionary biology
cognitive science
complex systems theory
The system’s function is primary; transparency is a human desire layered on top.
For me:
Yes, it is logically and philosophically valid to argue that explainability research is not inherently meaningful if:
the system’s internal logic is fundamentally distant from human understanding,
and the system can be validated purely through external behavior.
However, in engineering, law, safety, and ethics, explainability remains critical even if the internal mechanisms can never be fully understood.
So I operate with the logic that:
We may never understand the internal logic fully, but we still need explainability tools to understand the consequences of that logic.
